From b9bed81e6e7aa934bafa4c130b1f70246bd656c8 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Thu, 21 Jun 2018 17:35:07 -0700 Subject: [PATCH 001/152] Initial V2 alloc runner --- client/allocrunnerv2/alloc_runner.go | 160 +++++++ client/allocrunnerv2/alloc_runner_getters.go | 15 + client/allocrunnerv2/alloc_runner_hooks.go | 176 ++++++++ client/allocrunnerv2/alloc_runner_test.go | 91 ++++ client/allocrunnerv2/config/config.go | 26 ++ client/allocrunnerv2/interfaces/factory.go | 10 + client/allocrunnerv2/interfaces/runner.go | 20 + .../interfaces/runner_lifecycle.go | 37 ++ .../interfaces/task_lifecycle.go | 92 ++++ client/allocrunnerv2/state/state.go | 38 ++ .../allocrunnerv2/taskrunner/state/state.go | 47 ++ .../allocrunnerv2/taskrunner/task_runner.go | 425 ++++++++++++++++++ .../taskrunner/task_runner_getters.go | 15 + .../taskrunner/task_runner_hooks.go | 285 ++++++++++++ client/allocrunnerv2/tr_shim.go | 31 ++ client/allocrunnerv2/util.go | 1 + client/driver/env/env.go | 10 + 17 files changed, 1479 insertions(+) create mode 100644 client/allocrunnerv2/alloc_runner.go create mode 100644 client/allocrunnerv2/alloc_runner_getters.go create mode 100644 client/allocrunnerv2/alloc_runner_hooks.go create mode 100644 client/allocrunnerv2/alloc_runner_test.go create mode 100644 client/allocrunnerv2/config/config.go create mode 100644 client/allocrunnerv2/interfaces/factory.go create mode 100644 client/allocrunnerv2/interfaces/runner.go create mode 100644 client/allocrunnerv2/interfaces/runner_lifecycle.go create mode 100644 client/allocrunnerv2/interfaces/task_lifecycle.go create mode 100644 client/allocrunnerv2/state/state.go create mode 100644 client/allocrunnerv2/taskrunner/state/state.go create mode 100644 client/allocrunnerv2/taskrunner/task_runner.go create mode 100644 client/allocrunnerv2/taskrunner/task_runner_getters.go create mode 100644 client/allocrunnerv2/taskrunner/task_runner_hooks.go create mode 100644 client/allocrunnerv2/tr_shim.go create mode 100644 client/allocrunnerv2/util.go diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go new file mode 100644 index 000000000..8eaa4112e --- /dev/null +++ b/client/allocrunnerv2/alloc_runner.go @@ -0,0 +1,160 @@ +package allocrunnerv2 + +import ( + "context" + "fmt" + "sync" + + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocdir" + "github.com/hashicorp/nomad/client/allocrunnerv2/config" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunnerv2/state" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" + "github.com/hashicorp/nomad/nomad/structs" +) + +// allocRunner is used to run all the tasks in a given allocation +type allocRunner struct { + // ctx is the alloc runners context. It is cancelled when all related + // activity for this allocation should be terminated. + ctx context.Context + + // ctxCancel is used to cancel the alloc runners cancel + ctxCancel context.CancelFunc + + // Logger is the logger for the alloc runner. + logger log.Logger + + // Config is the configuration for the alloc runner. + config *config.Config + + // waitCh is closed when the alloc runner has transitioned to a terminal + // state + waitCh chan struct{} + + // Alloc captures the allocation being run. + alloc *structs.Allocation + allocLock sync.Mutex + + // state captures the state of the alloc runner + state *state.State + + // allocDir is used to build the allocations directory structure. + allocDir *allocdir.AllocDir + + // runnerHooks are alloc runner lifecycle hooks that should be run on state + // transistions. + runnerHooks []interfaces.RunnerHook + + // tasks are the set of task runners + tasks map[string]*taskrunner.TaskRunner + taskLock sync.Mutex +} + +// NewAllocRunner returns a new allocation runner. +func NewAllocRunner(ctx context.Context, config *config.Config) (*allocRunner, error) { + // Create a context for the runner + arCtx, arCancel := context.WithCancel(ctx) + + ar := &allocRunner{ + ctx: arCtx, + ctxCancel: arCancel, + config: config, + alloc: config.Allocation, + waitCh: make(chan struct{}), + } + + // Create the logger based on the allocation ID + ar.logger = config.Logger.With("alloc_id", ar.ID()) + + // Initialize the runners hooks. + ar.initRunnerHooks() + + return ar, nil +} + +func (ar *allocRunner) WaitCh() <-chan struct{} { + return ar.waitCh +} + +// XXX How does alloc Restart work +// Run is the main go-routine that executes all the tasks. +func (ar *allocRunner) Run() { + // Close the wait channel + defer close(ar.waitCh) + + // Run the prerun hooks + // XXX Equivalent to TR.Prerun hook + if err := ar.prerun(); err != nil { + ar.logger.Error("prerun failed", "error", err) + goto POST + } + + // Run the runners + if err := ar.runImpl(); err != nil { + ar.logger.Error("starting tasks failed", "error", err) + } + +POST: + // Run the postrun hooks + // XXX Equivalent to TR.Poststop hook + if err := ar.postrun(); err != nil { + ar.logger.Error("postrun failed", "error", err) + } +} + +// runImpl is used to run the runners. +func (ar *allocRunner) runImpl() error { + // Grab the task group + tg := ar.alloc.Job.LookupTaskGroup(ar.alloc.TaskGroup) + if tg == nil { + // XXX Fail and exit + ar.logger.Error("failed to lookup task group", "task_group", ar.alloc.TaskGroup) + return fmt.Errorf("failed to lookup task group %q", ar.alloc.TaskGroup) + } + + for _, task := range tg.Tasks { + if err := ar.runTask(task); err != nil { + return err + } + } + + // Block until all tasks are done. + var wg sync.WaitGroup + ar.taskLock.Lock() + for _, task := range ar.tasks { + wg.Add(1) + go func() { + <-task.WaitCh() + wg.Done() + }() + } + ar.taskLock.Unlock() + + wg.Wait() + return nil +} + +// runTask is used to run a task. +func (ar *allocRunner) runTask(task *structs.Task) error { + // Create the runner + config := &taskrunner.Config{ + Parent: &allocRunnerShim{ar}, + Task: task, + Logger: ar.logger, + } + tr, err := taskrunner.NewTaskRunner(config) + if err != nil { + return err + } + + // Start the runner + go tr.Run() + + // Store the runner + ar.taskLock.Lock() + ar.tasks[task.Name] = tr + ar.taskLock.Unlock() + return nil +} diff --git a/client/allocrunnerv2/alloc_runner_getters.go b/client/allocrunnerv2/alloc_runner_getters.go new file mode 100644 index 000000000..92f6c5faa --- /dev/null +++ b/client/allocrunnerv2/alloc_runner_getters.go @@ -0,0 +1,15 @@ +package allocrunnerv2 + +import "github.com/hashicorp/nomad/nomad/structs" + +func (ar *allocRunner) ID() string { + ar.allocLock.Lock() + defer ar.allocLock.Unlock() + return ar.alloc.ID +} + +func (ar *allocRunner) Alloc() *structs.Allocation { + ar.allocLock.Lock() + defer ar.allocLock.Unlock() + return ar.alloc +} diff --git a/client/allocrunnerv2/alloc_runner_hooks.go b/client/allocrunnerv2/alloc_runner_hooks.go new file mode 100644 index 000000000..085020813 --- /dev/null +++ b/client/allocrunnerv2/alloc_runner_hooks.go @@ -0,0 +1,176 @@ +package allocrunnerv2 + +import ( + "context" + "fmt" + "time" + + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" +) + +// initRunnerHooks intializes the runners hooks. +func (ar *allocRunner) initRunnerHooks() { + hookLogger := ar.logger.Named("runner_hook") + ar.runnerHooks = make([]interfaces.RunnerHook, 0, 3) + + // Create the alloc directory hook. This is run first to ensure the + // directoy path exists for other hooks. + ar.runnerHooks = append(ar.runnerHooks, newAllocDirHook(ar, hookLogger)) +} + +// prerun is used to run the runners prerun hooks. +func (ar *allocRunner) prerun() error { + if ar.logger.IsTrace() { + start := time.Now() + ar.logger.Trace("running pre-run hooks", "start", start) + defer func() { + end := time.Now() + ar.logger.Trace("finished pre-run hooks", "end", end, "duration", end.Sub(start)) + }() + } + + for _, hook := range ar.runnerHooks { + pre, ok := hook.(interfaces.RunnerPrerunHook) + if !ok { + continue + } + + name := pre.Name() + var start time.Time + if ar.logger.IsTrace() { + start = time.Now() + ar.logger.Trace("running pre-run hook", "name", name, "start", start) + } + + if err := pre.Prerun(); err != nil { + return fmt.Errorf("hook %q failed: %v", name, err) + } + + if ar.logger.IsTrace() { + end := time.Now() + ar.logger.Trace("finished pre-run hooks", "name", name, "end", end, "duration", end.Sub(start)) + } + } + + return nil +} + +// postrun is used to run the runners postrun hooks. +func (ar *allocRunner) postrun() error { + if ar.logger.IsTrace() { + start := time.Now() + ar.logger.Trace("running post-run hooks", "start", start) + defer func() { + end := time.Now() + ar.logger.Trace("finished post-run hooks", "end", end, "duration", end.Sub(start)) + }() + } + + for _, hook := range ar.runnerHooks { + post, ok := hook.(interfaces.RunnerPostrunHook) + if !ok { + continue + } + + name := post.Name() + var start time.Time + if ar.logger.IsTrace() { + start = time.Now() + ar.logger.Trace("running post-run hook", "name", name, "start", start) + } + + if err := post.Postrun(); err != nil { + return fmt.Errorf("hook %q failed: %v", name, err) + } + + if ar.logger.IsTrace() { + end := time.Now() + ar.logger.Trace("finished post-run hooks", "name", name, "end", end, "duration", end.Sub(start)) + } + } + + return nil +} + +/* +What state is needed to transfer: +*/ + +/* +AR Hooks: +Alloc Dir Build: +Needs to know the folder to create + +Alloc Migrate +Needs access to RPC + +Alloc Health Watcher: +Requires: Access to consul to watch health, access to every task event, task status change +*/ + +type allocDirHook struct { + runner *allocRunner + logger log.Logger +} + +func newAllocDirHook(runner *allocRunner, logger log.Logger) *allocDirHook { + ad := &allocDirHook{ + runner: runner, + } + ad.logger = logger.Named(ad.Name()) + return ad +} + +func (h *allocDirHook) Name() string { + return "alloc_dir" +} + +func (h *allocDirHook) Prerun() error { + return h.runner.allocDir.Build() +} + +func (h *allocDirHook) Destroy() error { + return h.runner.allocDir.Destroy() +} + +// TODO +type allocHealthWatcherHook struct { + runner *allocRunner + logger log.Logger + ctx context.Context + cancelFn context.CancelFunc +} + +func newAllocHealthWatcherHook(runner *allocRunner, logger log.Logger) *allocHealthWatcherHook { + ctx, cancelFn := context.WithCancel(runner.ctx) + ad := &allocHealthWatcherHook{ + runner: runner, + ctx: ctx, + cancelFn: cancelFn, + } + + ad.logger = logger.Named(ad.Name()) + return ad +} + +func (h *allocHealthWatcherHook) Name() string { + return "alloc_health_watcher" +} + +func (h *allocHealthWatcherHook) Prerun() error { + return nil +} + +func (h *allocHealthWatcherHook) Update() error { + // Cancel the old watcher and create a new one + h.cancelFn() + + // TODO create the new one + return nil +} + +func (h *allocHealthWatcherHook) Destroy() error { + h.cancelFn() + return nil +} diff --git a/client/allocrunnerv2/alloc_runner_test.go b/client/allocrunnerv2/alloc_runner_test.go new file mode 100644 index 000000000..b7c8ee282 --- /dev/null +++ b/client/allocrunnerv2/alloc_runner_test.go @@ -0,0 +1,91 @@ +package allocrunnerv2 + +import ( + "context" + "testing" + + "github.com/hashicorp/nomad/client/allocrunnerv2/config" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + clientconfig "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/nomad/mock" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/stretchr/testify/require" +) + +func testAllocRunnerFromAlloc(t *testing.T, alloc *structs.Allocation) *allocRunner { + cconf := clientconfig.DefaultConfig() + config := &config.Config{ + ClientConfig: cconf, + Logger: testlog.HCLogger(t).With("unit_test", t.Name()), + Allocation: alloc, + } + + ar, err := NewAllocRunner(context.Background(), config) + if err != nil { + t.Fatalf("Failed to create test alloc runner: %v", err) + } + + return ar + +} + +func testAllocRunner(t *testing.T) *allocRunner { + return testAllocRunnerFromAlloc(t, mock.Alloc()) +} + +// preRun is a test RunnerHook that captures whether Prerun was called on it +type preRun struct{ run bool } + +func (p *preRun) Name() string { return "pre" } +func (p *preRun) Prerun() error { + p.run = true + return nil +} + +// postRun is a test RunnerHook that captures whether Postrun was called on it +type postRun struct{ run bool } + +func (p *postRun) Name() string { return "post" } +func (p *postRun) Postrun() error { + p.run = true + return nil +} + +// Tests that prerun only runs pre run hooks. +func TestAllocRunner_Prerun_Basic(t *testing.T) { + t.Parallel() + require := require.New(t) + ar := testAllocRunner(t) + + // Overwrite the hooks with test hooks + pre := &preRun{} + post := &postRun{} + ar.runnerHooks = []interfaces.RunnerHook{pre, post} + + // Run the hooks + require.NoError(ar.prerun()) + + // Assert only the pre is run + require.True(pre.run) + require.False(post.run) +} + +// Tests that postrun only runs post run hooks. +func TestAllocRunner_Postrun_Basic(t *testing.T) { + t.Parallel() + require := require.New(t) + ar := testAllocRunner(t) + + // Overwrite the hooks with test hooks + pre := &preRun{} + post := &postRun{} + ar.runnerHooks = []interfaces.RunnerHook{pre, post} + + // Run the hooks + require.NoError(ar.postrun()) + + // Assert only the pre is run + require.True(post.run) + require.False(pre.run) +} diff --git a/client/allocrunnerv2/config/config.go b/client/allocrunnerv2/config/config.go new file mode 100644 index 000000000..342d751f1 --- /dev/null +++ b/client/allocrunnerv2/config/config.go @@ -0,0 +1,26 @@ +package config + +import ( + "github.com/boltdb/bolt" + log "github.com/hashicorp/go-hclog" + clientconfig "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/nomad/structs" +) + +// Config holds the configuration for creating an allocation runner. +type Config struct { + // Logger is the logger for the allocation runner. + Logger log.Logger + + // ClientConfig is the clients configuration. + ClientConfig *clientconfig.Config + + // Allocation captures the allocation that should be run. + Allocation *structs.Allocation + + // StateDB is used to store and restore state. + StateDB *bolt.DB + + // XXX Can have a OnStateTransistion hook that we can use to update the + // server +} diff --git a/client/allocrunnerv2/interfaces/factory.go b/client/allocrunnerv2/interfaces/factory.go new file mode 100644 index 000000000..fd9afc96b --- /dev/null +++ b/client/allocrunnerv2/interfaces/factory.go @@ -0,0 +1,10 @@ +package interfaces + +import ( + "context" + + "github.com/hashicorp/nomad/client/allocrunnerv2/config" +) + +// AllocRunnerFactory is the factory method for retrieving an allocation runner. +type AllocRunnerFactory func(ctx context.Context, config *config.Config) (AllocRunner, error) diff --git a/client/allocrunnerv2/interfaces/runner.go b/client/allocrunnerv2/interfaces/runner.go new file mode 100644 index 000000000..a888484f7 --- /dev/null +++ b/client/allocrunnerv2/interfaces/runner.go @@ -0,0 +1,20 @@ +package interfaces + +import "github.com/hashicorp/nomad/client/allocrunnerv2/state" + +// AllocRunner is the interface for an allocation runner. +type AllocRunner interface { + // ID returns the ID of the allocation being run. + ID() string + + // Run starts the runner and begins executing all the tasks as part of the + // allocation. + Run() + + // State returns a copy of the runners state object + State() *state.State +} + +// TaskRunner is the interface for a task runner. +type TaskRunner interface { +} diff --git a/client/allocrunnerv2/interfaces/runner_lifecycle.go b/client/allocrunnerv2/interfaces/runner_lifecycle.go new file mode 100644 index 000000000..5ccece2e8 --- /dev/null +++ b/client/allocrunnerv2/interfaces/runner_lifecycle.go @@ -0,0 +1,37 @@ +package interfaces + +import ( + "github.com/hashicorp/nomad/client/allocrunnerv2/state" +) + +// RunnnerHook is a lifecycle hook into the life cycle of an allocation runner. +type RunnerHook interface { + Name() string +} + +type RunnerPrerunHook interface { + RunnerHook + Prerun() error +} + +type RunnerPostrunHook interface { + RunnerHook + Postrun() error +} + +type RunnerDestroyHook interface { + RunnerHook + Destroy() error +} + +type RunnerUpdateHook interface { + RunnerHook + Update() error +} + +// XXX Not sure yet +type RunnerHookFactory func(target HookTarget) (RunnerHook, error) +type HookTarget interface { + // State retrieves a copy of the target alloc runners state. + State() *state.State +} diff --git a/client/allocrunnerv2/interfaces/task_lifecycle.go b/client/allocrunnerv2/interfaces/task_lifecycle.go new file mode 100644 index 000000000..9dfcbf2b8 --- /dev/null +++ b/client/allocrunnerv2/interfaces/task_lifecycle.go @@ -0,0 +1,92 @@ +package interfaces + +import ( + "github.com/hashicorp/nomad/nomad/structs" +) + +/* + pre-run post-run pre-stop post-stop + | | | | + | | | | + --------> run ------> exited ----------> not restart ---------> garbage collect + +*/ + +// TaskHook is a lifecycle hook into the life cycle of a task runner. +type TaskHook interface { + Name() string +} + +type TaskPrerunRequest struct { + // HookData is previously set data by the hook + HookData map[string]string + + // Task is the task to run + Task *structs.Task + + // Vault token may optionally be set if a Vault token is available + VaultToken string + + // TaskDir is the task's directory on the host + TaskDir string +} + +type TaskPrerunResponse struct { + // Env is the environment variables to set for the task + Env map[string]string + + // HookData allows the hook to emit data to be passed in the next time it is + // run + HookData map[string]string + + // DoOnce lets the hook indicate that it should only be run once + DoOnce bool +} + +type TaskPrerunHook interface { + RunnerHook + Prerun(*TaskPrerunRequest, *TaskPrerunResponse) error +} + +// XXX If we want consul style hooks, need to have something that runs after the +// tasks starts +type TaskPostrunRequest struct { + // Network info +} +type TaskPostrunResponse struct{} + +type TaskPostrunHook interface { + RunnerHook + Postrun() error + //Postrun(*TaskPostrunRequest, *TaskPostrunResponse) error +} + +type TaskPoststopRequest struct{} +type TaskPoststopResponse struct{} + +type TaskPoststopHook interface { + RunnerHook + Postrun() error + //Postrun(*TaskPostrunRequest, *TaskPostrunResponse) error +} + +type TaskDestroyRequest struct{} +type TaskDestroyResponse struct{} + +type TaskDestroyHook interface { + RunnerHook + Destroy() error + //Destroy(*TaskDestroyRequest, *TaskDestroyResponse) error +} + +type TaskUpdateRequest struct { + Alloc string + Vault string // Don't need message bus then +} +type TaskUpdateResponse struct{} + +type TaskUpdateHook interface { + RunnerHook + Update() error + //Update(*TaskUpdateRequest, *TaskUpdateResponse) error +} diff --git a/client/allocrunnerv2/state/state.go b/client/allocrunnerv2/state/state.go new file mode 100644 index 000000000..4105e2656 --- /dev/null +++ b/client/allocrunnerv2/state/state.go @@ -0,0 +1,38 @@ +package state + +import ( + "sync" + + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + "github.com/hashicorp/nomad/nomad/structs" +) + +// State captures the state of the allocation runner. +type State struct { + sync.RWMutex + + // ClientState captures the overall state of the allocation + ClientState string + + // ClientDesc is an optional human readable description of the + // allocations client state + ClientDesc string + + // TaskStates captures the state of individual tasks. + TaskStates map[string]*state.State + + // DeploymentStatus captures the status of the deployment + DeploymentStatus *structs.AllocDeploymentStatus +} + +type PersistentState struct { + ClientState string +} + +func (s *State) PersistentState() *PersistentState { + s.RLock() + defer s.RUnlock() + return &PersistentState{ + ClientState: s.ClientState, + } +} diff --git a/client/allocrunnerv2/taskrunner/state/state.go b/client/allocrunnerv2/taskrunner/state/state.go new file mode 100644 index 000000000..c46ab8ace --- /dev/null +++ b/client/allocrunnerv2/taskrunner/state/state.go @@ -0,0 +1,47 @@ +package state + +import ( + "sync" + + "github.com/hashicorp/nomad/helper" + "github.com/hashicorp/nomad/nomad/structs" +) + +type State struct { + sync.RWMutex + Task *structs.TaskState + Hooks map[string]*HookState + + // VaultToken is the current Vault token for the task + VaultToken string +} + +// Copy should be called with the lock held +func (s *State) Copy() *State { + // Create a copy + c := &State{ + Task: s.Task.Copy(), + Hooks: make(map[string]*HookState, len(s.Hooks)), + VaultToken: s.VaultToken, + } + + // Copy the hooks + for h, state := range s.Hooks { + c.Hooks[h] = state.Copy() + } + + return c +} + +type HookState struct { + SuccessfulOnce bool + Data map[string]string + LastError error +} + +func (h *HookState) Copy() *HookState { + c := new(HookState) + *c = *h + c.Data = helper.CopyMapStringString(c.Data) + return c +} diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go new file mode 100644 index 000000000..426c3967b --- /dev/null +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -0,0 +1,425 @@ +package taskrunner + +import ( + "context" + "fmt" + "sync" + "time" + + metrics "github.com/armon/go-metrics" + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocdir" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" + "github.com/hashicorp/nomad/client/allocrunnerv2/config" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + arstate "github.com/hashicorp/nomad/client/allocrunnerv2/state" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + "github.com/hashicorp/nomad/client/driver" + "github.com/hashicorp/nomad/client/driver/env" + dstructs "github.com/hashicorp/nomad/client/driver/structs" + "github.com/hashicorp/nomad/nomad/structs" +) + +type allocRunner interface { + State() *arstate.State // XXX Do we need it + Alloc() *structs.Allocation + ID() string + Config() *config.Config + GetAllocDir() *allocdir.AllocDir + StateUpdated(*state.State) error +} + +type TaskRunner struct { + // config is the task runners configuration. + config *Config + + // state captures the state of the task runner + state *state.State + + // ctx is the alloc runners context. It is cancelled when all related + // activity for this allocation should be terminated. + ctx context.Context + + // ctxCancel is used to cancel the alloc runners cancel + ctxCancel context.CancelFunc + + // Logger is the logger for the task runner. + logger log.Logger + + // allocRunner is the parent allocRunner + allocRunner allocRunner + + // waitCh is closed when the task runner has transitioned to a terminal + // state + waitCh chan struct{} + + // driver is the driver for the task. + driver driver.Driver + + // handle is the handle to the currently running driver + handle driver.DriverHandle + handleLock sync.Mutex + + // task is the task beign run + task *structs.Task + taskLock sync.RWMutex + + // taskDir is the directory structure for this task. + taskDir *allocdir.TaskDir + + // envBuilder is used to build the task's environment + envBuilder *env.Builder + + // restartTracker is used to decide if the task should be restarted. + restartTracker *restarts.RestartTracker + + // runnerHooks are task runner lifecycle hooks that should be run on state + // transistions. + runnerHooks []interfaces.TaskHook + + // baseLabels are used when emitting tagged metrics. All task runner metrics + // will have these tags, and optionally more. + baseLabels []metrics.Label +} + +type Config struct { + Parent allocRunner + Task *structs.Task + Logger log.Logger + + // State is optionally restored task state + State *state.State +} + +func NewTaskRunner(config *Config) (*TaskRunner, error) { + // Create a context for the runner + trCtx, trCancel := context.WithCancel(context.Background()) + + // Initialize the environment builder + envBuilder := env.NewBuilder( + config.Parent.Config().ClientConfig.Node, + config.Parent.Alloc(), + config.Task, + config.Parent.Config().ClientConfig.Region) + + tr := &TaskRunner{ + config: config, + allocRunner: config.Parent, + state: config.State, + ctx: trCtx, + ctxCancel: trCancel, + task: config.Task, + taskDir: config.Parent.GetAllocDir().NewTaskDir(config.Task.Name), + envBuilder: envBuilder, + waitCh: make(chan struct{}), + } + + // Create the logger based on the allocation ID + tr.logger = config.Logger.Named("task_runner").With("task", config.Task.Name) + + // Build the restart tracker. + alloc := tr.allocRunner.Alloc() + tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup) + if tg == nil { + tr.logger.Error("alloc missing task group") + return nil, fmt.Errorf("alloc missing task group") + } + tr.restartTracker = restarts.NewRestartTracker(tg.RestartPolicy, alloc.Job.Type) + + // Initialize the task state + tr.initState() + + // Get the driver + if err := tr.initDriver(); err != nil { + tr.logger.Error("failed to create driver", "error", err) + return nil, err + } + + // Initialize the runners hooks. + tr.initHooks() + + // Initialize base labels + tr.initLabels() + + return tr, nil +} + +func (tr *TaskRunner) initState() { + if tr.state == nil { + tr.state = &state.State{ + Task: &structs.TaskState{ + State: structs.TaskStatePending, + }, + Hooks: make(map[string]*state.HookState), + } + } +} + +func (tr *TaskRunner) initLabels() { + alloc := tr.allocRunner.Alloc() + tr.baseLabels = []metrics.Label{ + { + Name: "job", + Value: alloc.Job.Name, + }, + { + Name: "task_group", + Value: alloc.TaskGroup, + }, + { + Name: "alloc_id", + Value: alloc.ID, + }, + { + Name: "task", + Value: tr.Name(), + }, + } +} + +func (tr *TaskRunner) WaitCh() <-chan struct{} { + return tr.waitCh +} + +func (tr *TaskRunner) Update() { + // XXX +} + +func (tr *TaskRunner) Stop() { + // XXX + tr.ctxCancel() +} + +func (tr *TaskRunner) Run() { + defer close(tr.waitCh) + + var err error + var restart bool + var restartWait time.Duration + var waitRes *dstructs.WaitResult + for { + // Run the prerun hooks + if err = tr.prerun(); err != nil { + tr.logger.Error("prerun failed", "error", err) + tr.restartTracker.SetStartError(err) + goto RESTART + } + + // Run and wait + waitRes, err = tr.runDriver() + if err != nil { + tr.logger.Error("running driver failed", "error", err) + tr.restartTracker.SetStartError(err) + goto RESTART + } + tr.restartTracker.SetWaitResult(waitRes) + + // Run the postrun hooks + if err = tr.postrun(); err != nil { + tr.logger.Error("postrun failed", "error", err) + } + + // Check if the context is closed already and go straight to destroy + if err := tr.ctx.Err(); err != nil { + goto DESTROY + } + + RESTART: + // Actually restart by sleeping and also watching for destroy events + restart, restartWait = tr.shouldRestart() + if restart { + select { + case <-time.After(restartWait): + continue + case <-tr.ctx.Done(): + } + } + + DESTROY: + // Run the destroy hooks + if err = tr.destroy(); err != nil { + tr.logger.Error("postrun failed", "error", err) + } + + tr.logger.Debug("task run loop exiting") + return + } +} + +func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { + // Determine if we should restart + state, when := tr.restartTracker.GetState() + reason := tr.restartTracker.GetReason() + switch state { + case structs.TaskNotRestarting, structs.TaskTerminated: + tr.logger.Info("not restarting task", "reason", reason) + if state == structs.TaskNotRestarting { + tr.setState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskNotRestarting).SetRestartReason(reason).SetFailsTask()) + } + return false, 0 + case structs.TaskRestarting: + tr.logger.Info("restarting task", "reason", reason, "delay", when) + tr.setState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskRestarting).SetRestartDelay(when).SetRestartReason(reason)) + return true, 0 + default: + tr.logger.Error("restart tracker returned unknown state", "state", state) + return true, when + } +} + +// runDriver runs the driver and waits for it to exit +func (tr *TaskRunner) runDriver() (*dstructs.WaitResult, error) { + // Run prestart + ctx := driver.NewExecContext(tr.taskDir, tr.envBuilder.Build()) + _, err := tr.driver.Prestart(ctx, tr.task) + if err != nil { + tr.logger.Error("driver pre-start failed", "error", err) + return nil, err + } + + // Create a new context for Start since the environment may have been updated. + ctx = driver.NewExecContext(tr.taskDir, tr.envBuilder.Build()) + + // Start the job + sresp, err := tr.driver.Start(ctx, tr.task) + if err != nil { + tr.logger.Warn("driver start failed", "error", err) + return nil, err + } + + // Wait on the handle + tr.handleLock.Lock() + handle := sresp.Handle + tr.handle = handle + tr.handleLock.Unlock() + + // Emit an event that we started + tr.setState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) + + // Wait for the task to exit + waitRes := <-handle.WaitCh() + return waitRes, nil +} + +// initDriver creates the driver for the task +func (tr *TaskRunner) initDriver() error { + // Create a task-specific event emitter callback to expose minimal + // state to drivers + eventEmitter := func(m string, args ...interface{}) { + msg := fmt.Sprintf(m, args...) + tr.logger.Debug("driver event", "event", msg) + tr.setState("", structs.NewTaskEvent(structs.TaskDriverMessage).SetDriverMessage(msg)) + } + + alloc := tr.allocRunner.Alloc() + driverCtx := driver.NewDriverContext( + alloc.Job.Name, + alloc.TaskGroup, + tr.Name(), + tr.allocRunner.ID(), + tr.allocRunner.Config().ClientConfig, // XXX Why does it need this + tr.allocRunner.Config().ClientConfig.Node, // XXX THIS I NEED TO FIX + tr.logger.StandardLogger(nil), // XXX Should pass this through + eventEmitter) + + driver, err := driver.NewDriver(tr.task.Driver, driverCtx) + if err != nil { + return err + } + + tr.driver = driver + return nil +} + +func (tr *TaskRunner) setState(state string, event *structs.TaskEvent) { + // Ensure the event is populated with human readable strings + event.PopulateEventDisplayMessage() + + // Lock our state + tr.state.Lock() + defer tr.state.Unlock() + task := tr.state.Task + + // Update the state of the task + if state != "" { + task.State = state + } + + // Handle the event + if event == nil { + if event.FailsTask { + task.Failed = true + } + + if event.Type == structs.TaskRestarting { + if !tr.allocRunner.Config().ClientConfig.DisableTaggedMetrics { + metrics.IncrCounterWithLabels([]string{"client", "allocs", "restart"}, 1, tr.baseLabels) + } + //if r.config.BackwardsCompatibleMetrics { + //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "restart"}, 1) + //} + task.Restarts++ + task.LastRestart = time.Unix(0, event.Time) + } + appendTaskEvent(task, event) + } + + // Handle the state transistion. + switch state { + case structs.TaskStateRunning: + // Capture the start time if it is just starting + if task.State != structs.TaskStateRunning { + task.StartedAt = time.Now().UTC() + if !tr.allocRunner.Config().ClientConfig.DisableTaggedMetrics { + metrics.IncrCounterWithLabels([]string{"client", "allocs", "running"}, 1, tr.baseLabels) + } + //if r.config.BackwardsCompatibleMetrics { + //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "running"}, 1) + //} + } + case structs.TaskStateDead: + // Capture the finished time if not already set + if task.FinishedAt.IsZero() { + task.FinishedAt = time.Now().UTC() + } + + // Emitting metrics to indicate task complete and failures + if task.Failed { + if !tr.allocRunner.Config().ClientConfig.DisableTaggedMetrics { + metrics.IncrCounterWithLabels([]string{"client", "allocs", "failed"}, 1, tr.baseLabels) + } + //if r.config.BackwardsCompatibleMetrics { + //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "failed"}, 1) + //} + } else { + if !tr.allocRunner.Config().ClientConfig.DisableTaggedMetrics { + metrics.IncrCounterWithLabels([]string{"client", "allocs", "complete"}, 1, tr.baseLabels) + } + //if r.config.BackwardsCompatibleMetrics { + //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "complete"}, 1) + //} + } + } + + // Create a copy and notify the alloc runner of the transition + if err := tr.allocRunner.StateUpdated(tr.state.Copy()); err != nil { + tr.logger.Error("failed to save state", "error", err) + } +} + +// appendTaskEvent updates the task status by appending the new event. +func appendTaskEvent(state *structs.TaskState, event *structs.TaskEvent) { + capacity := 10 + if state.Events == nil { + state.Events = make([]*structs.TaskEvent, 0, capacity) + } + + // If we hit capacity, then shift it. + if len(state.Events) == capacity { + old := state.Events + state.Events = make([]*structs.TaskEvent, 0, capacity) + state.Events = append(state.Events, old[1:]...) + } + + state.Events = append(state.Events, event) +} diff --git a/client/allocrunnerv2/taskrunner/task_runner_getters.go b/client/allocrunnerv2/taskrunner/task_runner_getters.go new file mode 100644 index 000000000..7c8e8c679 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/task_runner_getters.go @@ -0,0 +1,15 @@ +package taskrunner + +import "github.com/hashicorp/nomad/nomad/structs" + +func (tr *TaskRunner) Name() string { + tr.taskLock.RLock() + defer tr.taskLock.RUnlock() + return tr.task.Name +} + +func (tr *TaskRunner) Task() *structs.Task { + tr.taskLock.RLock() + defer tr.taskLock.RUnlock() + return tr.task +} diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go new file mode 100644 index 000000000..ac92b0b9a --- /dev/null +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -0,0 +1,285 @@ +package taskrunner + +import ( + "fmt" + "time" + + log "github.com/hashicorp/go-hclog" + + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + cconfig "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/driver" + "github.com/hashicorp/nomad/nomad/structs" +) + +// initHooks intializes the tasks hooks. +func (tr *TaskRunner) initHooks() { + hookLogger := tr.logger.Named("task_hook") + tr.runnerHooks = make([]interfaces.TaskHook, 0, 3) + + // Create the task directory hook. This is run first to ensure the + // directoy path exists for other hooks. + tr.runnerHooks = append(tr.runnerHooks, newTaskDirHook(tr, hookLogger)) +} + +// prerun is used to run the runners prerun hooks. +func (tr *TaskRunner) prerun() error { + // Determine if the allocation is terminaland we should avoid running + // pre-run hooks. + alloc := tr.allocRunner.Alloc() + if alloc.TerminalStatus() { + tr.logger.Trace("skipping pre-run hooks since allocation is terminal") + return nil + } + + if tr.logger.IsTrace() { + start := time.Now() + tr.logger.Trace("running pre-run hooks", "start", start) + defer func() { + end := time.Now() + tr.logger.Trace("finished pre-run hooks", "end", end, "duration", end.Sub(start)) + }() + } + + for _, hook := range tr.runnerHooks { + pre, ok := hook.(interfaces.TaskPrerunHook) + if !ok { + continue + } + + name := pre.Name() + var start time.Time + if tr.logger.IsTrace() { + start = time.Now() + tr.logger.Trace("running pre-run hook", "name", name, "start", start) + } + + // Build the request + req := interfaces.TaskPrerunRequest{ + Task: tr.Task(), + TaskDir: tr.taskDir.Dir, + } + + tr.state.RLock() + hookState := tr.state.Hooks[name] + if hookState != nil { + req.HookData = hookState.Data + } + + req.VaultToken = tr.state.VaultToken + tr.state.RUnlock() + + if hookState.SuccessfulOnce { + tr.logger.Trace("skipping hook since it was successfully run once", "name", name) + continue + } + + // Run the pre-run hook + var resp interfaces.TaskPrerunResponse + err := pre.Prerun(&req, &resp) + + // Store the hook state + { + tr.state.Lock() + hookState, ok := tr.state.Hooks[name] + if !ok { + hookState = &state.HookState{} + tr.state.Hooks[name] = hookState + } + + hookState.LastError = err + if resp.HookData != nil { + hookState.Data = resp.HookData + } + + if resp.DoOnce && err != nil { + hookState.SuccessfulOnce = true + } + + // XXX Detect if state has changed so that we can signal to the + // alloc runner precisly + if err := tr.allocRunner.StateUpdated(tr.state.Copy()); err != nil { + tr.logger.Error("failed to save state", "error", err) + } + tr.state.Unlock() + } + + if err != nil { + return structs.WrapRecoverable(fmt.Sprintf("pre-run hook %q failed: %v", name, err), err) + } + + // Store the environment variables returned by the hook + if len(resp.Env) != 0 { + tr.envBuilder.SetGenericEnv(resp.Env) + } + + if tr.logger.IsTrace() { + end := time.Now() + tr.logger.Trace("finished pre-run hooks", "name", name, "end", end, "duration", end.Sub(start)) + } + } + + return nil +} + +// postrun is used to run the runners postrun hooks. +func (tr *TaskRunner) postrun() error { + if tr.logger.IsTrace() { + start := time.Now() + tr.logger.Trace("running post-run hooks", "start", start) + defer func() { + end := time.Now() + tr.logger.Trace("finished post-run hooks", "end", end, "duration", end.Sub(start)) + }() + } + + for _, hook := range tr.runnerHooks { + post, ok := hook.(interfaces.TaskPostrunHook) + if !ok { + continue + } + + name := post.Name() + var start time.Time + if tr.logger.IsTrace() { + start = time.Now() + tr.logger.Trace("running post-run hook", "name", name, "start", start) + } + + // XXX We shouldn't exit on the first one + if err := post.Postrun(); err != nil { + return fmt.Errorf("post-run hook %q failed: %v", name, err) + } + + if tr.logger.IsTrace() { + end := time.Now() + tr.logger.Trace("finished post-run hooks", "name", name, "end", end, "duration", end.Sub(start)) + } + } + + return nil +} + +// destroy is used to run the runners destroy hooks. +// XXX Naming change +func (tr *TaskRunner) destroy() error { + if tr.logger.IsTrace() { + start := time.Now() + tr.logger.Trace("running destroy hooks", "start", start) + defer func() { + end := time.Now() + tr.logger.Trace("finished destroy hooks", "end", end, "duration", end.Sub(start)) + }() + } + + for _, hook := range tr.runnerHooks { + post, ok := hook.(interfaces.TaskDestroyHook) + if !ok { + continue + } + + name := post.Name() + var start time.Time + if tr.logger.IsTrace() { + start = time.Now() + tr.logger.Trace("running destroy hook", "name", name, "start", start) + } + + // XXX We shouldn't exit on the first one + if err := post.Destroy(); err != nil { + return fmt.Errorf("destroy hook %q failed: %v", name, err) + } + + if tr.logger.IsTrace() { + end := time.Now() + tr.logger.Trace("finished destroy hooks", "name", name, "end", end, "duration", end.Sub(start)) + } + } + + return nil +} + +type taskDirHook struct { + runner *TaskRunner + logger log.Logger +} + +func newTaskDirHook(runner *TaskRunner, logger log.Logger) *taskDirHook { + td := &taskDirHook{ + runner: runner, + } + td.logger = logger.Named(td.Name()) + return td +} + +func (h *taskDirHook) Name() string { + return "task_dir" +} + +func (h *taskDirHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { + cc := h.runner.allocRunner.Config().ClientConfig + chroot := cconfig.DefaultChrootEnv + if len(cc.ChrootEnv) > 0 { + chroot = cc.ChrootEnv + } + + // Emit the event that we are going to be building the task directory + h.runner.setState("", structs.NewTaskEvent(structs.TaskSetup).SetMessage(structs.TaskBuildingTaskDir)) + + // Build the task directory structure + fsi := h.runner.driver.FSIsolation() + err := h.runner.taskDir.Build(false, chroot, fsi) + if err != nil { + return err + } + + // Update the environment variables based on the built task directory + driver.SetEnvvars(h.runner.envBuilder, fsi, h.runner.taskDir, h.runner.allocRunner.Config().ClientConfig) + return nil +} + +/* +TR Hooks: + +> @schmichael +Task Validate: +Require: Client config, task definiton +Return: error +Implement: Prestart + +> DONE +Task Dir Build: +Requires: Folder structure, driver isolation, client config +Return env, error +Implement: Prestart + +> @alex +Vault: Task, RPC to talk to server to derive token, Node SecretID +Return vault token (Call a setter), error, env +Implement: Prestart + +> @alex +Consul Template: +Require: Task, alloc directory, way to signal/restart task, updates when vault token changes +Return env, error +Implement: Prestart and Update (for new Vault token) and Destroy + +> @schmichael +Consul Service Reg: +Require: Task, interpolation/ENV +Return: error +Implement: Postrun, Update, Prestop + +> @alex +Dispatch Payload: +Require: Alloc +Return error +Implement: Prerun + +> @schmichael +Artifacts: +Require: Folder structure, task, interpolation/ENV +Return: error +Implement: Prerun and Destroy +*/ diff --git a/client/allocrunnerv2/tr_shim.go b/client/allocrunnerv2/tr_shim.go new file mode 100644 index 000000000..7cab963c4 --- /dev/null +++ b/client/allocrunnerv2/tr_shim.go @@ -0,0 +1,31 @@ +package allocrunnerv2 + +import ( + "github.com/hashicorp/nomad/client/allocdir" + "github.com/hashicorp/nomad/client/allocrunnerv2/config" + "github.com/hashicorp/nomad/client/allocrunnerv2/state" + trstate "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" +) + +// allocRunnerShim is used to pass the alloc runner to the task runner, exposing +// private methods as public fields. +type allocRunnerShim struct { + *allocRunner +} + +func (a *allocRunnerShim) State() *state.State { + return a.state +} + +func (a *allocRunnerShim) Config() *config.Config { + return a.config +} + +func (a *allocRunnerShim) GetAllocDir() *allocdir.AllocDir { + return a.allocDir +} + +func (a *allocRunnerShim) StateUpdated(trState *trstate.State) error { + // XXX implement + return nil +} diff --git a/client/allocrunnerv2/util.go b/client/allocrunnerv2/util.go new file mode 100644 index 000000000..b833b6c2d --- /dev/null +++ b/client/allocrunnerv2/util.go @@ -0,0 +1 @@ +package allocrunnerv2 diff --git a/client/driver/env/env.go b/client/driver/env/env.go index 63c245504..ecd4ea2c8 100644 --- a/client/driver/env/env.go +++ b/client/driver/env/env.go @@ -362,6 +362,16 @@ func (b *Builder) UpdateTask(alloc *structs.Allocation, task *structs.Task) *Bui return b.setTask(task).setAlloc(alloc) } +func (b *Builder) SetGenericEnv(envs map[string]string) *Builder { + b.mu.Lock() + defer b.mu.Unlock() + for k, v := range envs { + b.envvars[k] = v + } + + return b +} + // setTask is called from NewBuilder to populate task related environment // variables. func (b *Builder) setTask(task *structs.Task) *Builder { From eae54e2954ccfb9a8fd654c328e222124b7861d5 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 27 Jun 2018 16:57:31 -0700 Subject: [PATCH 002/152] artifact task hook --- client/allocrunner/taskrunner/task_runner.go | 2 + .../interfaces/task_lifecycle.go | 14 +++-- .../allocrunnerv2/taskrunner/task_runner.go | 11 ++-- .../taskrunner/task_runner_hooks.go | 54 ++++++++++++++++++- 4 files changed, 69 insertions(+), 12 deletions(-) diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index 1c6c7ef84..469f69109 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -647,6 +647,8 @@ func (r *TaskRunner) validateTask() error { } } + //XXX Is this the right place for this? Seems like it could be done on + // the server when the job is submitted. Is this defense in depth? // Validate the artifacts for i, artifact := range r.task.Artifacts { // Verify the artifact doesn't escape the task directory. diff --git a/client/allocrunnerv2/interfaces/task_lifecycle.go b/client/allocrunnerv2/interfaces/task_lifecycle.go index 9dfcbf2b8..d5e1c28a9 100644 --- a/client/allocrunnerv2/interfaces/task_lifecycle.go +++ b/client/allocrunnerv2/interfaces/task_lifecycle.go @@ -1,6 +1,7 @@ package interfaces import ( + "github.com/hashicorp/nomad/client/driver/env" "github.com/hashicorp/nomad/nomad/structs" ) @@ -29,6 +30,9 @@ type TaskPrerunRequest struct { // TaskDir is the task's directory on the host TaskDir string + + // TaskEnv is the task's environment + TaskEnv *env.TaskEnv } type TaskPrerunResponse struct { @@ -44,7 +48,7 @@ type TaskPrerunResponse struct { } type TaskPrerunHook interface { - RunnerHook + TaskHook Prerun(*TaskPrerunRequest, *TaskPrerunResponse) error } @@ -56,7 +60,7 @@ type TaskPostrunRequest struct { type TaskPostrunResponse struct{} type TaskPostrunHook interface { - RunnerHook + TaskHook Postrun() error //Postrun(*TaskPostrunRequest, *TaskPostrunResponse) error } @@ -65,7 +69,7 @@ type TaskPoststopRequest struct{} type TaskPoststopResponse struct{} type TaskPoststopHook interface { - RunnerHook + TaskHook Postrun() error //Postrun(*TaskPostrunRequest, *TaskPostrunResponse) error } @@ -74,7 +78,7 @@ type TaskDestroyRequest struct{} type TaskDestroyResponse struct{} type TaskDestroyHook interface { - RunnerHook + TaskHook Destroy() error //Destroy(*TaskDestroyRequest, *TaskDestroyResponse) error } @@ -86,7 +90,7 @@ type TaskUpdateRequest struct { type TaskUpdateResponse struct{} type TaskUpdateHook interface { - RunnerHook + TaskHook Update() error //Update(*TaskUpdateRequest, *TaskUpdateResponse) error } diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 426c3967b..aae5ebc86 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -254,12 +254,12 @@ func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { case structs.TaskNotRestarting, structs.TaskTerminated: tr.logger.Info("not restarting task", "reason", reason) if state == structs.TaskNotRestarting { - tr.setState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskNotRestarting).SetRestartReason(reason).SetFailsTask()) + tr.SetState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskNotRestarting).SetRestartReason(reason).SetFailsTask()) } return false, 0 case structs.TaskRestarting: tr.logger.Info("restarting task", "reason", reason, "delay", when) - tr.setState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskRestarting).SetRestartDelay(when).SetRestartReason(reason)) + tr.SetState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskRestarting).SetRestartDelay(when).SetRestartReason(reason)) return true, 0 default: tr.logger.Error("restart tracker returned unknown state", "state", state) @@ -294,7 +294,7 @@ func (tr *TaskRunner) runDriver() (*dstructs.WaitResult, error) { tr.handleLock.Unlock() // Emit an event that we started - tr.setState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) + tr.SetState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) // Wait for the task to exit waitRes := <-handle.WaitCh() @@ -308,7 +308,7 @@ func (tr *TaskRunner) initDriver() error { eventEmitter := func(m string, args ...interface{}) { msg := fmt.Sprintf(m, args...) tr.logger.Debug("driver event", "event", msg) - tr.setState("", structs.NewTaskEvent(structs.TaskDriverMessage).SetDriverMessage(msg)) + tr.SetState("", structs.NewTaskEvent(structs.TaskDriverMessage).SetDriverMessage(msg)) } alloc := tr.allocRunner.Alloc() @@ -331,7 +331,8 @@ func (tr *TaskRunner) initDriver() error { return nil } -func (tr *TaskRunner) setState(state string, event *structs.TaskEvent) { +// SetState sets the task runners state. +func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { // Ensure the event is populated with human readable strings event.PopulateEventDisplayMessage() diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index ac92b0b9a..169403a19 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -6,6 +6,7 @@ import ( log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocrunner/getter" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" cconfig "github.com/hashicorp/nomad/client/config" @@ -20,7 +21,11 @@ func (tr *TaskRunner) initHooks() { // Create the task directory hook. This is run first to ensure the // directoy path exists for other hooks. - tr.runnerHooks = append(tr.runnerHooks, newTaskDirHook(tr, hookLogger)) + tr.runnerHooks = []interfaces.TaskHook{ + tr.runnerHooks, + newTaskDirHook(tr, hookLogger), + newArtifactHook(tr, hookLogger), + } } // prerun is used to run the runners prerun hooks. @@ -59,6 +64,7 @@ func (tr *TaskRunner) prerun() error { req := interfaces.TaskPrerunRequest{ Task: tr.Task(), TaskDir: tr.taskDir.Dir, + TaskEnv: tr.envBuilder.Build(), } tr.state.RLock() @@ -70,6 +76,9 @@ func (tr *TaskRunner) prerun() error { req.VaultToken = tr.state.VaultToken tr.state.RUnlock() + //XXX Can we assume everything only wants to be run until + //successful and simply keep track of which hooks have yet to + //run on failures+retries? if hookState.SuccessfulOnce { tr.logger.Trace("skipping hook since it was successfully run once", "name", name) continue @@ -225,7 +234,7 @@ func (h *taskDirHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interfaces } // Emit the event that we are going to be building the task directory - h.runner.setState("", structs.NewTaskEvent(structs.TaskSetup).SetMessage(structs.TaskBuildingTaskDir)) + h.runner.SetState("", structs.NewTaskEvent(structs.TaskSetup).SetMessage(structs.TaskBuildingTaskDir)) // Build the task directory structure fsi := h.runner.driver.FSIsolation() @@ -239,6 +248,47 @@ func (h *taskDirHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interfaces return nil } +type EventEmitter interface { + SetState(state string, event *structs.TaskEvent) +} + +// artifactHook downloads artifacts for a task. +type artifactHook struct { + eventEmitter EventEmitter + logger log.Logger +} + +func newArtifactHook(e EventEmitter, logger log.Logger) *artifactHook { + h := &artifactHook{ + eventEmitter: e, + } + h.logger = logger.Named(h.Name()) + return h +} + +func (*artifactHook) Name() string { + return "artifacts" +} + +func (h *artifactHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { + h.eventEmitter.SetState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskDownloadingArtifacts)) + + for _, artifact := range req.Task.Artifacts { + if err := getter.GetArtifact(req.TaskEnv, artifact, req.TaskDir); err != nil { + wrapped := fmt.Errorf("failed to download artifact %q: %v", artifact.GetterSource, err) + h.logger.Debug(wrapped.Error()) + h.eventEmitter.SetState(structs.TaskStatePending, + structs.NewTaskEvent(structs.TaskArtifactDownloadFailed).SetDownloadError(wrapped)) + return wrapped + } + } + + //XXX Should this be managed by task runner directly? Seems silly to + //make every hook specify it + resp.DoOnce = true + return nil +} + /* TR Hooks: From 9a6aa38b0ff1b3e0578dc707e8b1a7716ceec3ce Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 27 Jun 2018 17:27:03 -0700 Subject: [PATCH 003/152] begin adding AllocRunner.Update --- client/allocrunnerv2/alloc_runner.go | 85 +++++++++++-------- client/allocrunnerv2/alloc_runner_hooks.go | 2 +- .../allocrunnerv2/taskrunner/task_runner.go | 7 +- .../taskrunner/task_runner_hooks.go | 2 - 4 files changed, 55 insertions(+), 41 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 8eaa4112e..0da2a5faa 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -1,7 +1,6 @@ package allocrunnerv2 import ( - "context" "fmt" "sync" @@ -16,13 +15,6 @@ import ( // allocRunner is used to run all the tasks in a given allocation type allocRunner struct { - // ctx is the alloc runners context. It is cancelled when all related - // activity for this allocation should be terminated. - ctx context.Context - - // ctxCancel is used to cancel the alloc runners cancel - ctxCancel context.CancelFunc - // Logger is the logger for the alloc runner. logger log.Logger @@ -48,21 +40,19 @@ type allocRunner struct { runnerHooks []interfaces.RunnerHook // tasks are the set of task runners - tasks map[string]*taskrunner.TaskRunner - taskLock sync.Mutex + tasks map[string]*taskrunner.TaskRunner + + // updateCh receives allocation updates + updateCh chan *structs.Allocation } // NewAllocRunner returns a new allocation runner. -func NewAllocRunner(ctx context.Context, config *config.Config) (*allocRunner, error) { - // Create a context for the runner - arCtx, arCancel := context.WithCancel(ctx) - +func NewAllocRunner(config *config.Config) (*allocRunner, error) { ar := &allocRunner{ - ctx: arCtx, - ctxCancel: arCancel, - config: config, - alloc: config.Allocation, - waitCh: make(chan struct{}), + config: config, + alloc: config.Allocation, + waitCh: make(chan struct{}), + updateCh: make(chan *structs.Allocation), } // Create the logger based on the allocation ID @@ -84,6 +74,9 @@ func (ar *allocRunner) Run() { // Close the wait channel defer close(ar.waitCh) + var err error + var taskWaitCh <-chan struct{} + // Run the prerun hooks // XXX Equivalent to TR.Prerun hook if err := ar.prerun(); err != nil { @@ -92,10 +85,25 @@ func (ar *allocRunner) Run() { } // Run the runners - if err := ar.runImpl(); err != nil { + taskWaitCh, err = ar.runImpl() + if err != nil { ar.logger.Error("starting tasks failed", "error", err) } + for { + select { + case <-taskWaitCh: + // TaskRunners have all exited + case updated := <-ar.updateCh: + // Updated alloc received + //XXX Update hooks + //XXX Update ar.alloc + for _, tr := range ar.tasks { + tr.Update(updated) + } + } + } + POST: // Run the postrun hooks // XXX Equivalent to TR.Poststop hook @@ -105,35 +113,32 @@ POST: } // runImpl is used to run the runners. -func (ar *allocRunner) runImpl() error { +func (ar *allocRunner) runImpl() (<-chan struct{}, error) { // Grab the task group tg := ar.alloc.Job.LookupTaskGroup(ar.alloc.TaskGroup) if tg == nil { // XXX Fail and exit ar.logger.Error("failed to lookup task group", "task_group", ar.alloc.TaskGroup) - return fmt.Errorf("failed to lookup task group %q", ar.alloc.TaskGroup) + return nil, fmt.Errorf("failed to lookup task group %q", ar.alloc.TaskGroup) } for _, task := range tg.Tasks { if err := ar.runTask(task); err != nil { - return err + return nil, err } } - // Block until all tasks are done. - var wg sync.WaitGroup - ar.taskLock.Lock() - for _, task := range ar.tasks { - wg.Add(1) - go func() { + // Return a combined WaitCh that is closed when all task runners have + // exited. + waitCh := make(chan struct{}) + go func() { + defer close(waitCh) + for _, task := range ar.tasks { <-task.WaitCh() - wg.Done() - }() - } - ar.taskLock.Unlock() + } + }() - wg.Wait() - return nil + return waitCh, nil } // runTask is used to run a task. @@ -153,8 +158,14 @@ func (ar *allocRunner) runTask(task *structs.Task) error { go tr.Run() // Store the runner - ar.taskLock.Lock() ar.tasks[task.Name] = tr - ar.taskLock.Unlock() return nil } + +// Update the running allocation with a new version received from the server. +// +// This method is safe for calling concurrently with Run() and does not modify +// the passed in allocation. +func (ar *allocRunner) Update(update *structs.Allocation) { + ar.updateCh <- update +} diff --git a/client/allocrunnerv2/alloc_runner_hooks.go b/client/allocrunnerv2/alloc_runner_hooks.go index 085020813..e39d3fbdc 100644 --- a/client/allocrunnerv2/alloc_runner_hooks.go +++ b/client/allocrunnerv2/alloc_runner_hooks.go @@ -143,7 +143,7 @@ type allocHealthWatcherHook struct { } func newAllocHealthWatcherHook(runner *allocRunner, logger log.Logger) *allocHealthWatcherHook { - ctx, cancelFn := context.WithCancel(runner.ctx) + ctx, cancelFn := context.WithCancel(context.Background()) ad := &allocHealthWatcherHook{ runner: runner, ctx: ctx, diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index aae5ebc86..28b47a58c 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -177,11 +177,16 @@ func (tr *TaskRunner) initLabels() { } } +// WaitCh is closed when TaskRunner.Run exits. func (tr *TaskRunner) WaitCh() <-chan struct{} { return tr.waitCh } -func (tr *TaskRunner) Update() { +// Update the running allocation with a new version received from the server. +// +// This method is safe for calling concurrently with Run() and does not modify +// the passed in allocation. +func (tr *TaskRunner) Update(update *structs.Allocation) { // XXX } diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 169403a19..4f050a23a 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -17,12 +17,10 @@ import ( // initHooks intializes the tasks hooks. func (tr *TaskRunner) initHooks() { hookLogger := tr.logger.Named("task_hook") - tr.runnerHooks = make([]interfaces.TaskHook, 0, 3) // Create the task directory hook. This is run first to ensure the // directoy path exists for other hooks. tr.runnerHooks = []interfaces.TaskHook{ - tr.runnerHooks, newTaskDirHook(tr, hookLogger), newArtifactHook(tr, hookLogger), } From 0f7dcfdc9ab38e5e20c0fb396787fb3694a0e5d8 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 28 Jun 2018 17:01:05 -0700 Subject: [PATCH 004/152] example redis job "runs" on arv2! see below Tons left to do and lots of churn: 1. No state saving 2. No shutdown or gc 3. Removed AR factory *for now* 4. Made all "Config" structs local to the package they configure 5. Added allocID to GC to avoid a lookup Really hating how many things use *structs.Allocation. It's not bad without state saving, but if AllocRunner starts updating its copy things get racy fast. --- client/allocrunnerv2/alloc_runner.go | 117 +++++++++++++--- client/allocrunnerv2/alloc_runner_getters.go | 15 --- client/allocrunnerv2/{config => }/config.go | 6 +- client/allocrunnerv2/interfaces/factory.go | 10 -- .../allocrunnerv2/taskrunner/task_runner.go | 65 +++++---- .../taskrunner/task_runner_hooks.go | 16 ++- client/allocrunnerv2/tr_shim.go | 5 - client/client.go | 125 +++++++++++------- client/gc.go | 41 +++--- client/util.go | 32 ++--- 10 files changed, 248 insertions(+), 184 deletions(-) delete mode 100644 client/allocrunnerv2/alloc_runner_getters.go rename client/allocrunnerv2/{config => }/config.go (84%) delete mode 100644 client/allocrunnerv2/interfaces/factory.go diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 0da2a5faa..60a6fce30 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -2,14 +2,17 @@ package allocrunnerv2 import ( "fmt" + "path/filepath" "sync" log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunnerv2/config" + "github.com/hashicorp/nomad/client/allocrunner" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" "github.com/hashicorp/nomad/client/allocrunnerv2/state" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" + "github.com/hashicorp/nomad/client/config" + cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" ) @@ -18,8 +21,7 @@ type allocRunner struct { // Logger is the logger for the alloc runner. logger log.Logger - // Config is the configuration for the alloc runner. - config *config.Config + clientConfig *config.Config // waitCh is closed when the alloc runner has transitioned to a terminal // state @@ -27,7 +29,7 @@ type allocRunner struct { // Alloc captures the allocation being run. alloc *structs.Allocation - allocLock sync.Mutex + allocLock sync.RWMutex // state captures the state of the alloc runner state *state.State @@ -42,26 +44,31 @@ type allocRunner struct { // tasks are the set of task runners tasks map[string]*taskrunner.TaskRunner - // updateCh receives allocation updates + // updateCh receives allocation updates via the Update method updateCh chan *structs.Allocation } // NewAllocRunner returns a new allocation runner. -func NewAllocRunner(config *config.Config) (*allocRunner, error) { +func NewAllocRunner(config *Config) *allocRunner { ar := &allocRunner{ - config: config, - alloc: config.Allocation, - waitCh: make(chan struct{}), - updateCh: make(chan *structs.Allocation), + alloc: config.Alloc, + clientConfig: config.ClientConfig, + tasks: make(map[string]*taskrunner.TaskRunner), + waitCh: make(chan struct{}), + updateCh: make(chan *structs.Allocation), } + // Create alloc dir + //XXX update AllocDir to hc log + ar.allocDir = allocdir.NewAllocDir(nil, filepath.Join(config.ClientConfig.AllocDir, config.Alloc.ID)) + // Create the logger based on the allocation ID - ar.logger = config.Logger.With("alloc_id", ar.ID()) + ar.logger = config.Logger.With("alloc_id", config.Alloc.ID) // Initialize the runners hooks. ar.initRunnerHooks() - return ar, nil + return ar } func (ar *allocRunner) WaitCh() <-chan struct{} { @@ -115,15 +122,16 @@ POST: // runImpl is used to run the runners. func (ar *allocRunner) runImpl() (<-chan struct{}, error) { // Grab the task group - tg := ar.alloc.Job.LookupTaskGroup(ar.alloc.TaskGroup) + alloc := ar.Alloc() + tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup) if tg == nil { // XXX Fail and exit - ar.logger.Error("failed to lookup task group", "task_group", ar.alloc.TaskGroup) - return nil, fmt.Errorf("failed to lookup task group %q", ar.alloc.TaskGroup) + ar.logger.Error("failed to lookup task group", "task_group", alloc.TaskGroup) + return nil, fmt.Errorf("failed to lookup task group %q", alloc.TaskGroup) } for _, task := range tg.Tasks { - if err := ar.runTask(task); err != nil { + if err := ar.runTask(alloc, task); err != nil { return nil, err } } @@ -142,12 +150,14 @@ func (ar *allocRunner) runImpl() (<-chan struct{}, error) { } // runTask is used to run a task. -func (ar *allocRunner) runTask(task *structs.Task) error { +func (ar *allocRunner) runTask(alloc *structs.Allocation, task *structs.Task) error { // Create the runner config := &taskrunner.Config{ - Parent: &allocRunnerShim{ar}, - Task: task, - Logger: ar.logger, + Alloc: alloc, + ClientConfig: ar.clientConfig, + Task: task, + TaskDir: ar.allocDir.NewTaskDir(task.Name), + Logger: ar.logger, } tr, err := taskrunner.NewTaskRunner(config) if err != nil { @@ -162,6 +172,20 @@ func (ar *allocRunner) runTask(task *structs.Task) error { return nil } +// Alloc returns the current allocation being run by this runner. +//XXX how do we handle mutate the state saving stuff +func (ar *allocRunner) Alloc() *structs.Allocation { + ar.allocLock.RLock() + defer ar.allocLock.RUnlock() + return ar.alloc +} + +// SaveState does all the state related stuff. Who knows. FIXME +//XXX +func (ar *allocRunner) SaveState() error { + return nil +} + // Update the running allocation with a new version received from the server. // // This method is safe for calling concurrently with Run() and does not modify @@ -169,3 +193,56 @@ func (ar *allocRunner) runTask(task *structs.Task) error { func (ar *allocRunner) Update(update *structs.Allocation) { ar.updateCh <- update } + +// Destroy the alloc runner by stopping it if it is still running and cleaning +// up all of its resources. +// +// This method is safe for calling concurrently with Run(). Callers must +// receive on WaitCh() to block until alloc runner has stopped and been +// destroyed. +//XXX TODO +func (ar *allocRunner) Destroy() { + //TODO +} + +// IsDestroyed returns true if the alloc runner has been destroyed (stopped and +// garbage collected). +// +// This method is safe for calling concurrently with Run(). Callers must +// receive on WaitCh() to block until alloc runner has stopped and been +// destroyed. +//XXX TODO +func (ar *allocRunner) IsDestroyed() bool { + return false +} + +// IsWaiting returns true if the alloc runner is waiting for its previous +// allocation to terminate. +// +// This method is safe for calling concurrently with Run(). +//XXX TODO +func (ar *allocRunner) IsWaiting() bool { + return false +} + +// IsMigrating returns true if the alloc runner is migrating data from its +// previous allocation. +// +// This method is safe for calling concurrently with Run(). +//XXX TODO +func (ar *allocRunner) IsMigrating() bool { + return false +} + +// StatsReporter needs implementing +//XXX +func (ar *allocRunner) StatsReporter() allocrunner.AllocStatsReporter { + return noopStatsReporter{} +} + +//FIXME implement +type noopStatsReporter struct{} + +func (noopStatsReporter) LatestAllocStats(taskFilter string) (*cstructs.AllocResourceUsage, error) { + return nil, fmt.Errorf("not implemented") +} diff --git a/client/allocrunnerv2/alloc_runner_getters.go b/client/allocrunnerv2/alloc_runner_getters.go deleted file mode 100644 index 92f6c5faa..000000000 --- a/client/allocrunnerv2/alloc_runner_getters.go +++ /dev/null @@ -1,15 +0,0 @@ -package allocrunnerv2 - -import "github.com/hashicorp/nomad/nomad/structs" - -func (ar *allocRunner) ID() string { - ar.allocLock.Lock() - defer ar.allocLock.Unlock() - return ar.alloc.ID -} - -func (ar *allocRunner) Alloc() *structs.Allocation { - ar.allocLock.Lock() - defer ar.allocLock.Unlock() - return ar.alloc -} diff --git a/client/allocrunnerv2/config/config.go b/client/allocrunnerv2/config.go similarity index 84% rename from client/allocrunnerv2/config/config.go rename to client/allocrunnerv2/config.go index 342d751f1..11bcd3063 100644 --- a/client/allocrunnerv2/config/config.go +++ b/client/allocrunnerv2/config.go @@ -1,4 +1,4 @@ -package config +package allocrunnerv2 import ( "github.com/boltdb/bolt" @@ -15,8 +15,8 @@ type Config struct { // ClientConfig is the clients configuration. ClientConfig *clientconfig.Config - // Allocation captures the allocation that should be run. - Allocation *structs.Allocation + // Alloc captures the allocation that should be run. + Alloc *structs.Allocation // StateDB is used to store and restore state. StateDB *bolt.DB diff --git a/client/allocrunnerv2/interfaces/factory.go b/client/allocrunnerv2/interfaces/factory.go deleted file mode 100644 index fd9afc96b..000000000 --- a/client/allocrunnerv2/interfaces/factory.go +++ /dev/null @@ -1,10 +0,0 @@ -package interfaces - -import ( - "context" - - "github.com/hashicorp/nomad/client/allocrunnerv2/config" -) - -// AllocRunnerFactory is the factory method for retrieving an allocation runner. -type AllocRunnerFactory func(ctx context.Context, config *config.Config) (AllocRunner, error) diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 28b47a58c..36a5a9f60 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -10,10 +10,10 @@ import ( log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" - "github.com/hashicorp/nomad/client/allocrunnerv2/config" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" arstate "github.com/hashicorp/nomad/client/allocrunnerv2/state" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" dstructs "github.com/hashicorp/nomad/client/driver/structs" @@ -46,9 +46,6 @@ type TaskRunner struct { // Logger is the logger for the task runner. logger log.Logger - // allocRunner is the parent allocRunner - allocRunner allocRunner - // waitCh is closed when the task runner has transitioned to a terminal // state waitCh chan struct{} @@ -83,9 +80,11 @@ type TaskRunner struct { } type Config struct { - Parent allocRunner - Task *structs.Task - Logger log.Logger + Alloc *structs.Allocation + ClientConfig *config.Config + Task *structs.Task + TaskDir *allocdir.TaskDir + Logger log.Logger // State is optionally restored task state State *state.State @@ -97,28 +96,27 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { // Initialize the environment builder envBuilder := env.NewBuilder( - config.Parent.Config().ClientConfig.Node, - config.Parent.Alloc(), + config.ClientConfig.Node, + config.Alloc, config.Task, - config.Parent.Config().ClientConfig.Region) + config.ClientConfig.Region) tr := &TaskRunner{ - config: config, - allocRunner: config.Parent, - state: config.State, - ctx: trCtx, - ctxCancel: trCancel, - task: config.Task, - taskDir: config.Parent.GetAllocDir().NewTaskDir(config.Task.Name), - envBuilder: envBuilder, - waitCh: make(chan struct{}), + config: config, + state: config.State, + ctx: trCtx, + ctxCancel: trCancel, + task: config.Task, + taskDir: config.TaskDir, + envBuilder: envBuilder, + waitCh: make(chan struct{}), } // Create the logger based on the allocation ID tr.logger = config.Logger.Named("task_runner").With("task", config.Task.Name) // Build the restart tracker. - alloc := tr.allocRunner.Alloc() + alloc := config.Alloc tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup) if tg == nil { tr.logger.Error("alloc missing task group") @@ -156,7 +154,7 @@ func (tr *TaskRunner) initState() { } func (tr *TaskRunner) initLabels() { - alloc := tr.allocRunner.Alloc() + alloc := tr.config.Alloc tr.baseLabels = []metrics.Label{ { Name: "job", @@ -316,15 +314,15 @@ func (tr *TaskRunner) initDriver() error { tr.SetState("", structs.NewTaskEvent(structs.TaskDriverMessage).SetDriverMessage(msg)) } - alloc := tr.allocRunner.Alloc() + alloc := tr.config.Alloc driverCtx := driver.NewDriverContext( alloc.Job.Name, alloc.TaskGroup, tr.Name(), - tr.allocRunner.ID(), - tr.allocRunner.Config().ClientConfig, // XXX Why does it need this - tr.allocRunner.Config().ClientConfig.Node, // XXX THIS I NEED TO FIX - tr.logger.StandardLogger(nil), // XXX Should pass this through + alloc.ID, + tr.config.ClientConfig, // XXX Why does it need this + tr.config.ClientConfig.Node, // XXX THIS I NEED TO FIX + tr.logger.StandardLogger(nil), // XXX Should pass this through eventEmitter) driver, err := driver.NewDriver(tr.task.Driver, driverCtx) @@ -358,7 +356,7 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { } if event.Type == structs.TaskRestarting { - if !tr.allocRunner.Config().ClientConfig.DisableTaggedMetrics { + if !tr.config.ClientConfig.DisableTaggedMetrics { metrics.IncrCounterWithLabels([]string{"client", "allocs", "restart"}, 1, tr.baseLabels) } //if r.config.BackwardsCompatibleMetrics { @@ -376,7 +374,7 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { // Capture the start time if it is just starting if task.State != structs.TaskStateRunning { task.StartedAt = time.Now().UTC() - if !tr.allocRunner.Config().ClientConfig.DisableTaggedMetrics { + if !tr.config.ClientConfig.DisableTaggedMetrics { metrics.IncrCounterWithLabels([]string{"client", "allocs", "running"}, 1, tr.baseLabels) } //if r.config.BackwardsCompatibleMetrics { @@ -391,14 +389,14 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { // Emitting metrics to indicate task complete and failures if task.Failed { - if !tr.allocRunner.Config().ClientConfig.DisableTaggedMetrics { + if !tr.config.ClientConfig.DisableTaggedMetrics { metrics.IncrCounterWithLabels([]string{"client", "allocs", "failed"}, 1, tr.baseLabels) } //if r.config.BackwardsCompatibleMetrics { //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "failed"}, 1) //} } else { - if !tr.allocRunner.Config().ClientConfig.DisableTaggedMetrics { + if !tr.config.ClientConfig.DisableTaggedMetrics { metrics.IncrCounterWithLabels([]string{"client", "allocs", "complete"}, 1, tr.baseLabels) } //if r.config.BackwardsCompatibleMetrics { @@ -408,9 +406,10 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { } // Create a copy and notify the alloc runner of the transition - if err := tr.allocRunner.StateUpdated(tr.state.Copy()); err != nil { - tr.logger.Error("failed to save state", "error", err) - } + //FIXME + //if err := tr.allocRunner.StateUpdated(tr.state.Copy()); err != nil { + //tr.logger.Error("failed to save state", "error", err) + //} } // appendTaskEvent updates the task status by appending the new event. diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 4f050a23a..af40005e3 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -30,7 +30,7 @@ func (tr *TaskRunner) initHooks() { func (tr *TaskRunner) prerun() error { // Determine if the allocation is terminaland we should avoid running // pre-run hooks. - alloc := tr.allocRunner.Alloc() + alloc := tr.config.Alloc if alloc.TerminalStatus() { tr.logger.Trace("skipping pre-run hooks since allocation is terminal") return nil @@ -77,7 +77,7 @@ func (tr *TaskRunner) prerun() error { //XXX Can we assume everything only wants to be run until //successful and simply keep track of which hooks have yet to //run on failures+retries? - if hookState.SuccessfulOnce { + if hookState != nil && hookState.SuccessfulOnce { tr.logger.Trace("skipping hook since it was successfully run once", "name", name) continue } @@ -106,9 +106,11 @@ func (tr *TaskRunner) prerun() error { // XXX Detect if state has changed so that we can signal to the // alloc runner precisly - if err := tr.allocRunner.StateUpdated(tr.state.Copy()); err != nil { - tr.logger.Error("failed to save state", "error", err) - } + /* + if err := tr.allocRunner.StateUpdated(tr.state.Copy()); err != nil { + tr.logger.Error("failed to save state", "error", err) + } + */ tr.state.Unlock() } @@ -225,7 +227,7 @@ func (h *taskDirHook) Name() string { } func (h *taskDirHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { - cc := h.runner.allocRunner.Config().ClientConfig + cc := h.runner.config.ClientConfig chroot := cconfig.DefaultChrootEnv if len(cc.ChrootEnv) > 0 { chroot = cc.ChrootEnv @@ -242,7 +244,7 @@ func (h *taskDirHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interfaces } // Update the environment variables based on the built task directory - driver.SetEnvvars(h.runner.envBuilder, fsi, h.runner.taskDir, h.runner.allocRunner.Config().ClientConfig) + driver.SetEnvvars(h.runner.envBuilder, fsi, h.runner.taskDir, h.runner.config.ClientConfig) return nil } diff --git a/client/allocrunnerv2/tr_shim.go b/client/allocrunnerv2/tr_shim.go index 7cab963c4..b20246ea1 100644 --- a/client/allocrunnerv2/tr_shim.go +++ b/client/allocrunnerv2/tr_shim.go @@ -2,7 +2,6 @@ package allocrunnerv2 import ( "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunnerv2/config" "github.com/hashicorp/nomad/client/allocrunnerv2/state" trstate "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" ) @@ -17,10 +16,6 @@ func (a *allocRunnerShim) State() *state.State { return a.state } -func (a *allocRunnerShim) Config() *config.Config { - return a.config -} - func (a *allocRunnerShim) GetAllocDir() *allocdir.AllocDir { return a.allocDir } diff --git a/client/client.go b/client/client.go index b855ac9dd..2c690fbb5 100644 --- a/client/client.go +++ b/client/client.go @@ -29,6 +29,7 @@ import ( "github.com/hashicorp/consul/lib" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner" + "github.com/hashicorp/nomad/client/allocrunnerv2" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/servers" "github.com/hashicorp/nomad/client/state" @@ -99,6 +100,18 @@ type ClientStatsReporter interface { LatestHostStats() *stats.HostStats } +type AllocRunner interface { + StatsReporter() allocrunner.AllocStatsReporter + Destroy() + IsDestroyed() bool + IsWaiting() bool + IsMigrating() bool + WaitCh() <-chan struct{} + SaveState() error + Update(*structs.Allocation) + Alloc() *structs.Allocation +} + // Client is used to implement the client interaction with Nomad. Clients // are expected to register as a schedulable node to the servers, and to // run allocations as determined by the servers. @@ -150,7 +163,7 @@ type Client struct { // allocs maps alloc IDs to their AllocRunner. This map includes all // AllocRunners - running and GC'd - until the server GCs them. - allocs map[string]*allocrunner.AllocRunner + allocs map[string]AllocRunner allocLock sync.RWMutex // allocUpdates stores allocations that need to be synced to the server. @@ -225,7 +238,6 @@ func NewClient(cfg *config.Config, consulCatalog consul.CatalogAPI, consulServic tlsWrap: tlsWrap, streamingRpcs: structs.NewStreamingRpcRegistry(), logger: logger.StandardLogger(&hclog.StandardLoggerOptions{InferLevels: true}), - allocs: make(map[string]*allocrunner.AllocRunner), allocUpdates: make(chan *structs.Allocation, 64), shutdownCh: make(chan struct{}), triggerDiscoveryCh: make(chan struct{}), @@ -522,16 +534,12 @@ func (c *Client) Shutdown() error { // Destroy all the running allocations. if c.config.DevMode { - var wg sync.WaitGroup for _, ar := range c.getAllocRunners() { - wg.Add(1) - go func(ar *allocrunner.AllocRunner) { - ar.Destroy() - <-ar.WaitCh() - wg.Done() - }(ar) + ar.Destroy() + } + for _, ar := range c.getAllocRunners() { + <-ar.WaitCh() } - wg.Wait() } c.shutdown = true @@ -614,11 +622,15 @@ func (c *Client) GetAllocFS(allocID string) (allocdir.AllocDirFS, error) { c.allocLock.RLock() defer c.allocLock.RUnlock() - ar, ok := c.allocs[allocID] + _, ok := c.allocs[allocID] if !ok { return nil, structs.NewErrUnknownAllocation(allocID) } - return ar.GetAllocDir(), nil + + //XXX Experiment in splitting AllocDir and TaskDir structs as the FS + //API only needs the immutable AllocDir struct + //FIXME needs a better logger though -- good reason to still get it from alloc runner? + return allocdir.NewAllocDir(c.logger, filepath.Join(c.config.AllocDir, allocID)), nil } // GetClientAlloc returns the allocation from the client @@ -803,7 +815,7 @@ func (c *Client) saveState() error { wg.Add(len(runners)) for id, ar := range runners { - go func(id string, ar *allocrunner.AllocRunner) { + go func(id string, ar AllocRunner) { err := ar.SaveState() if err != nil { c.logger.Printf("[ERR] client: failed to save state for alloc %q: %v", id, err) @@ -820,10 +832,10 @@ func (c *Client) saveState() error { } // getAllocRunners returns a snapshot of the current set of alloc runners. -func (c *Client) getAllocRunners() map[string]*allocrunner.AllocRunner { +func (c *Client) getAllocRunners() map[string]AllocRunner { c.allocLock.RLock() defer c.allocLock.RUnlock() - runners := make(map[string]*allocrunner.AllocRunner, len(c.allocs)) + runners := make(map[string]AllocRunner, len(c.allocs)) for id, ar := range c.allocs { runners[id] = ar } @@ -1498,7 +1510,7 @@ func (c *Client) updateAllocStatus(alloc *structs.Allocation) { c.allocLock.RUnlock() if ok { - c.garbageCollector.MarkForCollection(ar) + c.garbageCollector.MarkForCollection(alloc.ID, ar) // Trigger a GC in case we're over thresholds and just // waiting for eligible allocs. @@ -1660,14 +1672,16 @@ OUTER: // need to pull all the allocations. var pull []string filtered := make(map[string]struct{}) - runners := c.getAllocRunners() var pullIndex uint64 for allocID, modifyIndex := range resp.Allocs { // Pull the allocation if we don't have an alloc runner for the // allocation or if the alloc runner requires an updated allocation. - runner, ok := runners[allocID] + //XXX Part of Client alloc index tracking exp + c.allocLock.RLock() + currentAR, ok := c.allocs[allocID] + c.allocLock.RUnlock() - if !ok || runner.ShouldUpdate(modifyIndex) { + if !ok || modifyIndex > currentAR.Alloc().AllocModifyIndex { // Only pull allocs that are required. Filtered // allocs might be at a higher index, so ignore // it. @@ -1798,14 +1812,14 @@ func (c *Client) watchNodeUpdates() { func (c *Client) runAllocs(update *allocUpdates) { // Get the existing allocs c.allocLock.RLock() - exist := make([]*structs.Allocation, 0, len(c.allocs)) - for _, ar := range c.allocs { - exist = append(exist, ar.Alloc()) + existing := make(map[string]uint64, len(c.allocs)) + for id, ar := range c.allocs { + existing[id] = ar.Alloc().AllocModifyIndex } c.allocLock.RUnlock() // Diff the existing and updated allocations - diff := diffAllocs(exist, update) + diff := diffAllocs(existing, update) c.logger.Printf("[DEBUG] client: %#v", diff) // Remove the old allocations @@ -1815,9 +1829,9 @@ func (c *Client) runAllocs(update *allocUpdates) { // Update the existing allocations for _, update := range diff.updated { - if err := c.updateAlloc(update.exist, update.updated); err != nil { + if err := c.updateAlloc(update); err != nil { c.logger.Printf("[ERR] client: failed to update alloc %q: %v", - update.exist.ID, err) + update.ID, err) } } @@ -1842,34 +1856,33 @@ func (c *Client) runAllocs(update *allocUpdates) { // removeAlloc is invoked when we should remove an allocation because it has // been removed by the server. -func (c *Client) removeAlloc(alloc *structs.Allocation) { +func (c *Client) removeAlloc(allocID string) { c.allocLock.Lock() - ar, ok := c.allocs[alloc.ID] + defer c.allocLock.Unlock() + ar, ok := c.allocs[allocID] if !ok { - c.allocLock.Unlock() - c.logger.Printf("[WARN] client: missing context for alloc '%s'", alloc.ID) + c.logger.Printf("[WARN] client: missing context for alloc '%s'", allocID) return } // Stop tracking alloc runner as it's been GC'd by the server - delete(c.allocs, alloc.ID) - c.allocLock.Unlock() + delete(c.allocs, allocID) // Ensure the GC has a reference and then collect. Collecting through the GC // applies rate limiting - c.garbageCollector.MarkForCollection(ar) + c.garbageCollector.MarkForCollection(allocID, ar) // GC immediately since the server has GC'd it - go c.garbageCollector.Collect(alloc.ID) + go c.garbageCollector.Collect(allocID) } // updateAlloc is invoked when we should update an allocation -func (c *Client) updateAlloc(exist, update *structs.Allocation) error { - c.allocLock.RLock() - ar, ok := c.allocs[exist.ID] - c.allocLock.RUnlock() +func (c *Client) updateAlloc(update *structs.Allocation) error { + c.allocLock.Lock() + defer c.allocLock.Unlock() + ar, ok := c.allocs[update.ID] if !ok { - c.logger.Printf("[WARN] client: missing context for alloc '%s'", exist.ID) + c.logger.Printf("[WARN] client: missing context for alloc %q", update.ID) return nil } @@ -1879,33 +1892,53 @@ func (c *Client) updateAlloc(exist, update *structs.Allocation) error { // addAlloc is invoked when we should add an allocation func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error { - // Check if we already have an alloc runner c.allocLock.Lock() defer c.allocLock.Unlock() + + // Check if we already have an alloc runner if _, ok := c.allocs[alloc.ID]; ok { c.logger.Printf("[DEBUG]: client: dropping duplicate add allocation request: %q", alloc.ID) return nil } + //FIXME disabled previous alloc waiting/migrating // get the previous alloc runner - if one exists - for the // blocking/migrating watcher - var prevAR *allocrunner.AllocRunner - if alloc.PreviousAllocation != "" { - prevAR = c.allocs[alloc.PreviousAllocation] - } + /* + var prevAR *allocrunner.AllocRunner + if alloc.PreviousAllocation != "" { + prevAR = c.allocs[alloc.PreviousAllocation] + } - c.configLock.RLock() - prevAlloc := allocrunner.NewAllocWatcher(alloc, prevAR, c, c.configCopy, c.logger, migrateToken) + c.configLock.RLock() + prevAlloc := allocrunner.NewAllocWatcher(alloc, prevAR, c, c.configCopy, c.logger, migrateToken) + */ // Copy the config since the node can be swapped out as it is being updated. // The long term fix is to pass in the config and node separately and then // we don't have to do a copy. - ar := allocrunner.NewAllocRunner(c.logger, c.configCopy.Copy(), c.stateDB, c.updateAllocStatus, alloc, c.vaultClient, c.consulService, prevAlloc) + //ar := allocrunner.NewAllocRunner(c.logger, c.configCopy.Copy(), c.stateDB, c.updateAllocStatus, alloc, c.vaultClient, c.consulService, prevAlloc) + //XXX FIXME + logger := hclog.New(&hclog.LoggerOptions{ + Name: "nomad", + Level: hclog.LevelFromString("DEBUG"), + }) + + c.configLock.RLock() + arConf := &allocrunnerv2.Config{ + Logger: logger, + ClientConfig: c.config, + Alloc: alloc, + StateDB: nil, //FIXME ? + } c.configLock.RUnlock() + ar := allocrunnerv2.NewAllocRunner(arConf) + // Store the alloc runner. c.allocs[alloc.ID] = ar + //XXX(schmichael) Why do we do this? if err := ar.SaveState(); err != nil { c.logger.Printf("[WARN] client: initial save state for alloc %q failed: %v", alloc.ID, err) } diff --git a/client/gc.go b/client/gc.go index 33084aa1d..4d4eb7f17 100644 --- a/client/gc.go +++ b/client/gc.go @@ -7,7 +7,6 @@ import ( "sync" "time" - "github.com/hashicorp/nomad/client/allocrunner" "github.com/hashicorp/nomad/client/stats" "github.com/hashicorp/nomad/nomad/structs" ) @@ -162,7 +161,7 @@ func (a *AllocGarbageCollector) keepUsageBelowThreshold() error { } // Destroy the alloc runner and wait until it exits - a.destroyAllocRunner(gcAlloc.allocRunner, reason) + a.destroyAllocRunner(gcAlloc.allocID, gcAlloc.allocRunner, reason) } return nil } @@ -170,12 +169,8 @@ func (a *AllocGarbageCollector) keepUsageBelowThreshold() error { // destroyAllocRunner is used to destroy an allocation runner. It will acquire a // lock to restrict parallelism and then destroy the alloc runner, returning // once the allocation has been destroyed. -func (a *AllocGarbageCollector) destroyAllocRunner(ar *allocrunner.AllocRunner, reason string) { - id := "" - if alloc := ar.Alloc(); alloc != nil { - id = alloc.ID - } - a.logger.Printf("[INFO] client.gc: garbage collecting allocation %s due to %s", id, reason) +func (a *AllocGarbageCollector) destroyAllocRunner(allocID string, ar AllocRunner, reason string) { + a.logger.Printf("[INFO] client.gc: garbage collecting allocation %s due to %s", allocID, reason) // Acquire the destroy lock select { @@ -191,7 +186,7 @@ func (a *AllocGarbageCollector) destroyAllocRunner(ar *allocrunner.AllocRunner, case <-a.shutdownCh: } - a.logger.Printf("[DEBUG] client.gc: garbage collected %q", ar.Alloc().ID) + a.logger.Printf("[DEBUG] client.gc: garbage collected %s", allocID) // Release the lock <-a.destroyCh @@ -205,7 +200,7 @@ func (a *AllocGarbageCollector) Stop() { // alloc was found and garbage collected; otherwise false. func (a *AllocGarbageCollector) Collect(allocID string) bool { if gcAlloc := a.allocRunners.Remove(allocID); gcAlloc != nil { - a.destroyAllocRunner(gcAlloc.allocRunner, "forced collection") + a.destroyAllocRunner(allocID, gcAlloc.allocRunner, "forced collection") return true } @@ -227,7 +222,7 @@ func (a *AllocGarbageCollector) CollectAll() { return } - go a.destroyAllocRunner(gcAlloc.allocRunner, "forced full node collection") + go a.destroyAllocRunner(gcAlloc.allocID, gcAlloc.allocRunner, "forced full node collection") } } @@ -257,7 +252,7 @@ func (a *AllocGarbageCollector) MakeRoomFor(allocations []*structs.Allocation) e } // Destroy the alloc runner and wait until it exits - a.destroyAllocRunner(gcAlloc.allocRunner, fmt.Sprintf("new allocations and over max (%d)", a.config.MaxAllocs)) + a.destroyAllocRunner(gcAlloc.allocID, gcAlloc.allocRunner, fmt.Sprintf("new allocations and over max (%d)", a.config.MaxAllocs)) } totalResource := &structs.AllocatedSharedResources{} @@ -330,7 +325,7 @@ func (a *AllocGarbageCollector) MakeRoomFor(allocations []*structs.Allocation) e } // Destroy the alloc runner and wait until it exits - a.destroyAllocRunner(ar, fmt.Sprintf("freeing %d MB for new allocations", allocDiskMB)) + a.destroyAllocRunner(gcAlloc.allocID, ar, fmt.Sprintf("freeing %d MB for new allocations", allocDiskMB)) diskCleared += allocDiskMB } @@ -338,13 +333,8 @@ func (a *AllocGarbageCollector) MakeRoomFor(allocations []*structs.Allocation) e } // MarkForCollection starts tracking an allocation for Garbage Collection -func (a *AllocGarbageCollector) MarkForCollection(ar *allocrunner.AllocRunner) { - if ar.Alloc() == nil { - a.destroyAllocRunner(ar, "alloc is nil") - return - } - - if a.allocRunners.Push(ar) { +func (a *AllocGarbageCollector) MarkForCollection(allocID string, ar AllocRunner) { + if a.allocRunners.Push(allocID, ar) { a.logger.Printf("[INFO] client.gc: marking allocation %v for GC", ar.Alloc().ID) } } @@ -353,7 +343,8 @@ func (a *AllocGarbageCollector) MarkForCollection(ar *allocrunner.AllocRunner) { // a PQ type GCAlloc struct { timeStamp time.Time - allocRunner *allocrunner.AllocRunner + allocID string + allocRunner AllocRunner index int } @@ -407,20 +398,20 @@ func NewIndexedGCAllocPQ() *IndexedGCAllocPQ { // Push an alloc runner into the GC queue. Returns true if alloc was added, // false if the alloc already existed. -func (i *IndexedGCAllocPQ) Push(ar *allocrunner.AllocRunner) bool { +func (i *IndexedGCAllocPQ) Push(allocID string, ar AllocRunner) bool { i.pqLock.Lock() defer i.pqLock.Unlock() - alloc := ar.Alloc() - if _, ok := i.index[alloc.ID]; ok { + if _, ok := i.index[allocID]; ok { // No work to do return false } gcAlloc := &GCAlloc{ timeStamp: time.Now(), + allocID: allocID, allocRunner: ar, } - i.index[alloc.ID] = gcAlloc + i.index[allocID] = gcAlloc heap.Push(&i.heap, gcAlloc) return true } diff --git a/client/util.go b/client/util.go index 02ce36a5e..e3662a303 100644 --- a/client/util.go +++ b/client/util.go @@ -7,16 +7,12 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) -type allocTuple struct { - exist, updated *structs.Allocation -} - // diffResult is used to return the sets that result from a diff type diffResult struct { added []*structs.Allocation - removed []*structs.Allocation - updated []allocTuple - ignore []*structs.Allocation + removed []string + updated []*structs.Allocation + ignore []string } func (d *diffResult) GoString() string { @@ -26,38 +22,34 @@ func (d *diffResult) GoString() string { // diffAllocs is used to diff the existing and updated allocations // to see what has happened. -func diffAllocs(existing []*structs.Allocation, allocs *allocUpdates) *diffResult { +func diffAllocs(existing map[string]uint64, allocs *allocUpdates) *diffResult { // Scan the existing allocations result := &diffResult{} - existIdx := make(map[string]struct{}) - for _, exist := range existing { - // Mark this as existing - existIdx[exist.ID] = struct{}{} - + for existID, existIndex := range existing { // Check if the alloc was updated or filtered because an update wasn't // needed. - alloc, pulled := allocs.pulled[exist.ID] - _, filtered := allocs.filtered[exist.ID] + alloc, pulled := allocs.pulled[existID] + _, filtered := allocs.filtered[existID] // If not updated or filtered, removed if !pulled && !filtered { - result.removed = append(result.removed, exist) + result.removed = append(result.removed, existID) continue } // Check for an update - if pulled && alloc.AllocModifyIndex > exist.AllocModifyIndex { - result.updated = append(result.updated, allocTuple{exist, alloc}) + if pulled && alloc.AllocModifyIndex > existIndex { + result.updated = append(result.updated, alloc) continue } // Ignore this - result.ignore = append(result.ignore, exist) + result.ignore = append(result.ignore, existID) } // Scan the updated allocations for any that are new for id, pulled := range allocs.pulled { - if _, ok := existIdx[id]; !ok { + if _, ok := existing[id]; !ok { result.added = append(result.added, pulled) } } From 637ef264ae2203e926c0d2f1dc659d17e443c5da Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 28 Jun 2018 17:20:13 -0700 Subject: [PATCH 005/152] Copy TR.Config vals to TR I think I like this pattern better as some Config vals are mutable (Alloc) and some aren't and some are used to derive other values and never used directly. Promoting them onto the TR struct is a little more work but is hopefully more clear as to how each value is used. --- .../allocrunnerv2/taskrunner/task_runner.go | 63 ++++++++++--------- .../taskrunner/task_runner_hooks.go | 6 +- 2 files changed, 36 insertions(+), 33 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 36a5a9f60..9a3403be3 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -11,7 +11,6 @@ import ( "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - arstate "github.com/hashicorp/nomad/client/allocrunnerv2/state" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver" @@ -20,18 +19,14 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) -type allocRunner interface { - State() *arstate.State // XXX Do we need it - Alloc() *structs.Allocation - ID() string - Config() *config.Config - GetAllocDir() *allocdir.AllocDir - StateUpdated(*state.State) error -} - type TaskRunner struct { - // config is the task runners configuration. - config *Config + // allocID is immutable so store a copy to access without locks + allocID string + + alloc *structs.Allocation + allocLock sync.Mutex + + clientConfig *config.Config // state captures the state of the task runner state *state.State @@ -102,14 +97,16 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { config.ClientConfig.Region) tr := &TaskRunner{ - config: config, - state: config.State, - ctx: trCtx, - ctxCancel: trCancel, - task: config.Task, - taskDir: config.TaskDir, - envBuilder: envBuilder, - waitCh: make(chan struct{}), + alloc: config.Alloc, + allocID: config.Alloc.ID, + clientConfig: config.ClientConfig, + task: config.Task, + taskDir: config.TaskDir, + envBuilder: envBuilder, + state: config.State, + ctx: trCtx, + ctxCancel: trCancel, + waitCh: make(chan struct{}), } // Create the logger based on the allocation ID @@ -154,7 +151,7 @@ func (tr *TaskRunner) initState() { } func (tr *TaskRunner) initLabels() { - alloc := tr.config.Alloc + alloc := tr.Alloc() tr.baseLabels = []metrics.Label{ { Name: "job", @@ -166,7 +163,7 @@ func (tr *TaskRunner) initLabels() { }, { Name: "alloc_id", - Value: alloc.ID, + Value: tr.allocID, }, { Name: "task", @@ -314,14 +311,14 @@ func (tr *TaskRunner) initDriver() error { tr.SetState("", structs.NewTaskEvent(structs.TaskDriverMessage).SetDriverMessage(msg)) } - alloc := tr.config.Alloc + alloc := tr.Alloc() driverCtx := driver.NewDriverContext( alloc.Job.Name, alloc.TaskGroup, tr.Name(), - alloc.ID, - tr.config.ClientConfig, // XXX Why does it need this - tr.config.ClientConfig.Node, // XXX THIS I NEED TO FIX + tr.allocID, + tr.clientConfig, // XXX Why does it need this + tr.clientConfig.Node, // XXX THIS I NEED TO FIX tr.logger.StandardLogger(nil), // XXX Should pass this through eventEmitter) @@ -356,7 +353,7 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { } if event.Type == structs.TaskRestarting { - if !tr.config.ClientConfig.DisableTaggedMetrics { + if !tr.clientConfig.DisableTaggedMetrics { metrics.IncrCounterWithLabels([]string{"client", "allocs", "restart"}, 1, tr.baseLabels) } //if r.config.BackwardsCompatibleMetrics { @@ -374,7 +371,7 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { // Capture the start time if it is just starting if task.State != structs.TaskStateRunning { task.StartedAt = time.Now().UTC() - if !tr.config.ClientConfig.DisableTaggedMetrics { + if !tr.clientConfig.DisableTaggedMetrics { metrics.IncrCounterWithLabels([]string{"client", "allocs", "running"}, 1, tr.baseLabels) } //if r.config.BackwardsCompatibleMetrics { @@ -389,14 +386,14 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { // Emitting metrics to indicate task complete and failures if task.Failed { - if !tr.config.ClientConfig.DisableTaggedMetrics { + if !tr.clientConfig.DisableTaggedMetrics { metrics.IncrCounterWithLabels([]string{"client", "allocs", "failed"}, 1, tr.baseLabels) } //if r.config.BackwardsCompatibleMetrics { //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "failed"}, 1) //} } else { - if !tr.config.ClientConfig.DisableTaggedMetrics { + if !tr.clientConfig.DisableTaggedMetrics { metrics.IncrCounterWithLabels([]string{"client", "allocs", "complete"}, 1, tr.baseLabels) } //if r.config.BackwardsCompatibleMetrics { @@ -412,6 +409,12 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { //} } +func (tr *TaskRunner) Alloc() *structs.Allocation { + tr.allocLock.Lock() + defer tr.allocLock.Unlock() + return tr.alloc +} + // appendTaskEvent updates the task status by appending the new event. func appendTaskEvent(state *structs.TaskState, event *structs.TaskEvent) { capacity := 10 diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index af40005e3..6daae44e9 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -30,7 +30,7 @@ func (tr *TaskRunner) initHooks() { func (tr *TaskRunner) prerun() error { // Determine if the allocation is terminaland we should avoid running // pre-run hooks. - alloc := tr.config.Alloc + alloc := tr.Alloc() if alloc.TerminalStatus() { tr.logger.Trace("skipping pre-run hooks since allocation is terminal") return nil @@ -227,7 +227,7 @@ func (h *taskDirHook) Name() string { } func (h *taskDirHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { - cc := h.runner.config.ClientConfig + cc := h.runner.clientConfig chroot := cconfig.DefaultChrootEnv if len(cc.ChrootEnv) > 0 { chroot = cc.ChrootEnv @@ -244,7 +244,7 @@ func (h *taskDirHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interfaces } // Update the environment variables based on the built task directory - driver.SetEnvvars(h.runner.envBuilder, fsi, h.runner.taskDir, h.runner.config.ClientConfig) + driver.SetEnvvars(h.runner.envBuilder, fsi, h.runner.taskDir, h.runner.clientConfig) return nil } From f473cd03d6ee9d2e7589e519ba2ae50b5b94f4c2 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Fri, 29 Jun 2018 14:53:31 -0700 Subject: [PATCH 006/152] tr: start update/shutdown logic --- .../allocrunnerv2/taskrunner/task_runner.go | 68 ++++++++++++------- 1 file changed, 42 insertions(+), 26 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 9a3403be3..cdbaa7ee5 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -31,16 +31,21 @@ type TaskRunner struct { // state captures the state of the task runner state *state.State - // ctx is the alloc runners context. It is cancelled when all related - // activity for this allocation should be terminated. + // ctx is the task runner's context and is done whe the task runner + // should exit. If a task runner is destroyed it will exit regardless + // of whether the context is done. ctx context.Context - // ctxCancel is used to cancel the alloc runners cancel + // ctxCancel is used to exit the task runner's Run loop (without + // stopping or destroying the task) ctxCancel context.CancelFunc // Logger is the logger for the task runner. logger log.Logger + // updateCh receives Alloc updates + updateCh chan *structs.Allocation + // waitCh is closed when the task runner has transitioned to a terminal // state waitCh chan struct{} @@ -94,7 +99,8 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { config.ClientConfig.Node, config.Alloc, config.Task, - config.ClientConfig.Region) + config.ClientConfig.Region, + ) tr := &TaskRunner{ alloc: config.Alloc, @@ -106,6 +112,7 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { state: config.State, ctx: trCtx, ctxCancel: trCancel, + updateCh: make(chan *structs.Allocation), waitCh: make(chan struct{}), } @@ -113,13 +120,12 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { tr.logger = config.Logger.Named("task_runner").With("task", config.Task.Name) // Build the restart tracker. - alloc := config.Alloc - tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup) + tg := tr.alloc.Job.LookupTaskGroup(tr.alloc.TaskGroup) if tg == nil { tr.logger.Error("alloc missing task group") return nil, fmt.Errorf("alloc missing task group") } - tr.restartTracker = restarts.NewRestartTracker(tg.RestartPolicy, alloc.Job.Type) + tr.restartTracker = restarts.NewRestartTracker(tg.RestartPolicy, tr.alloc.Job.Type) // Initialize the task state tr.initState() @@ -172,24 +178,6 @@ func (tr *TaskRunner) initLabels() { } } -// WaitCh is closed when TaskRunner.Run exits. -func (tr *TaskRunner) WaitCh() <-chan struct{} { - return tr.waitCh -} - -// Update the running allocation with a new version received from the server. -// -// This method is safe for calling concurrently with Run() and does not modify -// the passed in allocation. -func (tr *TaskRunner) Update(update *structs.Allocation) { - // XXX -} - -func (tr *TaskRunner) Stop() { - // XXX - tr.ctxCancel() -} - func (tr *TaskRunner) Run() { defer close(tr.waitCh) @@ -205,7 +193,7 @@ func (tr *TaskRunner) Run() { goto RESTART } - // Run and wait + // Run the task waitRes, err = tr.runDriver() if err != nil { tr.logger.Error("running driver failed", "error", err) @@ -409,6 +397,34 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { //} } +// WaitCh is closed when TaskRunner.Run exits. +func (tr *TaskRunner) WaitCh() <-chan struct{} { + return tr.waitCh +} + +// Update the running allocation with a new version received from the server. +// +// This method is safe for calling concurrently with Run() and does not modify +// the passed in allocation. +func (tr *TaskRunner) Update(update *structs.Allocation) { + select { + case tr.updateCh <- update: + case <-tr.WaitCh(): + //XXX Do we log here like we used to? If we're just + //shutting down it's not an error to drop the update as + //it will be applied on startup + } +} + +// Shutdown the task runner. Does not stop the task or garbage collect a +// stopped task. +// +// This method is safe for calling concurently with Run(). Callers must +// receive on WaitCh() to block until Run() has exited. +func (tr *TaskRunner) Shutdown() { + tr.ctxCancel() +} + func (tr *TaskRunner) Alloc() *structs.Allocation { tr.allocLock.Lock() defer tr.allocLock.Unlock() From 67874e761fed578d346b3ded5c6644ca7c744b9f Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Fri, 29 Jun 2018 15:39:22 -0700 Subject: [PATCH 007/152] tr: don't lock for immutable fields --- client/allocrunnerv2/taskrunner/task_runner.go | 11 +++++++---- .../allocrunnerv2/taskrunner/task_runner_getters.go | 6 ------ 2 files changed, 7 insertions(+), 10 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index cdbaa7ee5..be824b2dd 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -20,8 +20,10 @@ import ( ) type TaskRunner struct { - // allocID is immutable so store a copy to access without locks - allocID string + // allocID and taskName are immutable so store a copy to access without + // locks + allocID string + taskName string alloc *structs.Allocation allocLock sync.Mutex @@ -108,6 +110,7 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { clientConfig: config.ClientConfig, task: config.Task, taskDir: config.TaskDir, + taskName: config.Task.Name, envBuilder: envBuilder, state: config.State, ctx: trCtx, @@ -173,7 +176,7 @@ func (tr *TaskRunner) initLabels() { }, { Name: "task", - Value: tr.Name(), + Value: tr.taskName, }, } } @@ -303,7 +306,7 @@ func (tr *TaskRunner) initDriver() error { driverCtx := driver.NewDriverContext( alloc.Job.Name, alloc.TaskGroup, - tr.Name(), + tr.taskName, tr.allocID, tr.clientConfig, // XXX Why does it need this tr.clientConfig.Node, // XXX THIS I NEED TO FIX diff --git a/client/allocrunnerv2/taskrunner/task_runner_getters.go b/client/allocrunnerv2/taskrunner/task_runner_getters.go index 7c8e8c679..dd0a5ac20 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_getters.go +++ b/client/allocrunnerv2/taskrunner/task_runner_getters.go @@ -2,12 +2,6 @@ package taskrunner import "github.com/hashicorp/nomad/nomad/structs" -func (tr *TaskRunner) Name() string { - tr.taskLock.RLock() - defer tr.taskLock.RUnlock() - return tr.task.Name -} - func (tr *TaskRunner) Task() *structs.Task { tr.taskLock.RLock() defer tr.taskLock.RUnlock() From 561260d6fedb2805e6eb1fd560c693c49bf78abe Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Fri, 29 Jun 2018 15:39:54 -0700 Subject: [PATCH 008/152] tr: skip error/success saving All hooks only need to be run once. Since only one hook can fail per run there's no need to track errors on a per hook basis. --- .../allocrunnerv2/taskrunner/state/state.go | 4 +-- .../taskrunner/task_runner_hooks.go | 30 +++++-------------- 2 files changed, 9 insertions(+), 25 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/state/state.go b/client/allocrunnerv2/taskrunner/state/state.go index c46ab8ace..5c079a9e1 100644 --- a/client/allocrunnerv2/taskrunner/state/state.go +++ b/client/allocrunnerv2/taskrunner/state/state.go @@ -34,9 +34,7 @@ func (s *State) Copy() *State { } type HookState struct { - SuccessfulOnce bool - Data map[string]string - LastError error + Data map[string]string } func (h *HookState) Copy() *HookState { diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 6daae44e9..3be92e94c 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -28,6 +28,8 @@ func (tr *TaskRunner) initHooks() { // prerun is used to run the runners prerun hooks. func (tr *TaskRunner) prerun() error { + //XXX is this necessary? maybe we should have a generic cancelletion + // method instead of peeking into the alloc // Determine if the allocation is terminaland we should avoid running // pre-run hooks. alloc := tr.Alloc() @@ -68,23 +70,19 @@ func (tr *TaskRunner) prerun() error { tr.state.RLock() hookState := tr.state.Hooks[name] if hookState != nil { - req.HookData = hookState.Data + // Hook already ran, skip + tr.state.RUnlock() + continue } req.VaultToken = tr.state.VaultToken tr.state.RUnlock() - //XXX Can we assume everything only wants to be run until - //successful and simply keep track of which hooks have yet to - //run on failures+retries? - if hookState != nil && hookState.SuccessfulOnce { - tr.logger.Trace("skipping hook since it was successfully run once", "name", name) - continue - } - // Run the pre-run hook var resp interfaces.TaskPrerunResponse - err := pre.Prerun(&req, &resp) + if err := pre.Prerun(&req, &resp); err != nil { + return structs.WrapRecoverable(fmt.Sprintf("pre-run hook %q failed: %v", name, err), err) + } // Store the hook state { @@ -95,15 +93,10 @@ func (tr *TaskRunner) prerun() error { tr.state.Hooks[name] = hookState } - hookState.LastError = err if resp.HookData != nil { hookState.Data = resp.HookData } - if resp.DoOnce && err != nil { - hookState.SuccessfulOnce = true - } - // XXX Detect if state has changed so that we can signal to the // alloc runner precisly /* @@ -114,10 +107,6 @@ func (tr *TaskRunner) prerun() error { tr.state.Unlock() } - if err != nil { - return structs.WrapRecoverable(fmt.Sprintf("pre-run hook %q failed: %v", name, err), err) - } - // Store the environment variables returned by the hook if len(resp.Env) != 0 { tr.envBuilder.SetGenericEnv(resp.Env) @@ -283,9 +272,6 @@ func (h *artifactHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interface } } - //XXX Should this be managed by task runner directly? Seems silly to - //make every hook specify it - resp.DoOnce = true return nil } From c5504bd9393d11909f6cd4499da88c5bd125f70a Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Fri, 29 Jun 2018 16:22:18 -0700 Subject: [PATCH 009/152] tr: cleanup main loop and shutdown hook impl --- .../allocrunnerv2/taskrunner/task_runner.go | 93 ++++++++++--------- .../taskrunner/task_runner_hooks.go | 22 ++--- 2 files changed, 61 insertions(+), 54 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index be824b2dd..6d15e4874 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -15,7 +15,6 @@ import ( "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" - dstructs "github.com/hashicorp/nomad/client/driver/structs" "github.com/hashicorp/nomad/nomad/structs" ) @@ -34,12 +33,11 @@ type TaskRunner struct { state *state.State // ctx is the task runner's context and is done whe the task runner - // should exit. If a task runner is destroyed it will exit regardless - // of whether the context is done. + // should exit. Shutdown hooks are run. ctx context.Context - // ctxCancel is used to exit the task runner's Run loop (without - // stopping or destroying the task) + // ctxCancel is used to exit the task runner's Run loop without + // stopping the task. Shutdown hooks are run. ctxCancel context.CancelFunc // Logger is the logger for the task runner. @@ -56,8 +54,10 @@ type TaskRunner struct { driver driver.Driver // handle is the handle to the currently running driver - handle driver.DriverHandle - handleLock sync.Mutex + handle driver.DriverHandle + //XXX(schmichael) I think the handle is only manipulated in the Restore + // and Run loops, so there's never concurrent access. + //handleLock sync.Mutex // task is the task beign run task *structs.Task @@ -184,57 +184,69 @@ func (tr *TaskRunner) initLabels() { func (tr *TaskRunner) Run() { defer close(tr.waitCh) - var err error - var restart bool - var restartWait time.Duration - var waitRes *dstructs.WaitResult +MAIN: for { // Run the prerun hooks - if err = tr.prerun(); err != nil { + if err := tr.prerun(); err != nil { tr.logger.Error("prerun failed", "error", err) tr.restartTracker.SetStartError(err) goto RESTART } // Run the task - waitRes, err = tr.runDriver() - if err != nil { + if err := tr.runDriver(); err != nil { tr.logger.Error("running driver failed", "error", err) tr.restartTracker.SetStartError(err) goto RESTART } - tr.restartTracker.SetWaitResult(waitRes) // Run the postrun hooks - if err = tr.postrun(); err != nil { + if err := tr.postrun(); err != nil { tr.logger.Error("postrun failed", "error", err) } - // Check if the context is closed already and go straight to destroy - if err := tr.ctx.Err(); err != nil { - goto DESTROY + WAIT: + select { + case waitRes := <-tr.handle.WaitCh(): + tr.restartTracker.SetWaitResult(waitRes) + case _ = <-tr.updateCh: + //XXX Need to copy handleUpdate over + tr.logger.Warn("update not handled") + goto WAIT + case <-tr.ctx.Done(): + tr.logger.Debug("task runner cancelled") + break MAIN } RESTART: // Actually restart by sleeping and also watching for destroy events - restart, restartWait = tr.shouldRestart() - if restart { + restart, restartWait := tr.shouldRestart() + if !restart { + break + } + + deadline := time.Now().Add(restartWait) + timer := time.NewTimer(restartWait) + for time.Now().Before(deadline) { select { - case <-time.After(restartWait): - continue + case <-timer.C: + case _ = <-tr.updateCh: + //XXX Need to copy handleUpdate over + tr.logger.Warn("update not handled") case <-tr.ctx.Done(): + tr.logger.Debug("task runner cancelled") + break MAIN } } - - DESTROY: - // Run the destroy hooks - if err = tr.destroy(); err != nil { - tr.logger.Error("postrun failed", "error", err) - } - - tr.logger.Debug("task run loop exiting") - return + timer.Stop() } + + // Run the shutdown hooks + if err := tr.shutdown(); err != nil { + tr.logger.Error("postrun failed", "error", err) + } + + tr.logger.Debug("task run loop exiting") } func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { @@ -259,13 +271,13 @@ func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { } // runDriver runs the driver and waits for it to exit -func (tr *TaskRunner) runDriver() (*dstructs.WaitResult, error) { +func (tr *TaskRunner) runDriver() error { // Run prestart ctx := driver.NewExecContext(tr.taskDir, tr.envBuilder.Build()) _, err := tr.driver.Prestart(ctx, tr.task) if err != nil { tr.logger.Error("driver pre-start failed", "error", err) - return nil, err + return err } // Create a new context for Start since the environment may have been updated. @@ -275,21 +287,16 @@ func (tr *TaskRunner) runDriver() (*dstructs.WaitResult, error) { sresp, err := tr.driver.Start(ctx, tr.task) if err != nil { tr.logger.Warn("driver start failed", "error", err) - return nil, err + return err } // Wait on the handle - tr.handleLock.Lock() - handle := sresp.Handle - tr.handle = handle - tr.handleLock.Unlock() + tr.handle = sresp.Handle + //XXX need to capture the driver network // Emit an event that we started tr.SetState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) - - // Wait for the task to exit - waitRes := <-handle.WaitCh() - return waitRes, nil + return nil } // initDriver creates the driver for the task diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 3be92e94c..66863a0cc 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -54,12 +54,6 @@ func (tr *TaskRunner) prerun() error { } name := pre.Name() - var start time.Time - if tr.logger.IsTrace() { - start = time.Now() - tr.logger.Trace("running pre-run hook", "name", name, "start", start) - } - // Build the request req := interfaces.TaskPrerunRequest{ Task: tr.Task(), @@ -78,6 +72,13 @@ func (tr *TaskRunner) prerun() error { req.VaultToken = tr.state.VaultToken tr.state.RUnlock() + // Time the prerun hook + var start time.Time + if tr.logger.IsTrace() { + start = time.Now() + tr.logger.Trace("running pre-run hook", "name", name, "start", start) + } + // Run the pre-run hook var resp interfaces.TaskPrerunResponse if err := pre.Prerun(&req, &resp); err != nil { @@ -159,15 +160,14 @@ func (tr *TaskRunner) postrun() error { return nil } -// destroy is used to run the runners destroy hooks. -// XXX Naming change -func (tr *TaskRunner) destroy() error { +// shutdown is used to run the shutdown hooks. +func (tr *TaskRunner) shutdown() error { if tr.logger.IsTrace() { start := time.Now() - tr.logger.Trace("running destroy hooks", "start", start) + tr.logger.Trace("running poststop hooks", "start", start) defer func() { end := time.Now() - tr.logger.Trace("finished destroy hooks", "end", end, "duration", end.Sub(start)) + tr.logger.Trace("finished poststop hooks", "end", end, "duration", end.Sub(start)) }() } From 582c76a420d4d6c24e24a17adfec77e858455d23 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 10 Jul 2018 21:19:58 -0700 Subject: [PATCH 010/152] remove unused allocrunner shim --- client/allocrunnerv2/tr_shim.go | 26 -------------------------- 1 file changed, 26 deletions(-) delete mode 100644 client/allocrunnerv2/tr_shim.go diff --git a/client/allocrunnerv2/tr_shim.go b/client/allocrunnerv2/tr_shim.go deleted file mode 100644 index b20246ea1..000000000 --- a/client/allocrunnerv2/tr_shim.go +++ /dev/null @@ -1,26 +0,0 @@ -package allocrunnerv2 - -import ( - "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunnerv2/state" - trstate "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" -) - -// allocRunnerShim is used to pass the alloc runner to the task runner, exposing -// private methods as public fields. -type allocRunnerShim struct { - *allocRunner -} - -func (a *allocRunnerShim) State() *state.State { - return a.state -} - -func (a *allocRunnerShim) GetAllocDir() *allocdir.AllocDir { - return a.allocDir -} - -func (a *allocRunnerShim) StateUpdated(trState *trstate.State) error { - // XXX implement - return nil -} From 4f43ff5c51e83c40821cf892c52f8bc9d8882bd9 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 10 Jul 2018 21:21:12 -0700 Subject: [PATCH 011/152] pass statedb into allocrunnerv2 --- client/allocrunnerv2/alloc_runner.go | 6 ++++++ client/client.go | 4 ++-- 2 files changed, 8 insertions(+), 2 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 60a6fce30..533d036ee 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -5,6 +5,7 @@ import ( "path/filepath" "sync" + "github.com/boltdb/bolt" log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner" @@ -31,9 +32,12 @@ type allocRunner struct { alloc *structs.Allocation allocLock sync.RWMutex + //XXX implement for local state // state captures the state of the alloc runner state *state.State + stateDB *bolt.DB + // allocDir is used to build the allocations directory structure. allocDir *allocdir.AllocDir @@ -56,6 +60,7 @@ func NewAllocRunner(config *Config) *allocRunner { tasks: make(map[string]*taskrunner.TaskRunner), waitCh: make(chan struct{}), updateCh: make(chan *structs.Allocation), + stateDB: config.StateDB, } // Create alloc dir @@ -158,6 +163,7 @@ func (ar *allocRunner) runTask(alloc *structs.Allocation, task *structs.Task) er Task: task, TaskDir: ar.allocDir.NewTaskDir(task.Name), Logger: ar.logger, + StateDB: ar.stateDB, } tr, err := taskrunner.NewTaskRunner(config) if err != nil { diff --git a/client/client.go b/client/client.go index 2c690fbb5..64e4e1d00 100644 --- a/client/client.go +++ b/client/client.go @@ -1926,10 +1926,10 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error c.configLock.RLock() arConf := &allocrunnerv2.Config{ + Alloc: alloc, Logger: logger, ClientConfig: c.config, - Alloc: alloc, - StateDB: nil, //FIXME ? + StateDB: c.stateDB, } c.configLock.RUnlock() From ae89b7da958fe2e375e2b51871a526b4d17da8f9 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 10 Jul 2018 21:22:04 -0700 Subject: [PATCH 012/152] reimplement success state for tr hooks and state persistence splits apart local and remote persistence removes some locking *for now* --- .../interfaces/task_lifecycle.go | 4 +- client/allocrunnerv2/state/state.go | 4 - .../allocrunnerv2/taskrunner/state/state.go | 47 ++++++-- .../allocrunnerv2/taskrunner/task_runner.go | 114 +++++++++++++++--- .../taskrunner/task_runner_hooks.go | 27 ++--- helper/funcs.go | 32 +++++ 6 files changed, 174 insertions(+), 54 deletions(-) diff --git a/client/allocrunnerv2/interfaces/task_lifecycle.go b/client/allocrunnerv2/interfaces/task_lifecycle.go index d5e1c28a9..f8ef85664 100644 --- a/client/allocrunnerv2/interfaces/task_lifecycle.go +++ b/client/allocrunnerv2/interfaces/task_lifecycle.go @@ -43,8 +43,8 @@ type TaskPrerunResponse struct { // run HookData map[string]string - // DoOnce lets the hook indicate that it should only be run once - DoOnce bool + // Done lets the hook indicate that it should only be run once + Done bool } type TaskPrerunHook interface { diff --git a/client/allocrunnerv2/state/state.go b/client/allocrunnerv2/state/state.go index 4105e2656..eee301de6 100644 --- a/client/allocrunnerv2/state/state.go +++ b/client/allocrunnerv2/state/state.go @@ -3,7 +3,6 @@ package state import ( "sync" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" "github.com/hashicorp/nomad/nomad/structs" ) @@ -18,9 +17,6 @@ type State struct { // allocations client state ClientDesc string - // TaskStates captures the state of individual tasks. - TaskStates map[string]*state.State - // DeploymentStatus captures the status of the deployment DeploymentStatus *structs.AllocDeploymentStatus } diff --git a/client/allocrunnerv2/taskrunner/state/state.go b/client/allocrunnerv2/taskrunner/state/state.go index 5c079a9e1..bef07235f 100644 --- a/client/allocrunnerv2/taskrunner/state/state.go +++ b/client/allocrunnerv2/taskrunner/state/state.go @@ -1,28 +1,36 @@ package state import ( - "sync" - + "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/helper" - "github.com/hashicorp/nomad/nomad/structs" ) -type State struct { - sync.RWMutex - Task *structs.TaskState +// LocalState is Task state which is persisted for use when restarting Nomad +// agents. +type LocalState struct { Hooks map[string]*HookState // VaultToken is the current Vault token for the task VaultToken string + + // DriverNetwork is the network information returned by the task + // driver's Start method + DriverNetwork *structs.DriverNetwork +} + +func NewLocalState() *LocalState { + return &LocalState{ + Hooks: make(map[string]*HookState), + } } // Copy should be called with the lock held -func (s *State) Copy() *State { +func (s *LocalState) Copy() *LocalState { // Create a copy - c := &State{ - Task: s.Task.Copy(), - Hooks: make(map[string]*HookState, len(s.Hooks)), - VaultToken: s.VaultToken, + c := &LocalState{ + Hooks: make(map[string]*HookState, len(s.Hooks)), + VaultToken: s.VaultToken, + DriverNetwork: s.DriverNetwork, } // Copy the hooks @@ -34,7 +42,10 @@ func (s *State) Copy() *State { } type HookState struct { - Data map[string]string + // PrerunDone is true if the hook has run Prerun successfully and does + // not need to run again + PrerunDone bool + Data map[string]string } func (h *HookState) Copy() *HookState { @@ -43,3 +54,15 @@ func (h *HookState) Copy() *HookState { c.Data = helper.CopyMapStringString(c.Data) return c } + +func (h *HookState) Equal(o *HookState) bool { + if h == nil || o == nil { + return h == o + } + + if h.PrerunDone != o.PrerunDone { + return false + } + + return helper.CompareMapStringString(h.Data, o.Data) +} diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 6d15e4874..5f24ab0ac 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -1,12 +1,15 @@ package taskrunner import ( + "bytes" "context" "fmt" + "io" "sync" "time" metrics "github.com/armon/go-metrics" + "github.com/boltdb/bolt" log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" @@ -15,7 +18,16 @@ import ( "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" + oldstate "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/nomad/structs" + "github.com/ugorji/go/codec" + "golang.org/x/crypto/blake2b" +) + +var ( + // taskRunnerStateAllKey holds all the task runners state. At the moment + // there is no need to split it + taskRunnerStateAllKey = []byte("simple-all") ) type TaskRunner struct { @@ -29,8 +41,19 @@ type TaskRunner struct { clientConfig *config.Config - // state captures the state of the task runner - state *state.State + // state captures the state of the task for updating the allocation + state *structs.TaskState + + // localState captures the node-local state of the task for when the + // Nomad agent restarts + localState *state.LocalState + + // stateDB is for persisting localState + stateDB *bolt.DB + + // persistedHash is the hash of the last persisted state for skipping + // unnecessary writes + persistedHash []byte // ctx is the task runner's context and is done whe the task runner // should exit. Shutdown hooks are run. @@ -59,7 +82,7 @@ type TaskRunner struct { // and Run loops, so there's never concurrent access. //handleLock sync.Mutex - // task is the task beign run + // task is the task being run task *structs.Task taskLock sync.RWMutex @@ -88,8 +111,11 @@ type Config struct { TaskDir *allocdir.TaskDir Logger log.Logger - // State is optionally restored task state - State *state.State + // LocalState is optionally restored task state + LocalState *state.LocalState + + // StateDB is used to store and restore state. + StateDB *bolt.DB } func NewTaskRunner(config *Config) (*TaskRunner, error) { @@ -112,11 +138,14 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { taskDir: config.TaskDir, taskName: config.Task.Name, envBuilder: envBuilder, - state: config.State, - ctx: trCtx, - ctxCancel: trCancel, - updateCh: make(chan *structs.Allocation), - waitCh: make(chan struct{}), + //XXX Make a Copy to avoid races? + state: config.Alloc.TaskStates[config.Task.Name], + localState: config.LocalState, + stateDB: config.StateDB, + ctx: trCtx, + ctxCancel: trCancel, + updateCh: make(chan *structs.Allocation), + waitCh: make(chan struct{}), } // Create the logger based on the allocation ID @@ -150,13 +179,13 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { func (tr *TaskRunner) initState() { if tr.state == nil { - tr.state = &state.State{ - Task: &structs.TaskState{ - State: structs.TaskStatePending, - }, - Hooks: make(map[string]*state.HookState), + tr.state = &structs.TaskState{ + State: structs.TaskStatePending, } } + if tr.localState == nil { + tr.localState = state.NewLocalState() + } } func (tr *TaskRunner) initLabels() { @@ -329,15 +358,60 @@ func (tr *TaskRunner) initDriver() error { return nil } -// SetState sets the task runners state. +// persistLocalState persists local state to disk synchronously. +func (tr *TaskRunner) persistLocalState() error { + // buffer for writing to boltdb + var buf bytes.Buffer + + // Hash for skipping unnecessary writes + h, err := blake2b.New256(nil) + if err != nil { + // Programming error that should never happen! + return err + } + + // Multiplex writes to both + w := io.MultiWriter(h, &buf) + + // Encode as msgpack value + if err := codec.NewEncoder(w, structs.MsgpackHandle).Encode(&tr.localState); err != nil { + return fmt.Errorf("failed to serialize snapshot: %v", err) + } + + // If the hashes are equal, skip the write + hashVal := h.Sum(nil) + if bytes.Equal(hashVal, tr.persistedHash) { + return nil + } + + // Start the transaction. + return tr.stateDB.Batch(func(tx *bolt.Tx) error { + // Grab the task bucket + //XXX move into new state pkg + taskBkt, err := oldstate.GetTaskBucket(tx, tr.allocID, tr.taskName) + if err != nil { + return fmt.Errorf("failed to retrieve allocation bucket: %v", err) + } + + if err := oldstate.PutData(taskBkt, taskRunnerStateAllKey, buf.Bytes()); err != nil { + return fmt.Errorf("failed to write task_runner state: %v", err) + } + + // Store the hash that was persisted + tx.OnCommit(func() { + tr.persistedHash = hashVal + }) + + return nil + }) +} + +// SetState sets the task runners allocation state. func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { // Ensure the event is populated with human readable strings event.PopulateEventDisplayMessage() - // Lock our state - tr.state.Lock() - defer tr.state.Unlock() - task := tr.state.Task + task := tr.state // Update the state of the task if state != "" { diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 66863a0cc..ead99c8c5 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -61,16 +61,13 @@ func (tr *TaskRunner) prerun() error { TaskEnv: tr.envBuilder.Build(), } - tr.state.RLock() - hookState := tr.state.Hooks[name] - if hookState != nil { + origHookState := tr.localState.Hooks[name] + if origHookState != nil && origHookState.PrerunDone { // Hook already ran, skip - tr.state.RUnlock() continue } - req.VaultToken = tr.state.VaultToken - tr.state.RUnlock() + req.VaultToken = tr.localState.VaultToken // Time the prerun hook var start time.Time @@ -87,25 +84,23 @@ func (tr *TaskRunner) prerun() error { // Store the hook state { - tr.state.Lock() - hookState, ok := tr.state.Hooks[name] + hookState, ok := tr.localState.Hooks[name] if !ok { hookState = &state.HookState{} - tr.state.Hooks[name] = hookState + tr.localState.Hooks[name] = hookState } if resp.HookData != nil { hookState.Data = resp.HookData + hookState.PrerunDone = resp.Done } - // XXX Detect if state has changed so that we can signal to the - // alloc runner precisly - /* - if err := tr.allocRunner.StateUpdated(tr.state.Copy()); err != nil { - tr.logger.Error("failed to save state", "error", err) + // Persist local state if the hook state has changed + if !hookState.Equal(origHookState) { + if err := tr.persistLocalState(); err != nil { + return err } - */ - tr.state.Unlock() + } } // Store the environment variables returned by the hook diff --git a/helper/funcs.go b/helper/funcs.go index 083ab865b..89027b4cb 100644 --- a/helper/funcs.go +++ b/helper/funcs.go @@ -172,6 +172,38 @@ func SliceSetDisjoint(first, second []string) (bool, []string) { return false, flattened } +// CompareMapStringString returns true if the maps are equivalent. A nil and +// empty map are considered not equal. +func CompareMapStringString(a, b map[string]string) bool { + if a == nil || b == nil { + return a == nil && b == nil + } + + if len(a) != len(b) { + return false + } + + for k, v := range a { + v2, ok := b[k] + if !ok { + return false + } + if v != v2 { + return false + } + } + + // Already compared all known values in a so only test that keys from b + // exist in a + for k := range b { + if _, ok := a[k]; !ok { + return false + } + } + + return true +} + // Helpers for copying generic structures. func CopyMapStringString(m map[string]string) map[string]string { l := len(m) From b360f6f96ee84df08369c285b6a891cc80655894 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 11 Jul 2018 15:34:18 -0700 Subject: [PATCH 013/152] fix hclog level --- client/client.go | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/client/client.go b/client/client.go index 64e4e1d00..edbcd33cd 100644 --- a/client/client.go +++ b/client/client.go @@ -1921,7 +1921,7 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error //XXX FIXME logger := hclog.New(&hclog.LoggerOptions{ Name: "nomad", - Level: hclog.LevelFromString("DEBUG"), + Level: hclog.LevelFromString(c.configCopy.LogLevel), }) c.configLock.RLock() From 3a4f387fd3ec90ccefb123f36e1f178da156e8ca Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 11 Jul 2018 15:34:31 -0700 Subject: [PATCH 014/152] tr: fix setting done in existing hooks --- client/allocrunnerv2/taskrunner/task_runner_hooks.go | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index ead99c8c5..d3cc170b6 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -50,6 +50,7 @@ func (tr *TaskRunner) prerun() error { for _, hook := range tr.runnerHooks { pre, ok := hook.(interfaces.TaskPrerunHook) if !ok { + tr.logger.Trace("skipping: not a prerun hook", "name", hook.Name()) continue } @@ -63,7 +64,7 @@ func (tr *TaskRunner) prerun() error { origHookState := tr.localState.Hooks[name] if origHookState != nil && origHookState.PrerunDone { - // Hook already ran, skip + tr.logger.Trace("skipping: prerun hook already done", "name", pre.Name()) continue } @@ -229,6 +230,7 @@ func (h *taskDirHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interfaces // Update the environment variables based on the built task directory driver.SetEnvvars(h.runner.envBuilder, fsi, h.runner.taskDir, h.runner.clientConfig) + resp.Done = true return nil } @@ -267,6 +269,7 @@ func (h *artifactHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interface } } + resp.Done = true return nil } From ef213b864bdfa6138e4326c1dd7b229c889a3f6a Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 11 Jul 2018 15:42:35 -0700 Subject: [PATCH 015/152] tr: pass context to hooks --- client/allocrunnerv2/interfaces/task_lifecycle.go | 12 +++++++----- client/allocrunnerv2/taskrunner/task_runner_hooks.go | 8 +++++--- 2 files changed, 12 insertions(+), 8 deletions(-) diff --git a/client/allocrunnerv2/interfaces/task_lifecycle.go b/client/allocrunnerv2/interfaces/task_lifecycle.go index f8ef85664..5876dd236 100644 --- a/client/allocrunnerv2/interfaces/task_lifecycle.go +++ b/client/allocrunnerv2/interfaces/task_lifecycle.go @@ -1,6 +1,8 @@ package interfaces import ( + "context" + "github.com/hashicorp/nomad/client/driver/env" "github.com/hashicorp/nomad/nomad/structs" ) @@ -49,7 +51,7 @@ type TaskPrerunResponse struct { type TaskPrerunHook interface { TaskHook - Prerun(*TaskPrerunRequest, *TaskPrerunResponse) error + Prerun(context.Context, *TaskPrerunRequest, *TaskPrerunResponse) error } // XXX If we want consul style hooks, need to have something that runs after the @@ -62,7 +64,7 @@ type TaskPostrunResponse struct{} type TaskPostrunHook interface { TaskHook Postrun() error - //Postrun(*TaskPostrunRequest, *TaskPostrunResponse) error + //Postrun(context.Context, *TaskPostrunRequest, *TaskPostrunResponse) error } type TaskPoststopRequest struct{} @@ -71,7 +73,7 @@ type TaskPoststopResponse struct{} type TaskPoststopHook interface { TaskHook Postrun() error - //Postrun(*TaskPostrunRequest, *TaskPostrunResponse) error + //Postrun(context.Context, *TaskPostrunRequest, *TaskPostrunResponse) error } type TaskDestroyRequest struct{} @@ -80,7 +82,7 @@ type TaskDestroyResponse struct{} type TaskDestroyHook interface { TaskHook Destroy() error - //Destroy(*TaskDestroyRequest, *TaskDestroyResponse) error + //Destroy(context.Context, *TaskDestroyRequest, *TaskDestroyResponse) error } type TaskUpdateRequest struct { @@ -92,5 +94,5 @@ type TaskUpdateResponse struct{} type TaskUpdateHook interface { TaskHook Update() error - //Update(*TaskUpdateRequest, *TaskUpdateResponse) error + //Update(context.Context, *TaskUpdateRequest, *TaskUpdateResponse) error } diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index d3cc170b6..e0f25c5ee 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -1,6 +1,7 @@ package taskrunner import ( + "context" "fmt" "time" @@ -79,7 +80,7 @@ func (tr *TaskRunner) prerun() error { // Run the pre-run hook var resp interfaces.TaskPrerunResponse - if err := pre.Prerun(&req, &resp); err != nil { + if err := pre.Prerun(tr.ctx, &req, &resp); err != nil { return structs.WrapRecoverable(fmt.Sprintf("pre-run hook %q failed: %v", name, err), err) } @@ -211,7 +212,7 @@ func (h *taskDirHook) Name() string { return "task_dir" } -func (h *taskDirHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { +func (h *taskDirHook) Prerun(ctx context.Context, req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { cc := h.runner.clientConfig chroot := cconfig.DefaultChrootEnv if len(cc.ChrootEnv) > 0 { @@ -256,10 +257,11 @@ func (*artifactHook) Name() string { return "artifacts" } -func (h *artifactHook) Prerun(req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { +func (h *artifactHook) Prerun(ctx context.Context, req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { h.eventEmitter.SetState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskDownloadingArtifacts)) for _, artifact := range req.Task.Artifacts { + //XXX add ctx to GetArtifact to allow cancelling long downloads if err := getter.GetArtifact(req.TaskEnv, artifact, req.TaskDir); err != nil { wrapped := fmt.Errorf("failed to download artifact %q: %v", artifact.GetterSource, err) h.logger.Debug(wrapped.Error()) From 30d377eba49621672c85df36cc8e9e4850e4a281 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 11 Jul 2018 16:02:42 -0700 Subject: [PATCH 016/152] tr: improve skip log line --- client/allocrunnerv2/taskrunner/task_runner_hooks.go | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index e0f25c5ee..4792103dd 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -51,7 +51,7 @@ func (tr *TaskRunner) prerun() error { for _, hook := range tr.runnerHooks { pre, ok := hook.(interfaces.TaskPrerunHook) if !ok { - tr.logger.Trace("skipping: not a prerun hook", "name", hook.Name()) + tr.logger.Trace("skipping non-prerun hook", "name", hook.Name()) continue } @@ -65,7 +65,7 @@ func (tr *TaskRunner) prerun() error { origHookState := tr.localState.Hooks[name] if origHookState != nil && origHookState.PrerunDone { - tr.logger.Trace("skipping: prerun hook already done", "name", pre.Name()) + tr.logger.Trace("skipping done prerun hook", "name", pre.Name()) continue } From 80f6ce50c04834937f3f6c511f3f1feccf4c802d Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Thu, 12 Jul 2018 16:15:33 -0700 Subject: [PATCH 017/152] vault hook --- client/allocrunnerv2/alloc_runner.go | 18 +- client/allocrunnerv2/alloc_runner_test.go | 13 +- client/allocrunnerv2/config.go | 4 + .../interfaces/task_lifecycle.go | 9 +- client/allocrunnerv2/taskrunner/lifecycle.go | 24 ++ .../allocrunnerv2/taskrunner/state/state.go | 4 - .../allocrunnerv2/taskrunner/task_runner.go | 27 +- .../taskrunner/task_runner_getters.go | 18 + .../taskrunner/task_runner_hooks.go | 73 +++- client/allocrunnerv2/taskrunner/vault_hook.go | 390 ++++++++++++++++++ client/client.go | 1 + 11 files changed, 539 insertions(+), 42 deletions(-) create mode 100644 client/allocrunnerv2/taskrunner/lifecycle.go create mode 100644 client/allocrunnerv2/taskrunner/vault_hook.go diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 533d036ee..fe1d2fdae 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -12,7 +12,6 @@ import ( "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" "github.com/hashicorp/nomad/client/allocrunnerv2/state" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" - "github.com/hashicorp/nomad/client/config" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" ) @@ -22,7 +21,7 @@ type allocRunner struct { // Logger is the logger for the alloc runner. logger log.Logger - clientConfig *config.Config + config *Config // waitCh is closed when the alloc runner has transitioned to a terminal // state @@ -55,12 +54,12 @@ type allocRunner struct { // NewAllocRunner returns a new allocation runner. func NewAllocRunner(config *Config) *allocRunner { ar := &allocRunner{ - alloc: config.Alloc, - clientConfig: config.ClientConfig, - tasks: make(map[string]*taskrunner.TaskRunner), - waitCh: make(chan struct{}), - updateCh: make(chan *structs.Allocation), - stateDB: config.StateDB, + config: config, + alloc: config.Alloc, + tasks: make(map[string]*taskrunner.TaskRunner), + waitCh: make(chan struct{}), + updateCh: make(chan *structs.Allocation), + stateDB: config.StateDB, } // Create alloc dir @@ -159,11 +158,12 @@ func (ar *allocRunner) runTask(alloc *structs.Allocation, task *structs.Task) er // Create the runner config := &taskrunner.Config{ Alloc: alloc, - ClientConfig: ar.clientConfig, + ClientConfig: ar.config.ClientConfig, Task: task, TaskDir: ar.allocDir.NewTaskDir(task.Name), Logger: ar.logger, StateDB: ar.stateDB, + VaultClient: ar.config.Vault, } tr, err := taskrunner.NewTaskRunner(config) if err != nil { diff --git a/client/allocrunnerv2/alloc_runner_test.go b/client/allocrunnerv2/alloc_runner_test.go index b7c8ee282..af2e5fe10 100644 --- a/client/allocrunnerv2/alloc_runner_test.go +++ b/client/allocrunnerv2/alloc_runner_test.go @@ -1,10 +1,8 @@ package allocrunnerv2 import ( - "context" "testing" - "github.com/hashicorp/nomad/client/allocrunnerv2/config" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" clientconfig "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/helper/testlog" @@ -15,19 +13,14 @@ import ( func testAllocRunnerFromAlloc(t *testing.T, alloc *structs.Allocation) *allocRunner { cconf := clientconfig.DefaultConfig() - config := &config.Config{ + config := &Config{ ClientConfig: cconf, Logger: testlog.HCLogger(t).With("unit_test", t.Name()), - Allocation: alloc, - } - - ar, err := NewAllocRunner(context.Background(), config) - if err != nil { - t.Fatalf("Failed to create test alloc runner: %v", err) + Alloc: alloc, } + ar := NewAllocRunner(config) return ar - } func testAllocRunner(t *testing.T) *allocRunner { diff --git a/client/allocrunnerv2/config.go b/client/allocrunnerv2/config.go index 11bcd3063..9e6395158 100644 --- a/client/allocrunnerv2/config.go +++ b/client/allocrunnerv2/config.go @@ -4,6 +4,7 @@ import ( "github.com/boltdb/bolt" log "github.com/hashicorp/go-hclog" clientconfig "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" ) @@ -21,6 +22,9 @@ type Config struct { // StateDB is used to store and restore state. StateDB *bolt.DB + // Vault is the Vault client to use to retrieve Vault tokens + Vault vaultclient.VaultClient + // XXX Can have a OnStateTransistion hook that we can use to update the // server } diff --git a/client/allocrunnerv2/interfaces/task_lifecycle.go b/client/allocrunnerv2/interfaces/task_lifecycle.go index 5876dd236..ba81544be 100644 --- a/client/allocrunnerv2/interfaces/task_lifecycle.go +++ b/client/allocrunnerv2/interfaces/task_lifecycle.go @@ -72,8 +72,7 @@ type TaskPoststopResponse struct{} type TaskPoststopHook interface { TaskHook - Postrun() error - //Postrun(context.Context, *TaskPostrunRequest, *TaskPostrunResponse) error + Postrun(context.Context, *TaskPostrunRequest, *TaskPostrunResponse) error } type TaskDestroyRequest struct{} @@ -86,13 +85,11 @@ type TaskDestroyHook interface { } type TaskUpdateRequest struct { - Alloc string - Vault string // Don't need message bus then + VaultToken string } type TaskUpdateResponse struct{} type TaskUpdateHook interface { TaskHook - Update() error - //Update(context.Context, *TaskUpdateRequest, *TaskUpdateResponse) error + Update(context.Context, *TaskUpdateRequest, *TaskUpdateResponse) error } diff --git a/client/allocrunnerv2/taskrunner/lifecycle.go b/client/allocrunnerv2/taskrunner/lifecycle.go new file mode 100644 index 000000000..e93fbf065 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/lifecycle.go @@ -0,0 +1,24 @@ +package taskrunner + +import "os" + +// XXX These should probably all return an error and we should have predefined +// error types for the task not currently running +type TaskLifecycle interface { + Restart(source, reason string, failure bool) + Signal(source, reason string, s os.Signal) error + Kill(source, reason string, fail bool) +} + +func (tr *TaskRunner) Restart(source, reason string, failure bool) { + // TODO +} + +func (tr *TaskRunner) Signal(source, reason string, s os.Signal) error { + // TODO + return nil +} + +func (tr *TaskRunner) Kill(source, reason string, fail bool) { + // TODO +} diff --git a/client/allocrunnerv2/taskrunner/state/state.go b/client/allocrunnerv2/taskrunner/state/state.go index bef07235f..8abb71906 100644 --- a/client/allocrunnerv2/taskrunner/state/state.go +++ b/client/allocrunnerv2/taskrunner/state/state.go @@ -10,9 +10,6 @@ import ( type LocalState struct { Hooks map[string]*HookState - // VaultToken is the current Vault token for the task - VaultToken string - // DriverNetwork is the network information returned by the task // driver's Start method DriverNetwork *structs.DriverNetwork @@ -29,7 +26,6 @@ func (s *LocalState) Copy() *LocalState { // Create a copy c := &LocalState{ Hooks: make(map[string]*HookState, len(s.Hooks)), - VaultToken: s.VaultToken, DriverNetwork: s.DriverNetwork, } diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 5f24ab0ac..708a63df3 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -19,6 +19,7 @@ import ( "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" oldstate "github.com/hashicorp/nomad/client/state" + "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" "github.com/ugorji/go/codec" "golang.org/x/crypto/blake2b" @@ -46,7 +47,8 @@ type TaskRunner struct { // localState captures the node-local state of the task for when the // Nomad agent restarts - localState *state.LocalState + localState *state.LocalState + localStateLock sync.RWMutex // stateDB is for persisting localState stateDB *bolt.DB @@ -99,6 +101,14 @@ type TaskRunner struct { // transistions. runnerHooks []interfaces.TaskHook + // vaultClient is the client to use to derive and renew Vault tokens + vaultClient vaultclient.VaultClient + + // vaultToken is the current Vault token. It should be accessed with the + // getter. + vaultToken string + vaultTokenLock sync.Mutex + // baseLabels are used when emitting tagged metrics. All task runner metrics // will have these tags, and optionally more. baseLabels []metrics.Label @@ -111,6 +121,9 @@ type Config struct { TaskDir *allocdir.TaskDir Logger log.Logger + // VaultClient is the client to use to derive and renew Vault tokens + VaultClient vaultclient.VaultClient + // LocalState is optionally restored task state LocalState *state.LocalState @@ -138,6 +151,7 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { taskDir: config.TaskDir, taskName: config.Task.Name, envBuilder: envBuilder, + vaultClient: config.VaultClient, //XXX Make a Copy to avoid races? state: config.Alloc.TaskStates[config.Task.Name], localState: config.LocalState, @@ -374,7 +388,10 @@ func (tr *TaskRunner) persistLocalState() error { w := io.MultiWriter(h, &buf) // Encode as msgpack value - if err := codec.NewEncoder(w, structs.MsgpackHandle).Encode(&tr.localState); err != nil { + tr.localStateLock.Lock() + err = codec.NewEncoder(w, structs.MsgpackHandle).Encode(&tr.localState) + tr.localStateLock.Unlock() + if err != nil { return fmt.Errorf("failed to serialize snapshot: %v", err) } @@ -509,12 +526,6 @@ func (tr *TaskRunner) Shutdown() { tr.ctxCancel() } -func (tr *TaskRunner) Alloc() *structs.Allocation { - tr.allocLock.Lock() - defer tr.allocLock.Unlock() - return tr.alloc -} - // appendTaskEvent updates the task status by appending the new event. func appendTaskEvent(state *structs.TaskState, event *structs.TaskEvent) { capacity := 10 diff --git a/client/allocrunnerv2/taskrunner/task_runner_getters.go b/client/allocrunnerv2/taskrunner/task_runner_getters.go index dd0a5ac20..455cec467 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_getters.go +++ b/client/allocrunnerv2/taskrunner/task_runner_getters.go @@ -2,8 +2,26 @@ package taskrunner import "github.com/hashicorp/nomad/nomad/structs" +func (tr *TaskRunner) Alloc() *structs.Allocation { + tr.allocLock.Lock() + defer tr.allocLock.Unlock() + return tr.alloc +} + func (tr *TaskRunner) Task() *structs.Task { tr.taskLock.RLock() defer tr.taskLock.RUnlock() return tr.task } + +func (tr *TaskRunner) getVaultToken() string { + tr.vaultTokenLock.Lock() + defer tr.vaultTokenLock.Unlock() + return tr.vaultToken +} + +func (tr *TaskRunner) setVaultToken(token string) { + tr.vaultTokenLock.Lock() + defer tr.vaultTokenLock.Unlock() + tr.vaultToken = token +} diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 4792103dd..cb2382f8d 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -15,6 +15,10 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) +type EventEmitter interface { + SetState(state string, event *structs.TaskEvent) +} + // initHooks intializes the tasks hooks. func (tr *TaskRunner) initHooks() { hookLogger := tr.logger.Named("task_hook") @@ -25,6 +29,20 @@ func (tr *TaskRunner) initHooks() { newTaskDirHook(tr, hookLogger), newArtifactHook(tr, hookLogger), } + + // If Vault is enabled, add the hook + if task := tr.Task(); task.Vault != nil { + tr.runnerHooks = append(tr.runnerHooks, newVaultHook(&vaultHookConfig{ + vaultStanza: task.Vault, + client: tr.vaultClient, + events: tr, + lifecycle: tr, + updater: tr, + logger: hookLogger, + alloc: tr.Alloc(), + task: tr.taskName, + })) + } } // prerun is used to run the runners prerun hooks. @@ -63,13 +81,15 @@ func (tr *TaskRunner) prerun() error { TaskEnv: tr.envBuilder.Build(), } + tr.localStateLock.RLock() origHookState := tr.localState.Hooks[name] + tr.localStateLock.RUnlock() if origHookState != nil && origHookState.PrerunDone { tr.logger.Trace("skipping done prerun hook", "name", pre.Name()) continue } - req.VaultToken = tr.localState.VaultToken + req.VaultToken = tr.getVaultToken() // Time the prerun hook var start time.Time @@ -86,6 +106,7 @@ func (tr *TaskRunner) prerun() error { // Store the hook state { + tr.localStateLock.Lock() hookState, ok := tr.localState.Hooks[name] if !ok { hookState = &state.HookState{} @@ -96,6 +117,7 @@ func (tr *TaskRunner) prerun() error { hookState.Data = resp.HookData hookState.PrerunDone = resp.Done } + tr.localStateLock.Unlock() // Persist local state if the hook state has changed if !hookState.Equal(origHookState) { @@ -195,6 +217,51 @@ func (tr *TaskRunner) shutdown() error { return nil } +// update is used to run the runners update hooks. +func (tr *TaskRunner) updateHooks() { + if tr.logger.IsTrace() { + start := time.Now() + tr.logger.Trace("running update hooks", "start", start) + defer func() { + end := time.Now() + tr.logger.Trace("finished update hooks", "end", end, "duration", end.Sub(start)) + }() + } + + for _, hook := range tr.runnerHooks { + upd, ok := hook.(interfaces.TaskUpdateHook) + if !ok { + tr.logger.Trace("skipping non-update hook", "name", hook.Name()) + continue + } + + name := upd.Name() + + // Build the request + req := interfaces.TaskUpdateRequest{ + VaultToken: tr.getVaultToken(), + } + + // Time the prerun hook + var start time.Time + if tr.logger.IsTrace() { + start = time.Now() + tr.logger.Trace("running update hook", "name", name, "start", start) + } + + // Run the update hook + var resp interfaces.TaskUpdateResponse + if err := upd.Update(tr.ctx, &req, &resp); err != nil { + tr.logger.Error("update hook failed", "name", name, "error", err) + } + + if tr.logger.IsTrace() { + end := time.Now() + tr.logger.Trace("finished update hooks", "name", name, "end", end, "duration", end.Sub(start)) + } + } +} + type taskDirHook struct { runner *TaskRunner logger log.Logger @@ -235,10 +302,6 @@ func (h *taskDirHook) Prerun(ctx context.Context, req *interfaces.TaskPrerunRequ return nil } -type EventEmitter interface { - SetState(state string, event *structs.TaskEvent) -} - // artifactHook downloads artifacts for a task. type artifactHook struct { eventEmitter EventEmitter diff --git a/client/allocrunnerv2/taskrunner/vault_hook.go b/client/allocrunnerv2/taskrunner/vault_hook.go new file mode 100644 index 000000000..6cb3be915 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/vault_hook.go @@ -0,0 +1,390 @@ +package taskrunner + +import ( + "context" + "fmt" + "io/ioutil" + "os" + "path/filepath" + "sync" + "time" + + "github.com/hashicorp/consul-template/signals" + log "github.com/hashicorp/go-hclog" + + "github.com/hashicorp/nomad/client/allocdir" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/vaultclient" + "github.com/hashicorp/nomad/nomad/structs" +) + +const ( + // vaultBackoffBaseline is the baseline time for exponential backoff when + // attempting to retrieve a Vault token + vaultBackoffBaseline = 5 * time.Second + + // vaultBackoffLimit is the limit of the exponential backoff when attempting + // to retrieve a Vault token + vaultBackoffLimit = 3 * time.Minute + + // vaultTokenFile is the name of the file holding the Vault token inside the + // task's secret directory + vaultTokenFile = "vault_token" +) + +type vaultTokenUpdateHandler interface { + updatedVaultToken(token string) +} + +func (tr *TaskRunner) updatedVaultToken(token string) { + // Update the Vault token on the runner + tr.setVaultToken(token) + + // Update the tasks environment + tr.envBuilder.SetVaultToken(token, tr.task.Vault.Env) + + // Update the hooks with the new Vault token + tr.updateHooks() +} + +type vaultHookConfig struct { + vaultStanza *structs.Vault + client vaultclient.VaultClient + events EventEmitter + lifecycle TaskLifecycle + updater vaultTokenUpdateHandler + logger log.Logger + alloc *structs.Allocation + task string +} + +type vaultHook struct { + // vaultStanza is the vault stanza for the task + vaultStanza *structs.Vault + + // eventEmitter is used to emit events to the task + eventEmitter EventEmitter + + // lifecycle is used to signal, restart and kill a task + lifecycle TaskLifecycle + + // updater is used to update the Vault token + updater vaultTokenUpdateHandler + + // client is the Vault client to retrieve and renew the Vault token + client vaultclient.VaultClient + + // logger is used to log + logger log.Logger + + // ctx and cancel are used to kill the long running token manager + ctx context.Context + cancel context.CancelFunc + + // tokenPath is the path in which to read and write the token + tokenPath string + + // alloc is the allocation + alloc *structs.Allocation + + // taskName is the name of the task + taskName string + + // firstRun stores whether it is the first run for the hook + firstRun bool + + // future is used to wait on retrieving a Vault token + future *tokenFuture +} + +func newVaultHook(config *vaultHookConfig) *vaultHook { + ctx, cancel := context.WithCancel(context.Background()) + h := &vaultHook{ + vaultStanza: config.vaultStanza, + client: config.client, + eventEmitter: config.events, + lifecycle: config.lifecycle, + updater: config.updater, + alloc: config.alloc, + taskName: config.task, + firstRun: true, + ctx: ctx, + cancel: cancel, + future: newTokenFuture(), + } + h.logger = config.logger.Named(h.Name()) + return h +} + +func (*vaultHook) Name() string { + return "vault" +} + +func (h *vaultHook) Prerun(ctx context.Context, req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { + // If we have already run prerun before exit early. We do not use the + // PrerunDone value because we want to recover the token on restoration. + first := h.firstRun + h.firstRun = false + if !first { + return nil + } + + // Try to recover a token if it was previously written in the secrets + // directory + recoveredToken := "" + h.tokenPath = filepath.Join(req.TaskDir, allocdir.TaskSecrets, vaultTokenFile) + data, err := ioutil.ReadFile(h.tokenPath) + if err != nil { + if !os.IsNotExist(err) { + return fmt.Errorf("failed to recover vault token: %v", err) + } + + // Token file doesn't exist + } else { + // Store the recovered token + recoveredToken = string(data) + } + + // Launch the token manager + go h.run(recoveredToken) + + // Block until we get a token + select { + case <-h.future.Wait(): + case <-ctx.Done(): + return nil + } + + h.updater.updatedVaultToken(h.future.Get()) + return nil +} + +func (h *vaultHook) Poststop(ctx context.Context, req *interfaces.TaskPoststopRequest, resp *interfaces.TaskPoststopResponse) error { + // Shutdown any created manager + h.cancel() + return nil +} + +// run should be called in a go-routine and manages the derivation, renewal and +// handling of errors with the Vault token. The optional parameter allows +// setting the initial Vault token. This is useful when the Vault token is +// recovered off disk. +func (h *vaultHook) run(token string) { + // Helper for stopping token renewal + stopRenewal := func() { + if err := h.client.StopRenewToken(h.future.Get()); err != nil { + h.logger.Warn("failed to stop token renewal", "error", err) + } + } + + // updatedToken lets us store state between loops. If true, a new token + // has been retrieved and we need to apply the Vault change mode + var updatedToken bool + +OUTER: + for { + // Check if we should exit + select { + case <-h.ctx.Done(): + stopRenewal() + return + default: + } + + // Clear the token + h.future.Clear() + + // Check if there already is a token which can be the case for + // restoring the TaskRunner + if token == "" { + // Get a token + var exit bool + token, exit = h.deriveVaultToken() + if exit { + // Exit the manager + return + } + + // Write the token to disk + if err := h.writeToken(token); err != nil { + errorString := "failed to write Vault token to disk" + h.logger.Error(errorString, "error", err) + h.lifecycle.Kill("vault", errorString, true) + return + } + } + + // Start the renewal process + renewCh, err := h.client.RenewToken(token, 30) + + // An error returned means the token is not being renewed + if err != nil { + h.logger.Error("failed to start renewal of Vault token", "error", err) + token = "" + goto OUTER + } + + // The Vault token is valid now, so set it + h.future.Set(token) + + if updatedToken { + switch h.vaultStanza.ChangeMode { + case structs.VaultChangeModeSignal: + s, err := signals.Parse(h.vaultStanza.ChangeSignal) + if err != nil { + h.logger.Error("failed to parse signal", "error", err) + h.lifecycle.Kill("vault", fmt.Sprintf("failed to parse signal: %v", err), true) + return + } + + if err := h.lifecycle.Signal("vault", "new Vault token acquired", s); err != nil { + h.logger.Error("failed to send signal", "error", err) + h.lifecycle.Kill("vault", fmt.Sprintf("failed to send signal: %v", err), true) + return + } + case structs.VaultChangeModeRestart: + const noFailure = false + h.lifecycle.Restart("vault", "new Vault token acquired", noFailure) + case structs.VaultChangeModeNoop: + fallthrough + default: + h.logger.Error("invalid Vault change mode", "mode", h.vaultStanza.ChangeMode) + } + + // We have handled it + updatedToken = false + + // Call the handler + h.updater.updatedVaultToken(token) + } + + // Start watching for renewal errors + select { + case err := <-renewCh: + // Clear the token + token = "" + h.logger.Error("failed to renew Vault token", "error", err) + stopRenewal() + + // Check if we have to do anything + if h.vaultStanza.ChangeMode != structs.VaultChangeModeNoop { + updatedToken = true + } + case <-h.ctx.Done(): + stopRenewal() + return + } + } +} + +// deriveVaultToken derives the Vault token using exponential backoffs. It +// returns the Vault token and whether the manager should exit. +func (h *vaultHook) deriveVaultToken() (token string, exit bool) { + attempts := 0 + for { + tokens, err := h.client.DeriveToken(h.alloc, []string{h.taskName}) + if err == nil { + return tokens[h.taskName], false + } + + // Check if this is a server side error + if structs.IsServerSide(err) { + h.logger.Error("failed to derive Vault token", "error", err, "server_side", true) + h.lifecycle.Kill("vault", fmt.Sprintf("server error deriving vault token: %v", err), true) + return "", true + } + + // Check if we can't recover from the error + if !structs.IsRecoverable(err) { + h.logger.Error("failed to derive Vault token", "error", err, "recoverable", false) + h.lifecycle.Kill("vault", fmt.Sprintf("failed to derive token: %v", err), true) + return "", true + } + + // Handle the retry case + backoff := (1 << (2 * uint64(attempts))) * vaultBackoffBaseline + if backoff > vaultBackoffLimit { + backoff = vaultBackoffLimit + } + h.logger.Error("failed to derive Vault token", "error", err, "recoverable", true, "backoff", backoff) + + attempts++ + + // Wait till retrying + select { + case <-h.ctx.Done(): + return "", true + case <-time.After(backoff): + } + } +} + +// writeToken writes the given token to disk +func (h *vaultHook) writeToken(token string) error { + if err := ioutil.WriteFile(h.tokenPath, []byte(token), 0777); err != nil { + return fmt.Errorf("failed to write vault token: %v", err) + } + + return nil +} + +// tokenFuture stores the Vault token and allows consumers to block till a valid +// token exists +type tokenFuture struct { + waiting []chan struct{} + token string + set bool + m sync.Mutex +} + +// newTokenFuture returns a new token future without any token set +func newTokenFuture() *tokenFuture { + return &tokenFuture{} +} + +// Wait returns a channel that can be waited on. When this channel unblocks, a +// valid token will be available via the Get method +func (f *tokenFuture) Wait() <-chan struct{} { + f.m.Lock() + defer f.m.Unlock() + + c := make(chan struct{}) + if f.set { + close(c) + return c + } + + f.waiting = append(f.waiting, c) + return c +} + +// Set sets the token value and unblocks any caller of Wait +func (f *tokenFuture) Set(token string) *tokenFuture { + f.m.Lock() + defer f.m.Unlock() + + f.set = true + f.token = token + for _, w := range f.waiting { + close(w) + } + f.waiting = nil + return f +} + +// Clear clears the set vault token. +func (f *tokenFuture) Clear() *tokenFuture { + f.m.Lock() + defer f.m.Unlock() + + f.token = "" + f.set = false + return f +} + +// Get returns the set Vault token +func (f *tokenFuture) Get() string { + f.m.Lock() + defer f.m.Unlock() + return f.token +} diff --git a/client/client.go b/client/client.go index edbcd33cd..be4c1dee7 100644 --- a/client/client.go +++ b/client/client.go @@ -1930,6 +1930,7 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error Logger: logger, ClientConfig: c.config, StateDB: c.stateDB, + Vault: c.vaultClient, } c.configLock.RUnlock() From c9765deff1a21a073380c0a8f0bda2a49a5dfcc7 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Fri, 13 Jul 2018 09:45:29 -0700 Subject: [PATCH 018/152] address comments --- client/allocrunnerv2/alloc_runner.go | 24 ++++++++++++------- client/allocrunnerv2/taskrunner/vault_hook.go | 6 ++--- 2 files changed, 17 insertions(+), 13 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index fe1d2fdae..92d930255 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -12,7 +12,9 @@ import ( "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" "github.com/hashicorp/nomad/client/allocrunnerv2/state" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" + "github.com/hashicorp/nomad/client/config" cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" ) @@ -21,7 +23,10 @@ type allocRunner struct { // Logger is the logger for the alloc runner. logger log.Logger - config *Config + clientConfig *config.Config + + // vaultClient is the used to manage Vault tokens + vaultClient vaultclient.VaultClient // waitCh is closed when the alloc runner has transitioned to a terminal // state @@ -54,12 +59,13 @@ type allocRunner struct { // NewAllocRunner returns a new allocation runner. func NewAllocRunner(config *Config) *allocRunner { ar := &allocRunner{ - config: config, - alloc: config.Alloc, - tasks: make(map[string]*taskrunner.TaskRunner), - waitCh: make(chan struct{}), - updateCh: make(chan *structs.Allocation), - stateDB: config.StateDB, + clientConfig: config.ClientConfig, + vaultClient: config.Vault, + alloc: config.Alloc, + tasks: make(map[string]*taskrunner.TaskRunner), + waitCh: make(chan struct{}), + updateCh: make(chan *structs.Allocation), + stateDB: config.StateDB, } // Create alloc dir @@ -158,12 +164,12 @@ func (ar *allocRunner) runTask(alloc *structs.Allocation, task *structs.Task) er // Create the runner config := &taskrunner.Config{ Alloc: alloc, - ClientConfig: ar.config.ClientConfig, + ClientConfig: ar.clientConfig, Task: task, TaskDir: ar.allocDir.NewTaskDir(task.Name), Logger: ar.logger, StateDB: ar.stateDB, - VaultClient: ar.config.Vault, + VaultClient: ar.vaultClient, } tr, err := taskrunner.NewTaskRunner(config) if err != nil { diff --git a/client/allocrunnerv2/taskrunner/vault_hook.go b/client/allocrunnerv2/taskrunner/vault_hook.go index 6cb3be915..01e896ac6 100644 --- a/client/allocrunnerv2/taskrunner/vault_hook.go +++ b/client/allocrunnerv2/taskrunner/vault_hook.go @@ -184,11 +184,9 @@ func (h *vaultHook) run(token string) { OUTER: for { // Check if we should exit - select { - case <-h.ctx.Done(): + if h.ctx.Err() != nil { stopRenewal() return - default: } // Clear the token @@ -321,7 +319,7 @@ func (h *vaultHook) deriveVaultToken() (token string, exit bool) { // writeToken writes the given token to disk func (h *vaultHook) writeToken(token string) error { - if err := ioutil.WriteFile(h.tokenPath, []byte(token), 0777); err != nil { + if err := ioutil.WriteFile(h.tokenPath, []byte(token), 0666); err != nil { return fmt.Errorf("failed to write vault token: %v", err) } From 88aa0299a923a8114bd9f3b8b0e2ee9fb12ae04f Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Fri, 13 Jul 2018 13:45:57 -0700 Subject: [PATCH 019/152] Template hook --- .../taskrunner/interfaces/events.go | 8 + .../taskrunner/interfaces/lifecycle.go | 11 + client/allocrunnerv2/taskrunner/lifecycle.go | 8 - .../allocrunnerv2/taskrunner/task_runner.go | 5 + .../taskrunner/task_runner_hooks.go | 21 +- .../taskrunner/template/template.go | 679 +++++++++ .../taskrunner/template/template_test.go | 1342 +++++++++++++++++ .../allocrunnerv2/taskrunner/template_hook.go | 157 ++ client/allocrunnerv2/taskrunner/vault_hook.go | 9 +- 9 files changed, 2222 insertions(+), 18 deletions(-) create mode 100644 client/allocrunnerv2/taskrunner/interfaces/events.go create mode 100644 client/allocrunnerv2/taskrunner/interfaces/lifecycle.go create mode 100644 client/allocrunnerv2/taskrunner/template/template.go create mode 100644 client/allocrunnerv2/taskrunner/template/template_test.go create mode 100644 client/allocrunnerv2/taskrunner/template_hook.go diff --git a/client/allocrunnerv2/taskrunner/interfaces/events.go b/client/allocrunnerv2/taskrunner/interfaces/events.go new file mode 100644 index 000000000..3f908cedf --- /dev/null +++ b/client/allocrunnerv2/taskrunner/interfaces/events.go @@ -0,0 +1,8 @@ +package interfaces + +import "github.com/hashicorp/nomad/nomad/structs" + +type EventEmitter interface { + SetState(state string, event *structs.TaskEvent) + EmitEvent(source, message string) +} diff --git a/client/allocrunnerv2/taskrunner/interfaces/lifecycle.go b/client/allocrunnerv2/taskrunner/interfaces/lifecycle.go new file mode 100644 index 000000000..e440943a3 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/interfaces/lifecycle.go @@ -0,0 +1,11 @@ +package interfaces + +import "os" + +// XXX These should probably all return an error and we should have predefined +// error types for the task not currently running +type TaskLifecycle interface { + Restart(source, reason string, failure bool) + Signal(source, reason string, s os.Signal) error + Kill(source, reason string, fail bool) +} diff --git a/client/allocrunnerv2/taskrunner/lifecycle.go b/client/allocrunnerv2/taskrunner/lifecycle.go index e93fbf065..4d3f7bd0f 100644 --- a/client/allocrunnerv2/taskrunner/lifecycle.go +++ b/client/allocrunnerv2/taskrunner/lifecycle.go @@ -2,14 +2,6 @@ package taskrunner import "os" -// XXX These should probably all return an error and we should have predefined -// error types for the task not currently running -type TaskLifecycle interface { - Restart(source, reason string, failure bool) - Signal(source, reason string, s os.Signal) error - Kill(source, reason string, fail bool) -} - func (tr *TaskRunner) Restart(source, reason string, failure bool) { // TODO } diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 708a63df3..f54d5d6bd 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -498,6 +498,11 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { //} } +func (tr *TaskRunner) EmitEvent(source, message string) { + event := structs.NewTaskEvent(source).SetMessage(message) + tr.SetState("", event) +} + // WaitCh is closed when TaskRunner.Run exits. func (tr *TaskRunner) WaitCh() <-chan struct{} { return tr.waitCh diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index cb2382f8d..5205754a1 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -9,16 +9,13 @@ import ( "github.com/hashicorp/nomad/client/allocrunner/getter" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + ti "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" cconfig "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/nomad/structs" ) -type EventEmitter interface { - SetState(state string, event *structs.TaskEvent) -} - // initHooks intializes the tasks hooks. func (tr *TaskRunner) initHooks() { hookLogger := tr.logger.Named("task_hook") @@ -43,6 +40,18 @@ func (tr *TaskRunner) initHooks() { task: tr.taskName, })) } + + // If there are templates is enabled, add the hook + if task := tr.Task(); len(task.Templates) != 0 { + tr.runnerHooks = append(tr.runnerHooks, newTemplateHook(&templateHookConfig{ + logger: hookLogger, + lifecycle: tr, + events: tr, + templates: task.Templates, + clientConfig: tr.clientConfig, + envBuilder: tr.envBuilder, + })) + } } // prerun is used to run the runners prerun hooks. @@ -304,11 +313,11 @@ func (h *taskDirHook) Prerun(ctx context.Context, req *interfaces.TaskPrerunRequ // artifactHook downloads artifacts for a task. type artifactHook struct { - eventEmitter EventEmitter + eventEmitter ti.EventEmitter logger log.Logger } -func newArtifactHook(e EventEmitter, logger log.Logger) *artifactHook { +func newArtifactHook(e ti.EventEmitter, logger log.Logger) *artifactHook { h := &artifactHook{ eventEmitter: e, } diff --git a/client/allocrunnerv2/taskrunner/template/template.go b/client/allocrunnerv2/taskrunner/template/template.go new file mode 100644 index 000000000..27dba209f --- /dev/null +++ b/client/allocrunnerv2/taskrunner/template/template.go @@ -0,0 +1,679 @@ +package template + +import ( + "fmt" + "math/rand" + "os" + "path/filepath" + "sort" + "strconv" + "strings" + "sync" + "time" + + ctconf "github.com/hashicorp/consul-template/config" + "github.com/hashicorp/consul-template/manager" + "github.com/hashicorp/consul-template/signals" + envparse "github.com/hashicorp/go-envparse" + multierror "github.com/hashicorp/go-multierror" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/driver/env" + "github.com/hashicorp/nomad/helper" + "github.com/hashicorp/nomad/nomad/structs" +) + +const ( + // consulTemplateSourceName is the source name when using the TaskHooks. + consulTemplateSourceName = "Template" + + // hostSrcOption is the Client option that determines whether the template + // source may be from the host + hostSrcOption = "template.allow_host_source" + + // missingDepEventLimit is the number of missing dependencies that will be + // logged before we switch to showing just the number of missing + // dependencies. + missingDepEventLimit = 3 + + // DefaultMaxTemplateEventRate is the default maximum rate at which a + // template event should be fired. + DefaultMaxTemplateEventRate = 3 * time.Second +) + +// TaskTemplateManager is used to run a set of templates for a given task +type TaskTemplateManager struct { + // config holds the template managers configuration + config *TaskTemplateManagerConfig + + // lookup allows looking up the set of Nomad templates by their consul-template ID + lookup map[string][]*structs.Template + + // runner is the consul-template runner + runner *manager.Runner + + // signals is a lookup map from the string representation of a signal to its + // actual signal + signals map[string]os.Signal + + // shutdownCh is used to signal and started goroutine to shutdown + shutdownCh chan struct{} + + // shutdown marks whether the manager has been shutdown + shutdown bool + shutdownLock sync.Mutex +} + +// TaskTemplateManagerConfig is used to configure an instance of the +// TaskTemplateManager +type TaskTemplateManagerConfig struct { + // UnblockCh is closed when the template has been rendered + UnblockCh chan struct{} + + // Lifecycle is used to interact with the task the template manager is being + // run for + Lifecycle interfaces.TaskLifecycle + + // Events is used to emit events for the task + Events interfaces.EventEmitter + + // Templates is the set of templates we are managing + Templates []*structs.Template + + // ClientConfig is the Nomad Client configuration + ClientConfig *config.Config + + // VaultToken is the Vault token for the task. + VaultToken string + + // TaskDir is the task's directory + TaskDir string + + // EnvBuilder is the environment variable builder for the task. + EnvBuilder *env.Builder + + // MaxTemplateEventRate is the maximum rate at which we should emit events. + MaxTemplateEventRate time.Duration + + // retryRate is only used for testing and is used to increase the retry rate + retryRate time.Duration +} + +// Validate validates the configuration. +func (c *TaskTemplateManagerConfig) Validate() error { + if c == nil { + return fmt.Errorf("Nil config passed") + } else if c.UnblockCh == nil { + return fmt.Errorf("Invalid unblock channel given") + } else if c.Lifecycle == nil { + return fmt.Errorf("Invalid lifecycle hooks given") + } else if c.Events == nil { + return fmt.Errorf("Invalid event hook given") + } else if c.ClientConfig == nil { + return fmt.Errorf("Invalid client config given") + } else if c.TaskDir == "" { + return fmt.Errorf("Invalid task directory given: %q", c.TaskDir) + } else if c.EnvBuilder == nil { + return fmt.Errorf("Invalid task environment given") + } else if c.MaxTemplateEventRate == 0 { + return fmt.Errorf("Invalid max template event rate given") + } + + return nil +} + +func NewTaskTemplateManager(config *TaskTemplateManagerConfig) (*TaskTemplateManager, error) { + // Check pre-conditions + if err := config.Validate(); err != nil { + return nil, err + } + + tm := &TaskTemplateManager{ + config: config, + shutdownCh: make(chan struct{}), + } + + // Parse the signals that we need + for _, tmpl := range config.Templates { + if tmpl.ChangeSignal == "" { + continue + } + + sig, err := signals.Parse(tmpl.ChangeSignal) + if err != nil { + return nil, fmt.Errorf("Failed to parse signal %q", tmpl.ChangeSignal) + } + + if tm.signals == nil { + tm.signals = make(map[string]os.Signal) + } + + tm.signals[tmpl.ChangeSignal] = sig + } + + // Build the consul-template runner + runner, lookup, err := templateRunner(config) + if err != nil { + return nil, err + } + tm.runner = runner + tm.lookup = lookup + + go tm.run() + return tm, nil +} + +// Stop is used to stop the consul-template runner +func (tm *TaskTemplateManager) Stop() { + tm.shutdownLock.Lock() + defer tm.shutdownLock.Unlock() + + if tm.shutdown { + return + } + + close(tm.shutdownCh) + tm.shutdown = true + + // Stop the consul-template runner + if tm.runner != nil { + tm.runner.Stop() + } +} + +// run is the long lived loop that handles errors and templates being rendered +func (tm *TaskTemplateManager) run() { + // Runner is nil if there is no templates + if tm.runner == nil { + // Unblock the start if there is nothing to do + close(tm.config.UnblockCh) + return + } + + // Start the runner + go tm.runner.Start() + + // Block till all the templates have been rendered + tm.handleFirstRender() + + // Detect if there was a shutdown. + select { + case <-tm.shutdownCh: + return + default: + } + + // Read environment variables from env templates before we unblock + envMap, err := loadTemplateEnv(tm.config.Templates, tm.config.TaskDir) + if err != nil { + tm.config.Lifecycle.Kill(consulTemplateSourceName, err.Error(), true) + return + } + tm.config.EnvBuilder.SetTemplateEnv(envMap) + + // Unblock the task + close(tm.config.UnblockCh) + + // If all our templates are change mode no-op, then we can exit here + if tm.allTemplatesNoop() { + return + } + + // handle all subsequent render events. + tm.handleTemplateRerenders(time.Now()) +} + +// handleFirstRender blocks till all templates have been rendered +func (tm *TaskTemplateManager) handleFirstRender() { + // missingDependencies is the set of missing dependencies. + var missingDependencies map[string]struct{} + + // eventTimer is used to trigger the firing of an event showing the missing + // dependencies. + eventTimer := time.NewTimer(tm.config.MaxTemplateEventRate) + if !eventTimer.Stop() { + <-eventTimer.C + } + + // outstandingEvent tracks whether there is an outstanding event that should + // be fired. + outstandingEvent := false + + // Wait till all the templates have been rendered +WAIT: + for { + select { + case <-tm.shutdownCh: + return + case err, ok := <-tm.runner.ErrCh: + if !ok { + continue + } + + tm.config.Lifecycle.Kill(consulTemplateSourceName, err.Error(), true) + case <-tm.runner.TemplateRenderedCh(): + // A template has been rendered, figure out what to do + events := tm.runner.RenderEvents() + + // Not all templates have been rendered yet + if len(events) < len(tm.lookup) { + continue + } + + for _, event := range events { + // This template hasn't been rendered + if event.LastWouldRender.IsZero() { + continue WAIT + } + } + + break WAIT + case <-tm.runner.RenderEventCh(): + events := tm.runner.RenderEvents() + joinedSet := make(map[string]struct{}) + for _, event := range events { + missing := event.MissingDeps + if missing == nil { + continue + } + + for _, dep := range missing.List() { + joinedSet[dep.String()] = struct{}{} + } + } + + // Check to see if the new joined set is the same as the old + different := len(joinedSet) != len(missingDependencies) + if !different { + for k := range joinedSet { + if _, ok := missingDependencies[k]; !ok { + different = true + break + } + } + } + + // Nothing to do + if !different { + continue + } + + // Update the missing set + missingDependencies = joinedSet + + // Update the event timer channel + if !outstandingEvent { + // We got new data so reset + outstandingEvent = true + eventTimer.Reset(tm.config.MaxTemplateEventRate) + } + case <-eventTimer.C: + if missingDependencies == nil { + continue + } + + // Clear the outstanding event + outstandingEvent = false + + // Build the missing set + missingSlice := make([]string, 0, len(missingDependencies)) + for k := range missingDependencies { + missingSlice = append(missingSlice, k) + } + sort.Strings(missingSlice) + + if l := len(missingSlice); l > missingDepEventLimit { + missingSlice[missingDepEventLimit] = fmt.Sprintf("and %d more", l-missingDepEventLimit) + missingSlice = missingSlice[:missingDepEventLimit+1] + } + + missingStr := strings.Join(missingSlice, ", ") + tm.config.Events.EmitEvent(consulTemplateSourceName, fmt.Sprintf("Missing: %s", missingStr)) + } + } +} + +// handleTemplateRerenders is used to handle template render events after they +// have all rendered. It takes action based on which set of templates re-render. +// The passed allRenderedTime is the time at which all templates have rendered. +// This is used to avoid signaling the task for any render event before hand. +func (tm *TaskTemplateManager) handleTemplateRerenders(allRenderedTime time.Time) { + // A lookup for the last time the template was handled + handledRenders := make(map[string]time.Time, len(tm.config.Templates)) + + for { + select { + case <-tm.shutdownCh: + return + case err, ok := <-tm.runner.ErrCh: + if !ok { + continue + } + + tm.config.Lifecycle.Kill(consulTemplateSourceName, err.Error(), true) + case <-tm.runner.TemplateRenderedCh(): + // A template has been rendered, figure out what to do + var handling []string + signals := make(map[string]struct{}) + restart := false + var splay time.Duration + + events := tm.runner.RenderEvents() + for id, event := range events { + + // First time through + if allRenderedTime.After(event.LastDidRender) || allRenderedTime.Equal(event.LastDidRender) { + handledRenders[id] = allRenderedTime + continue + } + + // We have already handled this one + if htime := handledRenders[id]; htime.After(event.LastDidRender) || htime.Equal(event.LastDidRender) { + continue + } + + // Lookup the template and determine what to do + tmpls, ok := tm.lookup[id] + if !ok { + tm.config.Lifecycle.Kill(consulTemplateSourceName, fmt.Sprintf("template runner returned unknown template id %q", id), true) + return + } + + // Read environment variables from templates + envMap, err := loadTemplateEnv(tm.config.Templates, tm.config.TaskDir) + if err != nil { + tm.config.Lifecycle.Kill(consulTemplateSourceName, err.Error(), true) + return + } + tm.config.EnvBuilder.SetTemplateEnv(envMap) + + for _, tmpl := range tmpls { + switch tmpl.ChangeMode { + case structs.TemplateChangeModeSignal: + signals[tmpl.ChangeSignal] = struct{}{} + case structs.TemplateChangeModeRestart: + restart = true + case structs.TemplateChangeModeNoop: + continue + } + + if tmpl.Splay > splay { + splay = tmpl.Splay + } + } + + handling = append(handling, id) + } + + if restart || len(signals) != 0 { + if splay != 0 { + ns := splay.Nanoseconds() + offset := rand.Int63n(ns) + t := time.Duration(offset) + + select { + case <-time.After(t): + case <-tm.shutdownCh: + return + } + } + + // Update handle time + for _, id := range handling { + handledRenders[id] = events[id].LastDidRender + } + + if restart { + const failure = false + tm.config.Lifecycle.Restart(consulTemplateSourceName, "template with change_mode restart re-rendered", failure) + } else if len(signals) != 0 { + var mErr multierror.Error + for signal := range signals { + err := tm.config.Lifecycle.Signal(consulTemplateSourceName, "template re-rendered", tm.signals[signal]) + if err != nil { + multierror.Append(&mErr, err) + } + } + + if err := mErr.ErrorOrNil(); err != nil { + flat := make([]os.Signal, 0, len(signals)) + for signal := range signals { + flat = append(flat, tm.signals[signal]) + } + tm.config.Lifecycle.Kill(consulTemplateSourceName, fmt.Sprintf("Sending signals %v failed: %v", flat, err), true) + } + } + } + } + } +} + +// allTemplatesNoop returns whether all the managed templates have change mode noop. +func (tm *TaskTemplateManager) allTemplatesNoop() bool { + for _, tmpl := range tm.config.Templates { + if tmpl.ChangeMode != structs.TemplateChangeModeNoop { + return false + } + } + + return true +} + +// templateRunner returns a consul-template runner for the given templates and a +// lookup by destination to the template. If no templates are in the config, a +// nil template runner and lookup is returned. +func templateRunner(config *TaskTemplateManagerConfig) ( + *manager.Runner, map[string][]*structs.Template, error) { + + if len(config.Templates) == 0 { + return nil, nil, nil + } + + // Parse the templates + ctmplMapping, err := parseTemplateConfigs(config) + if err != nil { + return nil, nil, err + } + + // Create the runner configuration. + runnerConfig, err := newRunnerConfig(config, ctmplMapping) + if err != nil { + return nil, nil, err + } + + runner, err := manager.NewRunner(runnerConfig, false, false) + if err != nil { + return nil, nil, err + } + + // Set Nomad's environment variables + runner.Env = config.EnvBuilder.Build().All() + + // Build the lookup + idMap := runner.TemplateConfigMapping() + lookup := make(map[string][]*structs.Template, len(idMap)) + for id, ctmpls := range idMap { + for _, ctmpl := range ctmpls { + templates := lookup[id] + templates = append(templates, ctmplMapping[ctmpl]) + lookup[id] = templates + } + } + + return runner, lookup, nil +} + +// parseTemplateConfigs converts the tasks templates in the config into +// consul-templates +func parseTemplateConfigs(config *TaskTemplateManagerConfig) (map[ctconf.TemplateConfig]*structs.Template, error) { + allowAbs := config.ClientConfig.ReadBoolDefault(hostSrcOption, true) + taskEnv := config.EnvBuilder.Build() + + ctmpls := make(map[ctconf.TemplateConfig]*structs.Template, len(config.Templates)) + for _, tmpl := range config.Templates { + var src, dest string + if tmpl.SourcePath != "" { + if filepath.IsAbs(tmpl.SourcePath) { + if !allowAbs { + return nil, fmt.Errorf("Specifying absolute template paths disallowed by client config: %q", tmpl.SourcePath) + } + + src = tmpl.SourcePath + } else { + src = filepath.Join(config.TaskDir, taskEnv.ReplaceEnv(tmpl.SourcePath)) + } + } + if tmpl.DestPath != "" { + dest = filepath.Join(config.TaskDir, taskEnv.ReplaceEnv(tmpl.DestPath)) + } + + ct := ctconf.DefaultTemplateConfig() + ct.Source = &src + ct.Destination = &dest + ct.Contents = &tmpl.EmbeddedTmpl + ct.LeftDelim = &tmpl.LeftDelim + ct.RightDelim = &tmpl.RightDelim + + // Set the permissions + if tmpl.Perms != "" { + v, err := strconv.ParseUint(tmpl.Perms, 8, 12) + if err != nil { + return nil, fmt.Errorf("Failed to parse %q as octal: %v", tmpl.Perms, err) + } + m := os.FileMode(v) + ct.Perms = &m + } + ct.Finalize() + + ctmpls[*ct] = tmpl + } + + return ctmpls, nil +} + +// newRunnerConfig returns a consul-template runner configuration, setting the +// Vault and Consul configurations based on the clients configs. +func newRunnerConfig(config *TaskTemplateManagerConfig, + templateMapping map[ctconf.TemplateConfig]*structs.Template) (*ctconf.Config, error) { + + cc := config.ClientConfig + conf := ctconf.DefaultConfig() + + // Gather the consul-template templates + flat := ctconf.TemplateConfigs(make([]*ctconf.TemplateConfig, 0, len(templateMapping))) + for ctmpl := range templateMapping { + local := ctmpl + flat = append(flat, &local) + } + conf.Templates = &flat + + // Go through the templates and determine the minimum Vault grace + vaultGrace := time.Duration(-1) + for _, tmpl := range templateMapping { + // Initial condition + if vaultGrace < 0 { + vaultGrace = tmpl.VaultGrace + } else if tmpl.VaultGrace < vaultGrace { + vaultGrace = tmpl.VaultGrace + } + } + + // Force faster retries + if config.retryRate != 0 { + rate := config.retryRate + conf.Consul.Retry.Backoff = &rate + } + + // Setup the Consul config + if cc.ConsulConfig != nil { + conf.Consul.Address = &cc.ConsulConfig.Addr + conf.Consul.Token = &cc.ConsulConfig.Token + + if cc.ConsulConfig.EnableSSL != nil && *cc.ConsulConfig.EnableSSL { + verify := cc.ConsulConfig.VerifySSL != nil && *cc.ConsulConfig.VerifySSL + conf.Consul.SSL = &ctconf.SSLConfig{ + Enabled: helper.BoolToPtr(true), + Verify: &verify, + Cert: &cc.ConsulConfig.CertFile, + Key: &cc.ConsulConfig.KeyFile, + CaCert: &cc.ConsulConfig.CAFile, + } + } + + if cc.ConsulConfig.Auth != "" { + parts := strings.SplitN(cc.ConsulConfig.Auth, ":", 2) + if len(parts) != 2 { + return nil, fmt.Errorf("Failed to parse Consul Auth config") + } + + conf.Consul.Auth = &ctconf.AuthConfig{ + Enabled: helper.BoolToPtr(true), + Username: &parts[0], + Password: &parts[1], + } + } + } + + // Setup the Vault config + // Always set these to ensure nothing is picked up from the environment + emptyStr := "" + conf.Vault.RenewToken = helper.BoolToPtr(false) + conf.Vault.Token = &emptyStr + if cc.VaultConfig != nil && cc.VaultConfig.IsEnabled() { + conf.Vault.Address = &cc.VaultConfig.Addr + conf.Vault.Token = &config.VaultToken + conf.Vault.Grace = helper.TimeToPtr(vaultGrace) + + if strings.HasPrefix(cc.VaultConfig.Addr, "https") || cc.VaultConfig.TLSCertFile != "" { + skipVerify := cc.VaultConfig.TLSSkipVerify != nil && *cc.VaultConfig.TLSSkipVerify + verify := !skipVerify + conf.Vault.SSL = &ctconf.SSLConfig{ + Enabled: helper.BoolToPtr(true), + Verify: &verify, + Cert: &cc.VaultConfig.TLSCertFile, + Key: &cc.VaultConfig.TLSKeyFile, + CaCert: &cc.VaultConfig.TLSCaFile, + CaPath: &cc.VaultConfig.TLSCaPath, + ServerName: &cc.VaultConfig.TLSServerName, + } + } else { + conf.Vault.SSL = &ctconf.SSLConfig{ + Enabled: helper.BoolToPtr(false), + Verify: helper.BoolToPtr(false), + Cert: &emptyStr, + Key: &emptyStr, + CaCert: &emptyStr, + CaPath: &emptyStr, + ServerName: &emptyStr, + } + } + } + + conf.Finalize() + return conf, nil +} + +// loadTemplateEnv loads task environment variables from all templates. +func loadTemplateEnv(tmpls []*structs.Template, taskDir string) (map[string]string, error) { + all := make(map[string]string, 50) + for _, t := range tmpls { + if !t.Envvars { + continue + } + f, err := os.Open(filepath.Join(taskDir, t.DestPath)) + if err != nil { + return nil, fmt.Errorf("error opening env template: %v", err) + } + defer f.Close() + + // Parse environment fil + vars, err := envparse.Parse(f) + if err != nil { + return nil, fmt.Errorf("error parsing env template %q: %v", t.DestPath, err) + } + for k, v := range vars { + all[k] = v + } + } + return all, nil +} diff --git a/client/allocrunnerv2/taskrunner/template/template_test.go b/client/allocrunnerv2/taskrunner/template/template_test.go new file mode 100644 index 000000000..102f6e745 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/template/template_test.go @@ -0,0 +1,1342 @@ +package template + +import ( + "fmt" + "io" + "io/ioutil" + "os" + "path/filepath" + "strings" + "testing" + "time" + + ctestutil "github.com/hashicorp/consul/testutil" + "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/driver/env" + "github.com/hashicorp/nomad/helper" + "github.com/hashicorp/nomad/nomad/mock" + "github.com/hashicorp/nomad/nomad/structs" + sconfig "github.com/hashicorp/nomad/nomad/structs/config" + "github.com/hashicorp/nomad/testutil" + "github.com/stretchr/testify/assert" + "github.com/stretchr/testify/require" +) + +const ( + // TestTaskName is the name of the injected task. It should appear in the + // environment variable $NOMAD_TASK_NAME + TestTaskName = "test-task" +) + +// MockTaskHooks is a mock of the TaskHooks interface useful for testing +type MockTaskHooks struct { + Restarts int + RestartCh chan struct{} + + Signals []os.Signal + SignalCh chan struct{} + + // SignalError is returned when Signal is called on the mock hook + SignalError error + + UnblockCh chan struct{} + + KillReason string + KillCh chan struct{} + + Events []string + EmitEventCh chan struct{} +} + +func NewMockTaskHooks() *MockTaskHooks { + return &MockTaskHooks{ + UnblockCh: make(chan struct{}, 1), + RestartCh: make(chan struct{}, 1), + SignalCh: make(chan struct{}, 1), + KillCh: make(chan struct{}, 1), + EmitEventCh: make(chan struct{}, 1), + } +} +func (m *MockTaskHooks) Restart(source, reason string, failure bool) { + m.Restarts++ + select { + case m.RestartCh <- struct{}{}: + default: + } +} + +func (m *MockTaskHooks) Signal(source, reason string, s os.Signal) error { + m.Signals = append(m.Signals, s) + select { + case m.SignalCh <- struct{}{}: + default: + } + + return m.SignalError +} + +func (m *MockTaskHooks) Kill(source, reason string, fail bool) { + m.KillReason = reason + select { + case m.KillCh <- struct{}{}: + default: + } +} + +func (m *MockTaskHooks) EmitEvent(source, message string) { + m.Events = append(m.Events, message) + select { + case m.EmitEventCh <- struct{}{}: + default: + } +} + +func (m *MockTaskHooks) SetState(state string, event *structs.TaskEvent) {} + +// testHarness is used to test the TaskTemplateManager by spinning up +// Consul/Vault as needed +type testHarness struct { + manager *TaskTemplateManager + mockHooks *MockTaskHooks + templates []*structs.Template + envBuilder *env.Builder + node *structs.Node + config *config.Config + vaultToken string + taskDir string + vault *testutil.TestVault + consul *ctestutil.TestServer + emitRate time.Duration +} + +// newTestHarness returns a harness starting a dev consul and vault server, +// building the appropriate config and creating a TaskTemplateManager +func newTestHarness(t *testing.T, templates []*structs.Template, consul, vault bool) *testHarness { + region := "global" + harness := &testHarness{ + mockHooks: NewMockTaskHooks(), + templates: templates, + node: mock.Node(), + config: &config.Config{Region: region}, + emitRate: DefaultMaxTemplateEventRate, + } + + // Build the task environment + a := mock.Alloc() + task := a.Job.TaskGroups[0].Tasks[0] + task.Name = TestTaskName + harness.envBuilder = env.NewBuilder(harness.node, a, task, region) + + // Make a tempdir + d, err := ioutil.TempDir("", "ct_test") + if err != nil { + t.Fatalf("Failed to make tmpdir: %v", err) + } + harness.taskDir = d + + if consul { + harness.consul, err = ctestutil.NewTestServer() + if err != nil { + t.Fatalf("error starting test Consul server: %v", err) + } + harness.config.ConsulConfig = &sconfig.ConsulConfig{ + Addr: harness.consul.HTTPAddr, + } + } + + if vault { + harness.vault = testutil.NewTestVault(t) + harness.config.VaultConfig = harness.vault.Config + harness.vaultToken = harness.vault.RootToken + } + + return harness +} + +func (h *testHarness) start(t *testing.T) { + if err := h.startWithErr(); err != nil { + t.Fatalf("failed to build task template manager: %v", err) + } +} + +func (h *testHarness) startWithErr() error { + var err error + h.manager, err = NewTaskTemplateManager(&TaskTemplateManagerConfig{ + UnblockCh: h.mockHooks.UnblockCh, + Lifecycle: h.mockHooks, + Events: h.mockHooks, + Templates: h.templates, + ClientConfig: h.config, + VaultToken: h.vaultToken, + TaskDir: h.taskDir, + EnvBuilder: h.envBuilder, + MaxTemplateEventRate: h.emitRate, + retryRate: 10 * time.Millisecond, + }) + + return err +} + +func (h *testHarness) setEmitRate(d time.Duration) { + h.emitRate = d +} + +// stop is used to stop any running Vault or Consul server plus the task manager +func (h *testHarness) stop() { + if h.vault != nil { + h.vault.Stop() + } + if h.consul != nil { + h.consul.Stop() + } + if h.manager != nil { + h.manager.Stop() + } + if h.taskDir != "" { + os.RemoveAll(h.taskDir) + } +} + +func TestTaskTemplateManager_InvalidConfig(t *testing.T) { + t.Parallel() + hooks := NewMockTaskHooks() + clientConfig := &config.Config{Region: "global"} + taskDir := "foo" + a := mock.Alloc() + envBuilder := env.NewBuilder(mock.Node(), a, a.Job.TaskGroups[0].Tasks[0], clientConfig.Region) + + cases := []struct { + name string + config *TaskTemplateManagerConfig + expectedErr string + }{ + { + name: "nil config", + config: nil, + expectedErr: "Nil config passed", + }, + { + name: "bad lifecycle hooks", + config: &TaskTemplateManagerConfig{ + UnblockCh: hooks.UnblockCh, + Events: hooks, + ClientConfig: clientConfig, + TaskDir: taskDir, + EnvBuilder: envBuilder, + MaxTemplateEventRate: DefaultMaxTemplateEventRate, + }, + expectedErr: "lifecycle hooks", + }, + { + name: "bad event hooks", + config: &TaskTemplateManagerConfig{ + UnblockCh: hooks.UnblockCh, + Lifecycle: hooks, + ClientConfig: clientConfig, + TaskDir: taskDir, + EnvBuilder: envBuilder, + MaxTemplateEventRate: DefaultMaxTemplateEventRate, + }, + expectedErr: "event hook", + }, + { + name: "bad client config", + config: &TaskTemplateManagerConfig{ + UnblockCh: hooks.UnblockCh, + Lifecycle: hooks, + Events: hooks, + TaskDir: taskDir, + EnvBuilder: envBuilder, + MaxTemplateEventRate: DefaultMaxTemplateEventRate, + }, + expectedErr: "client config", + }, + { + name: "bad task dir", + config: &TaskTemplateManagerConfig{ + UnblockCh: hooks.UnblockCh, + ClientConfig: clientConfig, + Lifecycle: hooks, + Events: hooks, + EnvBuilder: envBuilder, + MaxTemplateEventRate: DefaultMaxTemplateEventRate, + }, + expectedErr: "task directory", + }, + { + name: "bad env builder", + config: &TaskTemplateManagerConfig{ + UnblockCh: hooks.UnblockCh, + ClientConfig: clientConfig, + Lifecycle: hooks, + Events: hooks, + TaskDir: taskDir, + MaxTemplateEventRate: DefaultMaxTemplateEventRate, + }, + expectedErr: "task environment", + }, + { + name: "bad max event rate", + config: &TaskTemplateManagerConfig{ + UnblockCh: hooks.UnblockCh, + ClientConfig: clientConfig, + Lifecycle: hooks, + Events: hooks, + TaskDir: taskDir, + EnvBuilder: envBuilder, + }, + expectedErr: "template event rate", + }, + { + name: "valid", + config: &TaskTemplateManagerConfig{ + UnblockCh: hooks.UnblockCh, + ClientConfig: clientConfig, + Lifecycle: hooks, + Events: hooks, + TaskDir: taskDir, + EnvBuilder: envBuilder, + MaxTemplateEventRate: DefaultMaxTemplateEventRate, + }, + }, + { + name: "invalid signal", + config: &TaskTemplateManagerConfig{ + UnblockCh: hooks.UnblockCh, + Templates: []*structs.Template{ + { + DestPath: "foo", + EmbeddedTmpl: "hello, world", + ChangeMode: structs.TemplateChangeModeSignal, + ChangeSignal: "foobarbaz", + }, + }, + ClientConfig: clientConfig, + Lifecycle: hooks, + Events: hooks, + TaskDir: taskDir, + EnvBuilder: envBuilder, + MaxTemplateEventRate: DefaultMaxTemplateEventRate, + }, + expectedErr: "parse signal", + }, + } + + for _, c := range cases { + t.Run(c.name, func(t *testing.T) { + _, err := NewTaskTemplateManager(c.config) + if err != nil { + if c.expectedErr == "" { + t.Fatalf("unexpected error: %v", err) + } else if !strings.Contains(err.Error(), c.expectedErr) { + t.Fatalf("expected error to contain %q; got %q", c.expectedErr, err.Error()) + } + } else if c.expectedErr != "" { + t.Fatalf("expected an error to contain %q", c.expectedErr) + } + }) + } +} + +func TestTaskTemplateManager_HostPath(t *testing.T) { + t.Parallel() + // Make a template that will render immediately and write it to a tmp file + f, err := ioutil.TempFile("", "") + if err != nil { + t.Fatalf("Bad: %v", err) + } + defer f.Close() + defer os.Remove(f.Name()) + + content := "hello, world!" + if _, err := io.WriteString(f, content); err != nil { + t.Fatalf("Bad: %v", err) + } + + file := "my.tmpl" + template := &structs.Template{ + SourcePath: f.Name(), + DestPath: file, + ChangeMode: structs.TemplateChangeModeNoop, + } + + harness := newTestHarness(t, []*structs.Template{template}, false, false) + harness.start(t) + defer harness.stop() + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + // Check the file is there + path := filepath.Join(harness.taskDir, file) + raw, err := ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != content { + t.Fatalf("Unexpected template data; got %q, want %q", s, content) + } + + // Change the config to disallow host sources + harness = newTestHarness(t, []*structs.Template{template}, false, false) + harness.config.Options = map[string]string{ + hostSrcOption: "false", + } + if err := harness.startWithErr(); err == nil || !strings.Contains(err.Error(), "absolute") { + t.Fatalf("Expected absolute template path disallowed: %v", err) + } +} + +func TestTaskTemplateManager_Unblock_Static(t *testing.T) { + t.Parallel() + // Make a template that will render immediately + content := "hello, world!" + file := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: content, + DestPath: file, + ChangeMode: structs.TemplateChangeModeNoop, + } + + harness := newTestHarness(t, []*structs.Template{template}, false, false) + harness.start(t) + defer harness.stop() + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + // Check the file is there + path := filepath.Join(harness.taskDir, file) + raw, err := ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != content { + t.Fatalf("Unexpected template data; got %q, want %q", s, content) + } +} + +func TestTaskTemplateManager_Permissions(t *testing.T) { + t.Parallel() + // Make a template that will render immediately + content := "hello, world!" + file := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: content, + DestPath: file, + ChangeMode: structs.TemplateChangeModeNoop, + Perms: "777", + } + + harness := newTestHarness(t, []*structs.Template{template}, false, false) + harness.start(t) + defer harness.stop() + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + // Check the file is there + path := filepath.Join(harness.taskDir, file) + fi, err := os.Stat(path) + if err != nil { + t.Fatalf("Failed to stat file: %v", err) + } + + if m := fi.Mode(); m != os.ModePerm { + t.Fatalf("Got mode %v; want %v", m, os.ModePerm) + } +} + +func TestTaskTemplateManager_Unblock_Static_NomadEnv(t *testing.T) { + t.Parallel() + // Make a template that will render immediately + content := `Hello Nomad Task: {{env "NOMAD_TASK_NAME"}}` + expected := fmt.Sprintf("Hello Nomad Task: %s", TestTaskName) + file := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: content, + DestPath: file, + ChangeMode: structs.TemplateChangeModeNoop, + } + + harness := newTestHarness(t, []*structs.Template{template}, false, false) + harness.start(t) + defer harness.stop() + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + // Check the file is there + path := filepath.Join(harness.taskDir, file) + raw, err := ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != expected { + t.Fatalf("Unexpected template data; got %q, want %q", s, expected) + } +} + +func TestTaskTemplateManager_Unblock_Static_AlreadyRendered(t *testing.T) { + t.Parallel() + // Make a template that will render immediately + content := "hello, world!" + file := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: content, + DestPath: file, + ChangeMode: structs.TemplateChangeModeNoop, + } + + harness := newTestHarness(t, []*structs.Template{template}, false, false) + + // Write the contents + path := filepath.Join(harness.taskDir, file) + if err := ioutil.WriteFile(path, []byte(content), 0777); err != nil { + t.Fatalf("Failed to write data: %v", err) + } + + harness.start(t) + defer harness.stop() + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + // Check the file is there + path = filepath.Join(harness.taskDir, file) + raw, err := ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != content { + t.Fatalf("Unexpected template data; got %q, want %q", s, content) + } +} + +func TestTaskTemplateManager_Unblock_Consul(t *testing.T) { + t.Parallel() + // Make a template that will render based on a key in Consul + key := "foo" + content := "barbaz" + embedded := fmt.Sprintf(`{{key "%s"}}`, key) + file := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: embedded, + DestPath: file, + ChangeMode: structs.TemplateChangeModeNoop, + } + + harness := newTestHarness(t, []*structs.Template{template}, true, false) + harness.start(t) + defer harness.stop() + + // Ensure no unblock + select { + case <-harness.mockHooks.UnblockCh: + t.Fatalf("Task unblock should have not have been called") + case <-time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second): + } + + // Write the key to Consul + harness.consul.SetKV(t, key, []byte(content)) + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + // Check the file is there + path := filepath.Join(harness.taskDir, file) + raw, err := ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != content { + t.Fatalf("Unexpected template data; got %q, want %q", s, content) + } +} + +func TestTaskTemplateManager_Unblock_Vault(t *testing.T) { + t.Parallel() + require := require.New(t) + // Make a template that will render based on a key in Vault + vaultPath := "secret/data/password" + key := "password" + content := "barbaz" + embedded := fmt.Sprintf(`{{with secret "%s"}}{{.Data.data.%s}}{{end}}`, vaultPath, key) + file := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: embedded, + DestPath: file, + ChangeMode: structs.TemplateChangeModeNoop, + } + + harness := newTestHarness(t, []*structs.Template{template}, false, true) + harness.start(t) + defer harness.stop() + + // Ensure no unblock + select { + case <-harness.mockHooks.UnblockCh: + t.Fatalf("Task unblock should not have been called") + case <-time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second): + } + + // Write the secret to Vault + logical := harness.vault.Client.Logical() + _, err := logical.Write(vaultPath, map[string]interface{}{"data": map[string]interface{}{key: content}}) + require.NoError(err) + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + // Check the file is there + path := filepath.Join(harness.taskDir, file) + raw, err := ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != content { + t.Fatalf("Unexpected template data; got %q, want %q", s, content) + } +} + +func TestTaskTemplateManager_Unblock_Multi_Template(t *testing.T) { + t.Parallel() + // Make a template that will render immediately + staticContent := "hello, world!" + staticFile := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: staticContent, + DestPath: staticFile, + ChangeMode: structs.TemplateChangeModeNoop, + } + + // Make a template that will render based on a key in Consul + consulKey := "foo" + consulContent := "barbaz" + consulEmbedded := fmt.Sprintf(`{{key "%s"}}`, consulKey) + consulFile := "consul.tmpl" + template2 := &structs.Template{ + EmbeddedTmpl: consulEmbedded, + DestPath: consulFile, + ChangeMode: structs.TemplateChangeModeNoop, + } + + harness := newTestHarness(t, []*structs.Template{template, template2}, true, false) + harness.start(t) + defer harness.stop() + + // Ensure no unblock + select { + case <-harness.mockHooks.UnblockCh: + t.Fatalf("Task unblock should have not have been called") + case <-time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second): + } + + // Check that the static file has been rendered + path := filepath.Join(harness.taskDir, staticFile) + raw, err := ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != staticContent { + t.Fatalf("Unexpected template data; got %q, want %q", s, staticContent) + } + + // Write the key to Consul + harness.consul.SetKV(t, consulKey, []byte(consulContent)) + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + // Check the consul file is there + path = filepath.Join(harness.taskDir, consulFile) + raw, err = ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != consulContent { + t.Fatalf("Unexpected template data; got %q, want %q", s, consulContent) + } +} + +func TestTaskTemplateManager_Rerender_Noop(t *testing.T) { + t.Parallel() + // Make a template that will render based on a key in Consul + key := "foo" + content1 := "bar" + content2 := "baz" + embedded := fmt.Sprintf(`{{key "%s"}}`, key) + file := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: embedded, + DestPath: file, + ChangeMode: structs.TemplateChangeModeNoop, + } + + harness := newTestHarness(t, []*structs.Template{template}, true, false) + harness.start(t) + defer harness.stop() + + // Ensure no unblock + select { + case <-harness.mockHooks.UnblockCh: + t.Fatalf("Task unblock should have not have been called") + case <-time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second): + } + + // Write the key to Consul + harness.consul.SetKV(t, key, []byte(content1)) + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + // Check the file is there + path := filepath.Join(harness.taskDir, file) + raw, err := ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != content1 { + t.Fatalf("Unexpected template data; got %q, want %q", s, content1) + } + + // Update the key in Consul + harness.consul.SetKV(t, key, []byte(content2)) + + select { + case <-harness.mockHooks.RestartCh: + t.Fatalf("Noop ignored: %+v", harness.mockHooks) + case <-harness.mockHooks.SignalCh: + t.Fatalf("Noop ignored: %+v", harness.mockHooks) + case <-time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second): + } + + // Check the file has been updated + path = filepath.Join(harness.taskDir, file) + raw, err = ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != content2 { + t.Fatalf("Unexpected template data; got %q, want %q", s, content2) + } +} + +func TestTaskTemplateManager_Rerender_Signal(t *testing.T) { + t.Parallel() + // Make a template that renders based on a key in Consul and sends SIGALRM + key1 := "foo" + content1_1 := "bar" + content1_2 := "baz" + embedded1 := fmt.Sprintf(`{{key "%s"}}`, key1) + file1 := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: embedded1, + DestPath: file1, + ChangeMode: structs.TemplateChangeModeSignal, + ChangeSignal: "SIGALRM", + } + + // Make a template that renders based on a key in Consul and sends SIGBUS + key2 := "bam" + content2_1 := "cat" + content2_2 := "dog" + embedded2 := fmt.Sprintf(`{{key "%s"}}`, key2) + file2 := "my-second.tmpl" + template2 := &structs.Template{ + EmbeddedTmpl: embedded2, + DestPath: file2, + ChangeMode: structs.TemplateChangeModeSignal, + ChangeSignal: "SIGBUS", + } + + harness := newTestHarness(t, []*structs.Template{template, template2}, true, false) + harness.start(t) + defer harness.stop() + + // Ensure no unblock + select { + case <-harness.mockHooks.UnblockCh: + t.Fatalf("Task unblock should have not have been called") + case <-time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second): + } + + // Write the key to Consul + harness.consul.SetKV(t, key1, []byte(content1_1)) + harness.consul.SetKV(t, key2, []byte(content2_1)) + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + if len(harness.mockHooks.Signals) != 0 { + t.Fatalf("Should not have received any signals: %+v", harness.mockHooks) + } + + // Update the keys in Consul + harness.consul.SetKV(t, key1, []byte(content1_2)) + harness.consul.SetKV(t, key2, []byte(content2_2)) + + // Wait for signals + timeout := time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second) +OUTER: + for { + select { + case <-harness.mockHooks.RestartCh: + t.Fatalf("Restart with signal policy: %+v", harness.mockHooks) + case <-harness.mockHooks.SignalCh: + if len(harness.mockHooks.Signals) != 2 { + continue + } + break OUTER + case <-timeout: + t.Fatalf("Should have received two signals: %+v", harness.mockHooks) + } + } + + // Check the files have been updated + path := filepath.Join(harness.taskDir, file1) + raw, err := ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != content1_2 { + t.Fatalf("Unexpected template data; got %q, want %q", s, content1_2) + } + + path = filepath.Join(harness.taskDir, file2) + raw, err = ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != content2_2 { + t.Fatalf("Unexpected template data; got %q, want %q", s, content2_2) + } +} + +func TestTaskTemplateManager_Rerender_Restart(t *testing.T) { + t.Parallel() + // Make a template that renders based on a key in Consul and sends restart + key1 := "bam" + content1_1 := "cat" + content1_2 := "dog" + embedded1 := fmt.Sprintf(`{{key "%s"}}`, key1) + file1 := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: embedded1, + DestPath: file1, + ChangeMode: structs.TemplateChangeModeRestart, + } + + harness := newTestHarness(t, []*structs.Template{template}, true, false) + harness.start(t) + defer harness.stop() + + // Ensure no unblock + select { + case <-harness.mockHooks.UnblockCh: + t.Fatalf("Task unblock should have not have been called") + case <-time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second): + } + + // Write the key to Consul + harness.consul.SetKV(t, key1, []byte(content1_1)) + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + // Update the keys in Consul + harness.consul.SetKV(t, key1, []byte(content1_2)) + + // Wait for restart + timeout := time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second) +OUTER: + for { + select { + case <-harness.mockHooks.RestartCh: + break OUTER + case <-harness.mockHooks.SignalCh: + t.Fatalf("Signal with restart policy: %+v", harness.mockHooks) + case <-timeout: + t.Fatalf("Should have received a restart: %+v", harness.mockHooks) + } + } + + // Check the files have been updated + path := filepath.Join(harness.taskDir, file1) + raw, err := ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != content1_2 { + t.Fatalf("Unexpected template data; got %q, want %q", s, content1_2) + } +} + +func TestTaskTemplateManager_Interpolate_Destination(t *testing.T) { + t.Parallel() + // Make a template that will have its destination interpolated + content := "hello, world!" + file := "${node.unique.id}.tmpl" + template := &structs.Template{ + EmbeddedTmpl: content, + DestPath: file, + ChangeMode: structs.TemplateChangeModeNoop, + } + + harness := newTestHarness(t, []*structs.Template{template}, false, false) + harness.start(t) + defer harness.stop() + + // Ensure unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + // Check the file is there + actual := fmt.Sprintf("%s.tmpl", harness.node.ID) + path := filepath.Join(harness.taskDir, actual) + raw, err := ioutil.ReadFile(path) + if err != nil { + t.Fatalf("Failed to read rendered template from %q: %v", path, err) + } + + if s := string(raw); s != content { + t.Fatalf("Unexpected template data; got %q, want %q", s, content) + } +} + +func TestTaskTemplateManager_Signal_Error(t *testing.T) { + t.Parallel() + // Make a template that renders based on a key in Consul and sends SIGALRM + key1 := "foo" + content1 := "bar" + content2 := "baz" + embedded1 := fmt.Sprintf(`{{key "%s"}}`, key1) + file1 := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: embedded1, + DestPath: file1, + ChangeMode: structs.TemplateChangeModeSignal, + ChangeSignal: "SIGALRM", + } + + harness := newTestHarness(t, []*structs.Template{template}, true, false) + harness.start(t) + defer harness.stop() + + harness.mockHooks.SignalError = fmt.Errorf("test error") + + // Write the key to Consul + harness.consul.SetKV(t, key1, []byte(content1)) + + // Wait a little + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(2*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Should have received unblock: %+v", harness.mockHooks) + } + + // Write the key to Consul + harness.consul.SetKV(t, key1, []byte(content2)) + + // Wait for kill channel + select { + case <-harness.mockHooks.KillCh: + break + case <-time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Should have received a signals: %+v", harness.mockHooks) + } + + if !strings.Contains(harness.mockHooks.KillReason, "Sending signals") { + t.Fatalf("Unexpected error: %v", harness.mockHooks.KillReason) + } +} + +// TestTaskTemplateManager_Env asserts templates with the env flag set are read +// into the task's environment. +func TestTaskTemplateManager_Env(t *testing.T) { + t.Parallel() + template := &structs.Template{ + EmbeddedTmpl: ` +# Comment lines are ok + +FOO=bar +foo=123 +ANYTHING_goes=Spaces are=ok! +`, + DestPath: "test.env", + ChangeMode: structs.TemplateChangeModeNoop, + Envvars: true, + } + harness := newTestHarness(t, []*structs.Template{template}, true, false) + harness.start(t) + defer harness.stop() + + // Wait a little + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(2*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Should have received unblock: %+v", harness.mockHooks) + } + + // Validate environment + env := harness.envBuilder.Build().Map() + if len(env) < 3 { + t.Fatalf("expected at least 3 env vars but found %d:\n%#v\n", len(env), env) + } + if env["FOO"] != "bar" { + t.Errorf("expected FOO=bar but found %q", env["FOO"]) + } + if env["foo"] != "123" { + t.Errorf("expected foo=123 but found %q", env["foo"]) + } + if env["ANYTHING_goes"] != "Spaces are=ok!" { + t.Errorf("expected ANYTHING_GOES='Spaces are ok!' but found %q", env["ANYTHING_goes"]) + } +} + +// TestTaskTemplateManager_Env_Missing asserts the core env +// template processing function returns errors when files don't exist +func TestTaskTemplateManager_Env_Missing(t *testing.T) { + t.Parallel() + d, err := ioutil.TempDir("", "ct_env_missing") + if err != nil { + t.Fatalf("err: %v", err) + } + defer os.RemoveAll(d) + + // Fake writing the file so we don't have to run the whole template manager + err = ioutil.WriteFile(filepath.Join(d, "exists.env"), []byte("FOO=bar\n"), 0644) + if err != nil { + t.Fatalf("error writing template file: %v", err) + } + + templates := []*structs.Template{ + { + EmbeddedTmpl: "FOO=bar\n", + DestPath: "exists.env", + Envvars: true, + }, + { + EmbeddedTmpl: "WHAT=ever\n", + DestPath: "missing.env", + Envvars: true, + }, + } + + if vars, err := loadTemplateEnv(templates, d); err == nil { + t.Fatalf("expected an error but instead got env vars: %#v", vars) + } +} + +// TestTaskTemplateManager_Env_Multi asserts the core env +// template processing function returns combined env vars from multiple +// templates correctly. +func TestTaskTemplateManager_Env_Multi(t *testing.T) { + t.Parallel() + d, err := ioutil.TempDir("", "ct_env_missing") + if err != nil { + t.Fatalf("err: %v", err) + } + defer os.RemoveAll(d) + + // Fake writing the files so we don't have to run the whole template manager + err = ioutil.WriteFile(filepath.Join(d, "zzz.env"), []byte("FOO=bar\nSHARED=nope\n"), 0644) + if err != nil { + t.Fatalf("error writing template file 1: %v", err) + } + err = ioutil.WriteFile(filepath.Join(d, "aaa.env"), []byte("BAR=foo\nSHARED=yup\n"), 0644) + if err != nil { + t.Fatalf("error writing template file 2: %v", err) + } + + // Templates will get loaded in order (not alpha sorted) + templates := []*structs.Template{ + { + DestPath: "zzz.env", + Envvars: true, + }, + { + DestPath: "aaa.env", + Envvars: true, + }, + } + + vars, err := loadTemplateEnv(templates, d) + if err != nil { + t.Fatalf("expected an error but instead got env vars: %#v", vars) + } + if vars["FOO"] != "bar" { + t.Errorf("expected FOO=bar but found %q", vars["FOO"]) + } + if vars["BAR"] != "foo" { + t.Errorf("expected BAR=foo but found %q", vars["BAR"]) + } + if vars["SHARED"] != "yup" { + t.Errorf("expected FOO=bar but found %q", vars["yup"]) + } +} + +func TestTaskTemplateManager_Rerender_Env(t *testing.T) { + t.Parallel() + // Make a template that renders based on a key in Consul and sends restart + key1 := "bam" + key2 := "bar" + content1_1 := "cat" + content1_2 := "dog" + t1 := &structs.Template{ + EmbeddedTmpl: ` +FOO={{key "bam"}} +`, + DestPath: "test.env", + ChangeMode: structs.TemplateChangeModeRestart, + Envvars: true, + } + t2 := &structs.Template{ + EmbeddedTmpl: ` +BAR={{key "bar"}} +`, + DestPath: "test2.env", + ChangeMode: structs.TemplateChangeModeRestart, + Envvars: true, + } + + harness := newTestHarness(t, []*structs.Template{t1, t2}, true, false) + harness.start(t) + defer harness.stop() + + // Ensure no unblock + select { + case <-harness.mockHooks.UnblockCh: + t.Fatalf("Task unblock should have not have been called") + case <-time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second): + } + + // Write the key to Consul + harness.consul.SetKV(t, key1, []byte(content1_1)) + harness.consul.SetKV(t, key2, []byte(content1_1)) + + // Wait for the unblock + select { + case <-harness.mockHooks.UnblockCh: + case <-time.After(time.Duration(5*testutil.TestMultiplier()) * time.Second): + t.Fatalf("Task unblock should have been called") + } + + env := harness.envBuilder.Build().Map() + if v, ok := env["FOO"]; !ok || v != content1_1 { + t.Fatalf("Bad env for FOO: %v %v", v, ok) + } + if v, ok := env["BAR"]; !ok || v != content1_1 { + t.Fatalf("Bad env for BAR: %v %v", v, ok) + } + + // Update the keys in Consul + harness.consul.SetKV(t, key1, []byte(content1_2)) + + // Wait for restart + timeout := time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second) +OUTER: + for { + select { + case <-harness.mockHooks.RestartCh: + break OUTER + case <-harness.mockHooks.SignalCh: + t.Fatalf("Signal with restart policy: %+v", harness.mockHooks) + case <-timeout: + t.Fatalf("Should have received a restart: %+v", harness.mockHooks) + } + } + + env = harness.envBuilder.Build().Map() + if v, ok := env["FOO"]; !ok || v != content1_2 { + t.Fatalf("Bad env for FOO: %v %v", v, ok) + } + if v, ok := env["BAR"]; !ok || v != content1_1 { + t.Fatalf("Bad env for BAR: %v %v", v, ok) + } +} + +// TestTaskTemplateManager_Config_ServerName asserts the tls_server_name +// setting is propagated to consul-template's configuration. See #2776 +func TestTaskTemplateManager_Config_ServerName(t *testing.T) { + t.Parallel() + c := config.DefaultConfig() + c.VaultConfig = &sconfig.VaultConfig{ + Enabled: helper.BoolToPtr(true), + Addr: "https://localhost/", + TLSServerName: "notlocalhost", + } + config := &TaskTemplateManagerConfig{ + ClientConfig: c, + VaultToken: "token", + } + ctconf, err := newRunnerConfig(config, nil) + if err != nil { + t.Fatalf("unexpected error: %v", err) + } + + if *ctconf.Vault.SSL.ServerName != c.VaultConfig.TLSServerName { + t.Fatalf("expected %q but found %q", c.VaultConfig.TLSServerName, *ctconf.Vault.SSL.ServerName) + } +} + +// TestTaskTemplateManager_Config_VaultGrace asserts the vault_grace setting is +// propagated to consul-template's configuration. +func TestTaskTemplateManager_Config_VaultGrace(t *testing.T) { + t.Parallel() + assert := assert.New(t) + c := config.DefaultConfig() + c.Node = mock.Node() + c.VaultConfig = &sconfig.VaultConfig{ + Enabled: helper.BoolToPtr(true), + Addr: "https://localhost/", + TLSServerName: "notlocalhost", + } + + alloc := mock.Alloc() + config := &TaskTemplateManagerConfig{ + ClientConfig: c, + VaultToken: "token", + + // Make a template that will render immediately + Templates: []*structs.Template{ + { + EmbeddedTmpl: "bar", + DestPath: "foo", + ChangeMode: structs.TemplateChangeModeNoop, + VaultGrace: 10 * time.Second, + }, + { + EmbeddedTmpl: "baz", + DestPath: "bam", + ChangeMode: structs.TemplateChangeModeNoop, + VaultGrace: 100 * time.Second, + }, + }, + EnvBuilder: env.NewBuilder(c.Node, alloc, alloc.Job.TaskGroups[0].Tasks[0], c.Region), + } + + ctmplMapping, err := parseTemplateConfigs(config) + assert.Nil(err, "Parsing Templates") + + ctconf, err := newRunnerConfig(config, ctmplMapping) + assert.Nil(err, "Building Runner Config") + assert.NotNil(ctconf.Vault.Grace, "Vault Grace Pointer") + assert.Equal(10*time.Second, *ctconf.Vault.Grace, "Vault Grace Value") +} + +func TestTaskTemplateManager_BlockedEvents(t *testing.T) { + t.Parallel() + // Make a template that will render based on a key in Consul + var embedded string + for i := 0; i < 5; i++ { + embedded += fmt.Sprintf(`{{key "%d"}}`, i) + } + + file := "my.tmpl" + template := &structs.Template{ + EmbeddedTmpl: embedded, + DestPath: file, + ChangeMode: structs.TemplateChangeModeNoop, + } + + harness := newTestHarness(t, []*structs.Template{template}, true, false) + harness.setEmitRate(100 * time.Millisecond) + harness.start(t) + defer harness.stop() + + // Ensure that we get a blocked event + select { + case <-harness.mockHooks.UnblockCh: + t.Fatalf("Task unblock should have not have been called") + case <-harness.mockHooks.EmitEventCh: + case <-time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second): + t.Fatalf("timeout") + } + + // Check to see we got a correct message + event := harness.mockHooks.Events[0] + if !strings.Contains(event, "and 2 more") { + t.Fatalf("bad event: %q", event) + } + + // Write 3 keys to Consul + for i := 0; i < 3; i++ { + harness.consul.SetKV(t, fmt.Sprintf("%d", i), []byte{0xa}) + } + + // Ensure that we get a blocked event + select { + case <-harness.mockHooks.UnblockCh: + t.Fatalf("Task unblock should have not have been called") + case <-harness.mockHooks.EmitEventCh: + case <-time.After(time.Duration(1*testutil.TestMultiplier()) * time.Second): + t.Fatalf("timeout") + } + + // Check to see we got a correct message + event = harness.mockHooks.Events[len(harness.mockHooks.Events)-1] + if !strings.Contains(event, "Missing") || strings.Contains(event, "more") { + t.Fatalf("bad event: %q", event) + } +} diff --git a/client/allocrunnerv2/taskrunner/template_hook.go b/client/allocrunnerv2/taskrunner/template_hook.go new file mode 100644 index 000000000..38f1a434b --- /dev/null +++ b/client/allocrunnerv2/taskrunner/template_hook.go @@ -0,0 +1,157 @@ +package taskrunner + +import ( + "context" + "fmt" + "sync" + + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + ti "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/template" + "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/driver/env" + "github.com/hashicorp/nomad/nomad/structs" +) + +type templateHookConfig struct { + // logger is used to log + logger log.Logger + + // lifecycle is used to interact with the task's lifecycle + lifecycle ti.TaskLifecycle + + // events is used to emit events + events ti.EventEmitter + + // templates is the set of templates we are managing + templates []*structs.Template + + // clientConfig is the Nomad Client configuration + clientConfig *config.Config + + // envBuilder is the environment variable builder for the task. + envBuilder *env.Builder +} + +type templateHook struct { + config *templateHookConfig + + // logger is used to log + logger log.Logger + + // templateManager is used to manage any consul-templates this task may have + templateManager *template.TaskTemplateManager + managerLock sync.Mutex + + // vaultToken is the current Vault token + vaultToken string + + // taskDir is the task directory + taskDir string +} + +func newTemplateHook(config *templateHookConfig) *templateHook { + h := &templateHook{ + config: config, + } + h.logger = config.logger.Named(h.Name()) + return h +} + +func (*templateHook) Name() string { + return "template" +} + +func (h *templateHook) Prerun(ctx context.Context, req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { + h.managerLock.Lock() + defer h.managerLock.Unlock() + + // If we have already run prerun before exit early. + if h.templateManager != nil { + return nil + } + + // Store the current Vault token and the task directory + h.taskDir = req.TaskDir + h.vaultToken = req.VaultToken + unblockCh, err := h.newManager() + if err != nil { + return err + } + + // Wait for the template to render + select { + case <-ctx.Done(): + case <-unblockCh: + } + + return nil +} + +func (h *templateHook) newManager() (unblock chan struct{}, err error) { + unblock = make(chan struct{}) + m, err := template.NewTaskTemplateManager(&template.TaskTemplateManagerConfig{ + UnblockCh: unblock, + Lifecycle: h.config.lifecycle, + Events: h.config.events, + Templates: h.config.templates, + ClientConfig: h.config.clientConfig, + VaultToken: h.vaultToken, + TaskDir: h.taskDir, + EnvBuilder: h.config.envBuilder, + MaxTemplateEventRate: template.DefaultMaxTemplateEventRate, + }) + if err != nil { + h.logger.Error("failed to create template manager", "error", err) + return nil, err + } + + h.templateManager = m + return unblock, nil +} + +func (h *templateHook) Poststop(ctx context.Context, req *interfaces.TaskPoststopRequest, resp *interfaces.TaskPoststopResponse) error { + h.managerLock.Lock() + defer h.managerLock.Unlock() + + // Shutdown any created template + if h.templateManager != nil { + h.templateManager.Stop() + } + + return nil +} + +// Handle new Vault token +func (h *templateHook) Update(ctx context.Context, req *interfaces.TaskUpdateRequest, resp *interfaces.TaskUpdateResponse) error { + h.managerLock.Lock() + defer h.managerLock.Unlock() + + // Nothing to do + if h.templateManager == nil { + return nil + } + + // Check if the Vault token has changed + if req.VaultToken == h.vaultToken { + return nil + } else { + h.vaultToken = req.VaultToken + } + + // Shutdown the old template + h.templateManager.Stop() + h.templateManager = nil + + // Create the new template + if _, err := h.newManager(); err != nil { + err := fmt.Errorf("failed to build template manager: %v", err) + h.logger.Error("failed to build template manager", "error", err) + // XXX I think we can skip this + // r.setState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskSetupFailure).SetSetupError(err).SetFailsTask(), false) + h.config.lifecycle.Kill(h.Name(), err.Error(), true) + } + + return nil +} diff --git a/client/allocrunnerv2/taskrunner/vault_hook.go b/client/allocrunnerv2/taskrunner/vault_hook.go index 01e896ac6..be0702ab4 100644 --- a/client/allocrunnerv2/taskrunner/vault_hook.go +++ b/client/allocrunnerv2/taskrunner/vault_hook.go @@ -14,6 +14,7 @@ import ( "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + ti "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" ) @@ -50,8 +51,8 @@ func (tr *TaskRunner) updatedVaultToken(token string) { type vaultHookConfig struct { vaultStanza *structs.Vault client vaultclient.VaultClient - events EventEmitter - lifecycle TaskLifecycle + events ti.EventEmitter + lifecycle ti.TaskLifecycle updater vaultTokenUpdateHandler logger log.Logger alloc *structs.Allocation @@ -63,10 +64,10 @@ type vaultHook struct { vaultStanza *structs.Vault // eventEmitter is used to emit events to the task - eventEmitter EventEmitter + eventEmitter ti.EventEmitter // lifecycle is used to signal, restart and kill a task - lifecycle TaskLifecycle + lifecycle ti.TaskLifecycle // updater is used to update the Vault token updater vaultTokenUpdateHandler From 2599cf9d743d53cd0fd3b17285541d7f75586e07 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Mon, 16 Jul 2018 14:57:27 -0700 Subject: [PATCH 020/152] remove comment --- client/allocrunnerv2/taskrunner/template_hook.go | 2 -- 1 file changed, 2 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/template_hook.go b/client/allocrunnerv2/taskrunner/template_hook.go index 38f1a434b..bd5ae732b 100644 --- a/client/allocrunnerv2/taskrunner/template_hook.go +++ b/client/allocrunnerv2/taskrunner/template_hook.go @@ -148,8 +148,6 @@ func (h *templateHook) Update(ctx context.Context, req *interfaces.TaskUpdateReq if _, err := h.newManager(); err != nil { err := fmt.Errorf("failed to build template manager: %v", err) h.logger.Error("failed to build template manager", "error", err) - // XXX I think we can skip this - // r.setState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskSetupFailure).SetSetupError(err).SetFailsTask(), false) h.config.lifecycle.Kill(h.Name(), err.Error(), true) } From 86e81947b430e2d39216d487617bd7e499afbf12 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Mon, 16 Jul 2018 17:19:56 -0700 Subject: [PATCH 021/152] Hook renames --- client/allocrunnerv2/alloc_runner.go | 4 +- client/allocrunnerv2/alloc_runner_hooks.go | 2 +- .../interfaces/task_lifecycle.go | 49 ++--- .../allocrunnerv2/taskrunner/artifact_hook.go | 48 +++++ .../allocrunnerv2/taskrunner/state/state.go | 6 +- .../allocrunnerv2/taskrunner/task_dir_hook.go | 51 ++++++ .../allocrunnerv2/taskrunner/task_runner.go | 18 +- .../taskrunner/task_runner_hooks.go | 169 +++++------------- .../allocrunnerv2/taskrunner/template_hook.go | 4 +- client/allocrunnerv2/taskrunner/vault_hook.go | 8 +- 10 files changed, 193 insertions(+), 166 deletions(-) create mode 100644 client/allocrunnerv2/taskrunner/artifact_hook.go create mode 100644 client/allocrunnerv2/taskrunner/task_dir_hook.go diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 92d930255..f7d2aeabf 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -94,8 +94,8 @@ func (ar *allocRunner) Run() { var err error var taskWaitCh <-chan struct{} - // Run the prerun hooks - // XXX Equivalent to TR.Prerun hook + // Run the prestart hooks + // XXX Equivalent to TR.Prestart hook if err := ar.prerun(); err != nil { ar.logger.Error("prerun failed", "error", err) goto POST diff --git a/client/allocrunnerv2/alloc_runner_hooks.go b/client/allocrunnerv2/alloc_runner_hooks.go index e39d3fbdc..a3cac16f0 100644 --- a/client/allocrunnerv2/alloc_runner_hooks.go +++ b/client/allocrunnerv2/alloc_runner_hooks.go @@ -40,7 +40,7 @@ func (ar *allocRunner) prerun() error { var start time.Time if ar.logger.IsTrace() { start = time.Now() - ar.logger.Trace("running pre-run hook", "name", name, "start", start) + ar.logger.Trace("running prestart hook", "name", name, "start", start) } if err := pre.Prerun(); err != nil { diff --git a/client/allocrunnerv2/interfaces/task_lifecycle.go b/client/allocrunnerv2/interfaces/task_lifecycle.go index ba81544be..c0a11ee71 100644 --- a/client/allocrunnerv2/interfaces/task_lifecycle.go +++ b/client/allocrunnerv2/interfaces/task_lifecycle.go @@ -8,10 +8,13 @@ import ( ) /* - pre-run post-run pre-stop post-stop + prestart poststart exited stop | | | | | | | | --------> run ------> exited ----------> not restart ---------> garbage collect + | + | + kill -> exited -> stop */ @@ -20,7 +23,7 @@ type TaskHook interface { Name() string } -type TaskPrerunRequest struct { +type TaskPrestartRequest struct { // HookData is previously set data by the hook HookData map[string]string @@ -37,7 +40,7 @@ type TaskPrerunRequest struct { TaskEnv *env.TaskEnv } -type TaskPrerunResponse struct { +type TaskPrestartResponse struct { // Env is the environment variables to set for the task Env map[string]string @@ -49,39 +52,35 @@ type TaskPrerunResponse struct { Done bool } -type TaskPrerunHook interface { +type TaskPrestartHook interface { TaskHook - Prerun(context.Context, *TaskPrerunRequest, *TaskPrerunResponse) error + Prestart(context.Context, *TaskPrestartRequest, *TaskPrestartResponse) error } -// XXX If we want consul style hooks, need to have something that runs after the -// tasks starts -type TaskPostrunRequest struct { +type TaskPoststartRequest struct { // Network info } -type TaskPostrunResponse struct{} +type TaskPoststartResponse struct{} -type TaskPostrunHook interface { +type TaskPoststartHook interface { TaskHook - Postrun() error - //Postrun(context.Context, *TaskPostrunRequest, *TaskPostrunResponse) error + Poststart(context.Context, *TaskPoststartRequest, *TaskPoststartResponse) error } -type TaskPoststopRequest struct{} -type TaskPoststopResponse struct{} +type TaskKillRequest struct{} +type TaskKillResponse struct{} -type TaskPoststopHook interface { +type TaskKillHook interface { TaskHook - Postrun(context.Context, *TaskPostrunRequest, *TaskPostrunResponse) error + Kill(context.Context, *TaskKillRequest, *TaskKillResponse) error } -type TaskDestroyRequest struct{} -type TaskDestroyResponse struct{} +type TaskExitedRequest struct{} +type TaskExitedResponse struct{} -type TaskDestroyHook interface { +type TaskExitedHook interface { TaskHook - Destroy() error - //Destroy(context.Context, *TaskDestroyRequest, *TaskDestroyResponse) error + Exited(context.Context, *TaskExitedRequest, *TaskExitedResponse) error } type TaskUpdateRequest struct { @@ -93,3 +92,11 @@ type TaskUpdateHook interface { TaskHook Update(context.Context, *TaskUpdateRequest, *TaskUpdateResponse) error } + +type TaskStopRequest struct{} +type TaskStopResponse struct{} + +type TaskStopHook interface { + TaskHook + Stop(context.Context, *TaskStopRequest, *TaskStopResponse) error +} diff --git a/client/allocrunnerv2/taskrunner/artifact_hook.go b/client/allocrunnerv2/taskrunner/artifact_hook.go new file mode 100644 index 000000000..cceea6544 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/artifact_hook.go @@ -0,0 +1,48 @@ +package taskrunner + +import ( + "context" + "fmt" + + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocrunner/getter" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + ti "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/nomad/structs" +) + +// artifactHook downloads artifacts for a task. +type artifactHook struct { + eventEmitter ti.EventEmitter + logger log.Logger +} + +func newArtifactHook(e ti.EventEmitter, logger log.Logger) *artifactHook { + h := &artifactHook{ + eventEmitter: e, + } + h.logger = logger.Named(h.Name()) + return h +} + +func (*artifactHook) Name() string { + return "artifacts" +} + +func (h *artifactHook) Prestart(ctx context.Context, req *interfaces.TaskPrestartRequest, resp *interfaces.TaskPrestartResponse) error { + h.eventEmitter.SetState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskDownloadingArtifacts)) + + for _, artifact := range req.Task.Artifacts { + //XXX add ctx to GetArtifact to allow cancelling long downloads + if err := getter.GetArtifact(req.TaskEnv, artifact, req.TaskDir); err != nil { + wrapped := fmt.Errorf("failed to download artifact %q: %v", artifact.GetterSource, err) + h.logger.Debug(wrapped.Error()) + h.eventEmitter.SetState(structs.TaskStatePending, + structs.NewTaskEvent(structs.TaskArtifactDownloadFailed).SetDownloadError(wrapped)) + return wrapped + } + } + + resp.Done = true + return nil +} diff --git a/client/allocrunnerv2/taskrunner/state/state.go b/client/allocrunnerv2/taskrunner/state/state.go index 8abb71906..287fa5a8b 100644 --- a/client/allocrunnerv2/taskrunner/state/state.go +++ b/client/allocrunnerv2/taskrunner/state/state.go @@ -38,9 +38,9 @@ func (s *LocalState) Copy() *LocalState { } type HookState struct { - // PrerunDone is true if the hook has run Prerun successfully and does + // Prestart is true if the hook has run Prestart successfully and does // not need to run again - PrerunDone bool + PrestartDone bool Data map[string]string } @@ -56,7 +56,7 @@ func (h *HookState) Equal(o *HookState) bool { return h == o } - if h.PrerunDone != o.PrerunDone { + if h.PrestartDone != o.PrestartDone { return false } diff --git a/client/allocrunnerv2/taskrunner/task_dir_hook.go b/client/allocrunnerv2/taskrunner/task_dir_hook.go new file mode 100644 index 000000000..4d2840227 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/task_dir_hook.go @@ -0,0 +1,51 @@ +package taskrunner + +import ( + "context" + + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + cconfig "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/driver" + "github.com/hashicorp/nomad/nomad/structs" +) + +type taskDirHook struct { + runner *TaskRunner + logger log.Logger +} + +func newTaskDirHook(runner *TaskRunner, logger log.Logger) *taskDirHook { + td := &taskDirHook{ + runner: runner, + } + td.logger = logger.Named(td.Name()) + return td +} + +func (h *taskDirHook) Name() string { + return "task_dir" +} + +func (h *taskDirHook) Prestart(ctx context.Context, req *interfaces.TaskPrestartRequest, resp *interfaces.TaskPrestartResponse) error { + cc := h.runner.clientConfig + chroot := cconfig.DefaultChrootEnv + if len(cc.ChrootEnv) > 0 { + chroot = cc.ChrootEnv + } + + // Emit the event that we are going to be building the task directory + h.runner.SetState("", structs.NewTaskEvent(structs.TaskSetup).SetMessage(structs.TaskBuildingTaskDir)) + + // Build the task directory structure + fsi := h.runner.driver.FSIsolation() + err := h.runner.taskDir.Build(false, chroot, fsi) + if err != nil { + return err + } + + // Update the environment variables based on the built task directory + driver.SetEnvvars(h.runner.envBuilder, fsi, h.runner.taskDir, h.runner.clientConfig) + resp.Done = true + return nil +} diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index f54d5d6bd..a6c57489e 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -229,9 +229,9 @@ func (tr *TaskRunner) Run() { MAIN: for { - // Run the prerun hooks - if err := tr.prerun(); err != nil { - tr.logger.Error("prerun failed", "error", err) + // Run the prestart hooks + if err := tr.prestart(); err != nil { + tr.logger.Error("prestart failed", "error", err) tr.restartTracker.SetStartError(err) goto RESTART } @@ -243,9 +243,9 @@ MAIN: goto RESTART } - // Run the postrun hooks - if err := tr.postrun(); err != nil { - tr.logger.Error("postrun failed", "error", err) + // Run the poststart hooks + if err := tr.poststart(); err != nil { + tr.logger.Error("poststart failed", "error", err) } WAIT: @@ -284,9 +284,9 @@ MAIN: timer.Stop() } - // Run the shutdown hooks - if err := tr.shutdown(); err != nil { - tr.logger.Error("postrun failed", "error", err) + // Run the stop hooks + if err := tr.stop(); err != nil { + tr.logger.Error("stop failed", "error", err) } tr.logger.Debug("task run loop exiting") diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 5205754a1..415fb7f94 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -1,18 +1,11 @@ package taskrunner import ( - "context" "fmt" "time" - log "github.com/hashicorp/go-hclog" - - "github.com/hashicorp/nomad/client/allocrunner/getter" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - ti "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" - cconfig "github.com/hashicorp/nomad/client/config" - "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/nomad/structs" ) @@ -54,37 +47,37 @@ func (tr *TaskRunner) initHooks() { } } -// prerun is used to run the runners prerun hooks. -func (tr *TaskRunner) prerun() error { +// prestart is used to run the runners prestart hooks. +func (tr *TaskRunner) prestart() error { //XXX is this necessary? maybe we should have a generic cancelletion // method instead of peeking into the alloc // Determine if the allocation is terminaland we should avoid running - // pre-run hooks. + // prestart hooks. alloc := tr.Alloc() if alloc.TerminalStatus() { - tr.logger.Trace("skipping pre-run hooks since allocation is terminal") + tr.logger.Trace("skipping prestart hooks since allocation is terminal") return nil } if tr.logger.IsTrace() { start := time.Now() - tr.logger.Trace("running pre-run hooks", "start", start) + tr.logger.Trace("running prestart hooks", "start", start) defer func() { end := time.Now() - tr.logger.Trace("finished pre-run hooks", "end", end, "duration", end.Sub(start)) + tr.logger.Trace("finished prestart hooks", "end", end, "duration", end.Sub(start)) }() } for _, hook := range tr.runnerHooks { - pre, ok := hook.(interfaces.TaskPrerunHook) + pre, ok := hook.(interfaces.TaskPrestartHook) if !ok { - tr.logger.Trace("skipping non-prerun hook", "name", hook.Name()) + tr.logger.Trace("skipping non-prestart hook", "name", hook.Name()) continue } name := pre.Name() // Build the request - req := interfaces.TaskPrerunRequest{ + req := interfaces.TaskPrestartRequest{ Task: tr.Task(), TaskDir: tr.taskDir.Dir, TaskEnv: tr.envBuilder.Build(), @@ -93,24 +86,24 @@ func (tr *TaskRunner) prerun() error { tr.localStateLock.RLock() origHookState := tr.localState.Hooks[name] tr.localStateLock.RUnlock() - if origHookState != nil && origHookState.PrerunDone { - tr.logger.Trace("skipping done prerun hook", "name", pre.Name()) + if origHookState != nil && origHookState.PrestartDone { + tr.logger.Trace("skipping done prestart hook", "name", pre.Name()) continue } req.VaultToken = tr.getVaultToken() - // Time the prerun hook + // Time the prestart hook var start time.Time if tr.logger.IsTrace() { start = time.Now() - tr.logger.Trace("running pre-run hook", "name", name, "start", start) + tr.logger.Trace("running prestart hook", "name", name, "start", start) } - // Run the pre-run hook - var resp interfaces.TaskPrerunResponse - if err := pre.Prerun(tr.ctx, &req, &resp); err != nil { - return structs.WrapRecoverable(fmt.Sprintf("pre-run hook %q failed: %v", name, err), err) + // Run the prestart hook + var resp interfaces.TaskPrestartResponse + if err := pre.Prestart(tr.ctx, &req, &resp); err != nil { + return structs.WrapRecoverable(fmt.Sprintf("prestart hook %q failed: %v", name, err), err) } // Store the hook state @@ -124,7 +117,7 @@ func (tr *TaskRunner) prerun() error { if resp.HookData != nil { hookState.Data = resp.HookData - hookState.PrerunDone = resp.Done + hookState.PrestartDone = resp.Done } tr.localStateLock.Unlock() @@ -143,26 +136,26 @@ func (tr *TaskRunner) prerun() error { if tr.logger.IsTrace() { end := time.Now() - tr.logger.Trace("finished pre-run hooks", "name", name, "end", end, "duration", end.Sub(start)) + tr.logger.Trace("finished prestart hooks", "name", name, "end", end, "duration", end.Sub(start)) } } return nil } -// postrun is used to run the runners postrun hooks. -func (tr *TaskRunner) postrun() error { +// poststart is used to run the runners poststart hooks. +func (tr *TaskRunner) poststart() error { if tr.logger.IsTrace() { start := time.Now() - tr.logger.Trace("running post-run hooks", "start", start) + tr.logger.Trace("running poststart hooks", "start", start) defer func() { end := time.Now() - tr.logger.Trace("finished post-run hooks", "end", end, "duration", end.Sub(start)) + tr.logger.Trace("finished poststart hooks", "end", end, "duration", end.Sub(start)) }() } for _, hook := range tr.runnerHooks { - post, ok := hook.(interfaces.TaskPostrunHook) + post, ok := hook.(interfaces.TaskPoststartHook) if !ok { continue } @@ -171,36 +164,38 @@ func (tr *TaskRunner) postrun() error { var start time.Time if tr.logger.IsTrace() { start = time.Now() - tr.logger.Trace("running post-run hook", "name", name, "start", start) + tr.logger.Trace("running poststart hook", "name", name, "start", start) } + req := interfaces.TaskPoststartRequest{} + var resp interfaces.TaskPoststartResponse // XXX We shouldn't exit on the first one - if err := post.Postrun(); err != nil { - return fmt.Errorf("post-run hook %q failed: %v", name, err) + if err := post.Poststart(tr.ctx, &req, &resp); err != nil { + return fmt.Errorf("poststart hook %q failed: %v", name, err) } if tr.logger.IsTrace() { end := time.Now() - tr.logger.Trace("finished post-run hooks", "name", name, "end", end, "duration", end.Sub(start)) + tr.logger.Trace("finished poststart hooks", "name", name, "end", end, "duration", end.Sub(start)) } } return nil } -// shutdown is used to run the shutdown hooks. -func (tr *TaskRunner) shutdown() error { +// stop is used to run the stop hooks. +func (tr *TaskRunner) stop() error { if tr.logger.IsTrace() { start := time.Now() - tr.logger.Trace("running poststop hooks", "start", start) + tr.logger.Trace("running stop hooks", "start", start) defer func() { end := time.Now() - tr.logger.Trace("finished poststop hooks", "end", end, "duration", end.Sub(start)) + tr.logger.Trace("finished stop hooks", "end", end, "duration", end.Sub(start)) }() } for _, hook := range tr.runnerHooks { - post, ok := hook.(interfaces.TaskDestroyHook) + post, ok := hook.(interfaces.TaskStopHook) if !ok { continue } @@ -209,17 +204,19 @@ func (tr *TaskRunner) shutdown() error { var start time.Time if tr.logger.IsTrace() { start = time.Now() - tr.logger.Trace("running destroy hook", "name", name, "start", start) + tr.logger.Trace("running stop hook", "name", name, "start", start) } + req := interfaces.TaskStopRequest{} + var resp interfaces.TaskStopResponse // XXX We shouldn't exit on the first one - if err := post.Destroy(); err != nil { - return fmt.Errorf("destroy hook %q failed: %v", name, err) + if err := post.Stop(tr.ctx, &req, &resp); err != nil { + return fmt.Errorf("stop hook %q failed: %v", name, err) } if tr.logger.IsTrace() { end := time.Now() - tr.logger.Trace("finished destroy hooks", "name", name, "end", end, "duration", end.Sub(start)) + tr.logger.Trace("finished stop hooks", "name", name, "end", end, "duration", end.Sub(start)) } } @@ -251,7 +248,7 @@ func (tr *TaskRunner) updateHooks() { VaultToken: tr.getVaultToken(), } - // Time the prerun hook + // Time the update hook var start time.Time if tr.logger.IsTrace() { start = time.Now() @@ -271,82 +268,6 @@ func (tr *TaskRunner) updateHooks() { } } -type taskDirHook struct { - runner *TaskRunner - logger log.Logger -} - -func newTaskDirHook(runner *TaskRunner, logger log.Logger) *taskDirHook { - td := &taskDirHook{ - runner: runner, - } - td.logger = logger.Named(td.Name()) - return td -} - -func (h *taskDirHook) Name() string { - return "task_dir" -} - -func (h *taskDirHook) Prerun(ctx context.Context, req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { - cc := h.runner.clientConfig - chroot := cconfig.DefaultChrootEnv - if len(cc.ChrootEnv) > 0 { - chroot = cc.ChrootEnv - } - - // Emit the event that we are going to be building the task directory - h.runner.SetState("", structs.NewTaskEvent(structs.TaskSetup).SetMessage(structs.TaskBuildingTaskDir)) - - // Build the task directory structure - fsi := h.runner.driver.FSIsolation() - err := h.runner.taskDir.Build(false, chroot, fsi) - if err != nil { - return err - } - - // Update the environment variables based on the built task directory - driver.SetEnvvars(h.runner.envBuilder, fsi, h.runner.taskDir, h.runner.clientConfig) - resp.Done = true - return nil -} - -// artifactHook downloads artifacts for a task. -type artifactHook struct { - eventEmitter ti.EventEmitter - logger log.Logger -} - -func newArtifactHook(e ti.EventEmitter, logger log.Logger) *artifactHook { - h := &artifactHook{ - eventEmitter: e, - } - h.logger = logger.Named(h.Name()) - return h -} - -func (*artifactHook) Name() string { - return "artifacts" -} - -func (h *artifactHook) Prerun(ctx context.Context, req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { - h.eventEmitter.SetState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskDownloadingArtifacts)) - - for _, artifact := range req.Task.Artifacts { - //XXX add ctx to GetArtifact to allow cancelling long downloads - if err := getter.GetArtifact(req.TaskEnv, artifact, req.TaskDir); err != nil { - wrapped := fmt.Errorf("failed to download artifact %q: %v", artifact.GetterSource, err) - h.logger.Debug(wrapped.Error()) - h.eventEmitter.SetState(structs.TaskStatePending, - structs.NewTaskEvent(structs.TaskArtifactDownloadFailed).SetDownloadError(wrapped)) - return wrapped - } - } - - resp.Done = true - return nil -} - /* TR Hooks: @@ -377,17 +298,17 @@ Implement: Prestart and Update (for new Vault token) and Destroy Consul Service Reg: Require: Task, interpolation/ENV Return: error -Implement: Postrun, Update, Prestop +Implement: Poststart, Update, Kill, Exited > @alex Dispatch Payload: Require: Alloc Return error -Implement: Prerun +Implement: Prestart > @schmichael Artifacts: Require: Folder structure, task, interpolation/ENV Return: error -Implement: Prerun and Destroy +Implement: Prestart */ diff --git a/client/allocrunnerv2/taskrunner/template_hook.go b/client/allocrunnerv2/taskrunner/template_hook.go index bd5ae732b..56957aa6b 100644 --- a/client/allocrunnerv2/taskrunner/template_hook.go +++ b/client/allocrunnerv2/taskrunner/template_hook.go @@ -63,7 +63,7 @@ func (*templateHook) Name() string { return "template" } -func (h *templateHook) Prerun(ctx context.Context, req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { +func (h *templateHook) Prestart(ctx context.Context, req *interfaces.TaskPrestartRequest, resp *interfaces.TaskPrestartResponse) error { h.managerLock.Lock() defer h.managerLock.Unlock() @@ -111,7 +111,7 @@ func (h *templateHook) newManager() (unblock chan struct{}, err error) { return unblock, nil } -func (h *templateHook) Poststop(ctx context.Context, req *interfaces.TaskPoststopRequest, resp *interfaces.TaskPoststopResponse) error { +func (h *templateHook) Stop(ctx context.Context, req *interfaces.TaskStopRequest, resp *interfaces.TaskStopResponse) error { h.managerLock.Lock() defer h.managerLock.Unlock() diff --git a/client/allocrunnerv2/taskrunner/vault_hook.go b/client/allocrunnerv2/taskrunner/vault_hook.go index be0702ab4..4fa0b6896 100644 --- a/client/allocrunnerv2/taskrunner/vault_hook.go +++ b/client/allocrunnerv2/taskrunner/vault_hook.go @@ -121,9 +121,9 @@ func (*vaultHook) Name() string { return "vault" } -func (h *vaultHook) Prerun(ctx context.Context, req *interfaces.TaskPrerunRequest, resp *interfaces.TaskPrerunResponse) error { - // If we have already run prerun before exit early. We do not use the - // PrerunDone value because we want to recover the token on restoration. +func (h *vaultHook) Prestart(ctx context.Context, req *interfaces.TaskPrestartRequest, resp *interfaces.TaskPrestartResponse) error { + // If we have already run prestart before exit early. We do not use the + // PrestartDone value because we want to recover the token on restoration. first := h.firstRun h.firstRun = false if !first { @@ -160,7 +160,7 @@ func (h *vaultHook) Prerun(ctx context.Context, req *interfaces.TaskPrerunReques return nil } -func (h *vaultHook) Poststop(ctx context.Context, req *interfaces.TaskPoststopRequest, resp *interfaces.TaskPoststopResponse) error { +func (h *vaultHook) Stop(ctx context.Context, req *interfaces.TaskStopRequest, resp *interfaces.TaskStopResponse) error { // Shutdown any created manager h.cancel() return nil From 89b4ba9cc826b00d6e79fe25210e802141e0483b Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Mon, 16 Jul 2018 17:40:04 -0700 Subject: [PATCH 022/152] comments --- client/allocrunnerv2/interfaces/task_lifecycle.go | 10 ++++++++++ 1 file changed, 10 insertions(+) diff --git a/client/allocrunnerv2/interfaces/task_lifecycle.go b/client/allocrunnerv2/interfaces/task_lifecycle.go index c0a11ee71..95393b5f3 100644 --- a/client/allocrunnerv2/interfaces/task_lifecycle.go +++ b/client/allocrunnerv2/interfaces/task_lifecycle.go @@ -54,6 +54,8 @@ type TaskPrestartResponse struct { type TaskPrestartHook interface { TaskHook + + // Prestart is called before the task is started. Prestart(context.Context, *TaskPrestartRequest, *TaskPrestartResponse) error } @@ -64,6 +66,8 @@ type TaskPoststartResponse struct{} type TaskPoststartHook interface { TaskHook + + // Poststart is called after the task has started. Poststart(context.Context, *TaskPoststartRequest, *TaskPoststartResponse) error } @@ -72,6 +76,8 @@ type TaskKillResponse struct{} type TaskKillHook interface { TaskHook + + // Kill is called when a task is going to be killed. Kill(context.Context, *TaskKillRequest, *TaskKillResponse) error } @@ -80,6 +86,8 @@ type TaskExitedResponse struct{} type TaskExitedHook interface { TaskHook + + // Exited is called when a task exits and may or may not be restarted. Exited(context.Context, *TaskExitedRequest, *TaskExitedResponse) error } @@ -98,5 +106,7 @@ type TaskStopResponse struct{} type TaskStopHook interface { TaskHook + + // Stop is called after the task has exited and will not be started again. Stop(context.Context, *TaskStopRequest, *TaskStopResponse) error } From e401c660e70d2fb9e66367c8b756087e20d7d345 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Mon, 16 Jul 2018 14:37:27 -0700 Subject: [PATCH 023/152] Implement lifecycle hooks on the task runner --- .../taskrunner/restarts/restarts.go | 16 +++ .../taskrunner/restarts/restarts_test.go | 10 ++ .../interfaces/task_lifecycle.go | 28 ++++-- client/allocrunnerv2/taskrunner/errors.go | 11 +++ .../taskrunner/interfaces/events.go | 2 +- .../taskrunner/interfaces/lifecycle.go | 15 +-- client/allocrunnerv2/taskrunner/lifecycle.go | 99 +++++++++++++++++-- .../taskrunner/shutdown_delay_hook.go | 36 +++++++ .../allocrunnerv2/taskrunner/task_runner.go | 95 ++++++++++++------ .../taskrunner/task_runner_getters.go | 17 +++- .../taskrunner/task_runner_hooks.go | 48 ++++++++- .../taskrunner/template/template.go | 44 ++++++--- .../taskrunner/template/template_test.go | 43 ++++---- .../allocrunnerv2/taskrunner/template_hook.go | 5 +- client/allocrunnerv2/taskrunner/vault_hook.go | 32 ++++-- nomad/structs/structs.go | 6 ++ 16 files changed, 414 insertions(+), 93 deletions(-) create mode 100644 client/allocrunnerv2/taskrunner/errors.go create mode 100644 client/allocrunnerv2/taskrunner/shutdown_delay_hook.go diff --git a/client/allocrunner/taskrunner/restarts/restarts.go b/client/allocrunner/taskrunner/restarts/restarts.go index 05dd22958..ec7a2594d 100644 --- a/client/allocrunner/taskrunner/restarts/restarts.go +++ b/client/allocrunner/taskrunner/restarts/restarts.go @@ -36,6 +36,7 @@ func NewRestartTracker(policy *structs.RestartPolicy, jobType string) *RestartTr type RestartTracker struct { waitRes *dstructs.WaitResult startErr error + killed bool // Whether the task has been killed restartTriggered bool // Whether the task has been signalled to be restarted failure bool // Whether a failure triggered the restart count int // Current number of attempts. @@ -95,6 +96,14 @@ func (r *RestartTracker) SetRestartTriggered(failure bool) *RestartTracker { return r } +// SetKilled is used to mark that the task has been killed. +func (r *RestartTracker) SetKilled() *RestartTracker { + r.lock.Lock() + defer r.lock.Unlock() + r.killed = true + return r +} + // GetReason returns a human-readable description for the last state returned by // GetState. func (r *RestartTracker) GetReason() string { @@ -123,8 +132,15 @@ func (r *RestartTracker) GetState() (string, time.Duration) { r.waitRes = nil r.restartTriggered = false r.failure = false + r.killed = false }() + // Hot path if task was killed + if r.killed { + r.reason = "" + return structs.TaskKilled, 0 + } + // Hot path if a restart was triggered if r.restartTriggered { r.reason = "" diff --git a/client/allocrunner/taskrunner/restarts/restarts_test.go b/client/allocrunner/taskrunner/restarts/restarts_test.go index 7b8e5ea36..a671df83f 100644 --- a/client/allocrunner/taskrunner/restarts/restarts_test.go +++ b/client/allocrunner/taskrunner/restarts/restarts_test.go @@ -117,6 +117,16 @@ func TestClient_RestartTracker_ZeroAttempts(t *testing.T) { } } +func TestClient_RestartTracker_TaskKilled(t *testing.T) { + t.Parallel() + p := testPolicy(true, structs.RestartPolicyModeFail) + p.Attempts = 0 + rt := NewRestartTracker(p, structs.JobTypeService) + if state, when := rt.SetKilled().GetState(); state != structs.TaskKilled && when != 0 { + t.Fatalf("expect no restart; got %v %v", state, when) + } +} + func TestClient_RestartTracker_RestartTriggered(t *testing.T) { t.Parallel() p := testPolicy(true, structs.RestartPolicyModeFail) diff --git a/client/allocrunnerv2/interfaces/task_lifecycle.go b/client/allocrunnerv2/interfaces/task_lifecycle.go index 95393b5f3..01d098f8a 100644 --- a/client/allocrunnerv2/interfaces/task_lifecycle.go +++ b/client/allocrunnerv2/interfaces/task_lifecycle.go @@ -8,14 +8,28 @@ import ( ) /* - prestart poststart exited stop - | | | | - | | | | - --------> run ------> exited ----------> not restart ---------> garbage collect - | - | - kill -> exited -> stop + Restart + +--------------------------------------------------------+ + | | + | *Update | + | +-------+ | + | | | | + | | | | + | +---v-------+----+ | + +----v----+ | Running | +----+-----+ +--------------+ + | | *Prestart |----------------| *Exited | | *Stop | | + | Pending +-------------> *Poststart run +---^-----------> Exited +-----------> Terminal | + | | | upon entering | | | | NoRestart | | + +---------+ | running | | +----------+ +--------------+ + | | | + +--------+-------+ | + | | + +-----------+ + *Kill + (forces terminal) + +Link: http://stable.ascii-flow.appspot.com/#Draw4489375405966393064/1824429135 */ // TaskHook is a lifecycle hook into the life cycle of a task runner. diff --git a/client/allocrunnerv2/taskrunner/errors.go b/client/allocrunnerv2/taskrunner/errors.go new file mode 100644 index 000000000..bfe9b3dda --- /dev/null +++ b/client/allocrunnerv2/taskrunner/errors.go @@ -0,0 +1,11 @@ +package taskrunner + +import "errors" + +const ( + errTaskNotRunning = "Task not running" +) + +var ( + ErrTaskNotRunning = errors.New(errTaskNotRunning) +) diff --git a/client/allocrunnerv2/taskrunner/interfaces/events.go b/client/allocrunnerv2/taskrunner/interfaces/events.go index 3f908cedf..b9b623dff 100644 --- a/client/allocrunnerv2/taskrunner/interfaces/events.go +++ b/client/allocrunnerv2/taskrunner/interfaces/events.go @@ -4,5 +4,5 @@ import "github.com/hashicorp/nomad/nomad/structs" type EventEmitter interface { SetState(state string, event *structs.TaskEvent) - EmitEvent(source, message string) + EmitEvent(event *structs.TaskEvent) } diff --git a/client/allocrunnerv2/taskrunner/interfaces/lifecycle.go b/client/allocrunnerv2/taskrunner/interfaces/lifecycle.go index e440943a3..10e8edf82 100644 --- a/client/allocrunnerv2/taskrunner/interfaces/lifecycle.go +++ b/client/allocrunnerv2/taskrunner/interfaces/lifecycle.go @@ -1,11 +1,14 @@ package interfaces -import "os" +import ( + "context" + "os" + + "github.com/hashicorp/nomad/nomad/structs" +) -// XXX These should probably all return an error and we should have predefined -// error types for the task not currently running type TaskLifecycle interface { - Restart(source, reason string, failure bool) - Signal(source, reason string, s os.Signal) error - Kill(source, reason string, fail bool) + Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error + Signal(event *structs.TaskEvent, s os.Signal) error + Kill(ctx context.Context, event *structs.TaskEvent) error } diff --git a/client/allocrunnerv2/taskrunner/lifecycle.go b/client/allocrunnerv2/taskrunner/lifecycle.go index 4d3f7bd0f..9de87f528 100644 --- a/client/allocrunnerv2/taskrunner/lifecycle.go +++ b/client/allocrunnerv2/taskrunner/lifecycle.go @@ -1,16 +1,99 @@ package taskrunner -import "os" +import ( + "context" + "os" -func (tr *TaskRunner) Restart(source, reason string, failure bool) { - // TODO -} + "github.com/hashicorp/nomad/nomad/structs" +) + +func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error { + // Grab the handle + handle := tr.getDriverHandle() + + // Check it is running + if handle == nil { + return ErrTaskNotRunning + } + + // Emit the event + tr.EmitEvent(event) + + // Tell the restart tracker that a restart triggered the exit + tr.restartTracker.SetRestartTriggered(failure) + + // Kill the task using an exponential backoff in-case of failures. + destroySuccess, err := tr.handleDestroy(handle) + if !destroySuccess { + // We couldn't successfully destroy the resource created. + tr.logger.Error("failed to kill task. Resources may have been leaked", "error", err) + } + + // Drain the wait channel or wait for the request context to be cancelled + select { + case <-handle.WaitCh(): + case <-ctx.Done(): + return ctx.Err() + } -func (tr *TaskRunner) Signal(source, reason string, s os.Signal) error { - // TODO return nil } -func (tr *TaskRunner) Kill(source, reason string, fail bool) { - // TODO +func (tr *TaskRunner) Signal(event *structs.TaskEvent, s os.Signal) error { + // Grab the handle + handle := tr.getDriverHandle() + + // Check it is running + if handle == nil { + return ErrTaskNotRunning + } + + // Emit the event + tr.EmitEvent(event) + + // Send the signal + return handle.Signal(s) +} + +func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error { + // Grab the handle + handle := tr.getDriverHandle() + + // Check if the handle is running + if handle == nil { + return ErrTaskNotRunning + } + + // Emit the event + tr.EmitEvent(event) + + // Run the hooks prior to killing the task + tr.kill() + + // Tell the restart tracker that the task has been killed + tr.restartTracker.SetKilled() + + // Kill the task using an exponential backoff in-case of failures. + destroySuccess, destroyErr := tr.handleDestroy(handle) + if !destroySuccess { + // We couldn't successfully destroy the resource created. + tr.logger.Error("failed to kill task. Resources may have been leaked", "error", destroyErr) + } + + // Drain the wait channel or wait for the request context to be cancelled + select { + case <-handle.WaitCh(): + case <-ctx.Done(): + } + + // Store that the task has been destroyed and any associated error. + tr.SetState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskKilled).SetKillError(destroyErr)) + + if destroyErr != nil { + return destroyErr + } else if err := ctx.Err(); err != nil { + return err + } + + return nil } diff --git a/client/allocrunnerv2/taskrunner/shutdown_delay_hook.go b/client/allocrunnerv2/taskrunner/shutdown_delay_hook.go new file mode 100644 index 000000000..4bad97136 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/shutdown_delay_hook.go @@ -0,0 +1,36 @@ +package taskrunner + +import ( + "context" + "time" + + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" +) + +// shutdownDelayHook delays shutting down a task between deregistering it from +// Consul and actually killing it. +type shutdownDelayHook struct { + delay time.Duration + logger log.Logger +} + +func newShutdownDelayHook(delay time.Duration, logger log.Logger) *shutdownDelayHook { + h := &shutdownDelayHook{ + delay: delay, + } + h.logger = logger.Named(h.Name()) + return h +} + +func (*shutdownDelayHook) Name() string { + return "shutdown-delay" +} + +func (h *shutdownDelayHook) Kill(ctx context.Context, req *interfaces.TaskKillRequest, resp *interfaces.TaskKillResponse) error { + select { + case <-ctx.Done(): + case <-time.After(h.delay): + } + return nil +} diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index a6c57489e..e5fd74838 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -25,6 +25,20 @@ import ( "golang.org/x/crypto/blake2b" ) +const ( + // killBackoffBaseline is the baseline time for exponential backoff while + // killing a task. + killBackoffBaseline = 5 * time.Second + + // killBackoffLimit is the limit of the exponential backoff for killing + // the task. + killBackoffLimit = 2 * time.Minute + + // killFailureLimit is how many times we will attempt to kill a task before + // giving up and potentially leaking resources. + killFailureLimit = 5 +) + var ( // taskRunnerStateAllKey holds all the task runners state. At the moment // there is no need to split it @@ -79,10 +93,8 @@ type TaskRunner struct { driver driver.Driver // handle is the handle to the currently running driver - handle driver.DriverHandle - //XXX(schmichael) I think the handle is only manipulated in the Restore - // and Run loops, so there's never concurrent access. - //handleLock sync.Mutex + handle driver.DriverHandle + handleLock sync.Mutex // task is the task being run task *structs.Task @@ -226,9 +238,10 @@ func (tr *TaskRunner) initLabels() { func (tr *TaskRunner) Run() { defer close(tr.waitCh) + var handle driver.DriverHandle MAIN: - for { + for tr.ctx.Err() == nil { // Run the prestart hooks if err := tr.prestart(); err != nil { tr.logger.Error("prestart failed", "error", err) @@ -236,6 +249,10 @@ MAIN: goto RESTART } + if tr.ctx.Err() != nil { + break MAIN + } + // Run the task if err := tr.runDriver(); err != nil { tr.logger.Error("running driver failed", "error", err) @@ -248,24 +265,27 @@ MAIN: tr.logger.Error("poststart failed", "error", err) } - WAIT: + // Grab the handle + handle = tr.getDriverHandle() + select { - case waitRes := <-tr.handle.WaitCh(): + case waitRes := <-handle.WaitCh(): + // Clear the handle + tr.setDriverHandle(nil) + + // Store the wait result on the restart tracker tr.restartTracker.SetWaitResult(waitRes) - case _ = <-tr.updateCh: - //XXX Need to copy handleUpdate over - tr.logger.Warn("update not handled") - goto WAIT case <-tr.ctx.Done(): - tr.logger.Debug("task runner cancelled") - break MAIN + tr.logger.Debug("task killed") } + // TODO Need to run exited hooks + RESTART: // Actually restart by sleeping and also watching for destroy events restart, restartWait := tr.shouldRestart() if !restart { - break + break MAIN } deadline := time.Now().Add(restartWait) @@ -273,9 +293,6 @@ MAIN: for time.Now().Before(deadline) { select { case <-timer.C: - case _ = <-tr.updateCh: - //XXX Need to copy handleUpdate over - tr.logger.Warn("update not handled") case <-tr.ctx.Done(): tr.logger.Debug("task runner cancelled") break MAIN @@ -297,6 +314,9 @@ func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { state, when := tr.restartTracker.GetState() reason := tr.restartTracker.GetReason() switch state { + case structs.TaskKilled: + // The task was killed. Nothing to do + return false, 0 case structs.TaskNotRestarting, structs.TaskTerminated: tr.logger.Info("not restarting task", "reason", reason) if state == structs.TaskNotRestarting { @@ -333,8 +353,9 @@ func (tr *TaskRunner) runDriver() error { return err } - // Wait on the handle - tr.handle = sresp.Handle + // Grab the handle + tr.setDriverHandle(sresp.Handle) + //XXX need to capture the driver network // Emit an event that we started @@ -372,6 +393,30 @@ func (tr *TaskRunner) initDriver() error { return nil } +// handleDestroy kills the task handle. In the case that killing fails, +// handleDestroy will retry with an exponential backoff and will give up at a +// given limit. It returns whether the task was destroyed and the error +// associated with the last kill attempt. +func (tr *TaskRunner) handleDestroy(handle driver.DriverHandle) (destroyed bool, err error) { + // Cap the number of times we attempt to kill the task. + for i := 0; i < killFailureLimit; i++ { + if err = handle.Kill(); err != nil { + // Calculate the new backoff + backoff := (1 << (2 * uint64(i))) * killBackoffBaseline + if backoff > killBackoffLimit { + backoff = killBackoffLimit + } + + tr.logger.Error("failed to kill task", "backoff", backoff, "error", err) + time.Sleep(backoff) + } else { + // Kill was successful + return true, nil + } + } + return +} + // persistLocalState persists local state to disk synchronously. func (tr *TaskRunner) persistLocalState() error { // buffer for writing to boltdb @@ -498,8 +543,7 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { //} } -func (tr *TaskRunner) EmitEvent(source, message string) { - event := structs.NewTaskEvent(source).SetMessage(message) +func (tr *TaskRunner) EmitEvent(event *structs.TaskEvent) { tr.SetState("", event) } @@ -522,15 +566,6 @@ func (tr *TaskRunner) Update(update *structs.Allocation) { } } -// Shutdown the task runner. Does not stop the task or garbage collect a -// stopped task. -// -// This method is safe for calling concurently with Run(). Callers must -// receive on WaitCh() to block until Run() has exited. -func (tr *TaskRunner) Shutdown() { - tr.ctxCancel() -} - // appendTaskEvent updates the task status by appending the new event. func appendTaskEvent(state *structs.TaskState, event *structs.TaskEvent) { capacity := 10 diff --git a/client/allocrunnerv2/taskrunner/task_runner_getters.go b/client/allocrunnerv2/taskrunner/task_runner_getters.go index 455cec467..3ee76dff8 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_getters.go +++ b/client/allocrunnerv2/taskrunner/task_runner_getters.go @@ -1,6 +1,9 @@ package taskrunner -import "github.com/hashicorp/nomad/nomad/structs" +import ( + "github.com/hashicorp/nomad/client/driver" + "github.com/hashicorp/nomad/nomad/structs" +) func (tr *TaskRunner) Alloc() *structs.Allocation { tr.allocLock.Lock() @@ -25,3 +28,15 @@ func (tr *TaskRunner) setVaultToken(token string) { defer tr.vaultTokenLock.Unlock() tr.vaultToken = token } + +func (tr *TaskRunner) getDriverHandle() driver.DriverHandle { + tr.handleLock.Lock() + defer tr.handleLock.Unlock() + return tr.handle +} + +func (tr *TaskRunner) setDriverHandle(handle driver.DriverHandle) { + tr.handleLock.Lock() + defer tr.handleLock.Unlock() + tr.handle = handle +} diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 415fb7f94..9323e683e 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -1,6 +1,7 @@ package taskrunner import ( + "context" "fmt" "time" @@ -12,16 +13,18 @@ import ( // initHooks intializes the tasks hooks. func (tr *TaskRunner) initHooks() { hookLogger := tr.logger.Named("task_hook") + task := tr.Task() // Create the task directory hook. This is run first to ensure the // directoy path exists for other hooks. tr.runnerHooks = []interfaces.TaskHook{ newTaskDirHook(tr, hookLogger), newArtifactHook(tr, hookLogger), + newShutdownDelayHook(task.ShutdownDelay, hookLogger), } // If Vault is enabled, add the hook - if task := tr.Task(); task.Vault != nil { + if task.Vault != nil { tr.runnerHooks = append(tr.runnerHooks, newVaultHook(&vaultHookConfig{ vaultStanza: task.Vault, client: tr.vaultClient, @@ -35,7 +38,7 @@ func (tr *TaskRunner) initHooks() { } // If there are templates is enabled, add the hook - if task := tr.Task(); len(task.Templates) != 0 { + if len(task.Templates) != 0 { tr.runnerHooks = append(tr.runnerHooks, newTemplateHook(&templateHookConfig{ logger: hookLogger, lifecycle: tr, @@ -268,6 +271,47 @@ func (tr *TaskRunner) updateHooks() { } } +// kill is used to run the runners kill hooks. +func (tr *TaskRunner) kill() { + if tr.logger.IsTrace() { + start := time.Now() + tr.logger.Trace("running kill hooks", "start", start) + defer func() { + end := time.Now() + tr.logger.Trace("finished kill hooks", "end", end, "duration", end.Sub(start)) + }() + } + + for _, hook := range tr.runnerHooks { + upd, ok := hook.(interfaces.TaskKillHook) + if !ok { + tr.logger.Trace("skipping non-kill hook", "name", hook.Name()) + continue + } + + name := upd.Name() + + // Time the update hook + var start time.Time + if tr.logger.IsTrace() { + start = time.Now() + tr.logger.Trace("running kill hook", "name", name, "start", start) + } + + // Run the update hook + req := interfaces.TaskKillRequest{} + var resp interfaces.TaskKillResponse + if err := upd.Kill(context.Background(), &req, &resp); err != nil { + tr.logger.Error("kill hook failed", "name", name, "error", err) + } + + if tr.logger.IsTrace() { + end := time.Now() + tr.logger.Trace("finished kill hooks", "name", name, "end", end, "duration", end.Sub(start)) + } + } +} + /* TR Hooks: diff --git a/client/allocrunnerv2/taskrunner/template/template.go b/client/allocrunnerv2/taskrunner/template/template.go index 27dba209f..0685d1459 100644 --- a/client/allocrunnerv2/taskrunner/template/template.go +++ b/client/allocrunnerv2/taskrunner/template/template.go @@ -1,6 +1,7 @@ package template import ( + "context" "fmt" "math/rand" "os" @@ -206,7 +207,10 @@ func (tm *TaskTemplateManager) run() { // Read environment variables from env templates before we unblock envMap, err := loadTemplateEnv(tm.config.Templates, tm.config.TaskDir) if err != nil { - tm.config.Lifecycle.Kill(consulTemplateSourceName, err.Error(), true) + tm.config.Lifecycle.Kill(context.Background(), + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Template failed to read environment variables: %v", err))) return } tm.config.EnvBuilder.SetTemplateEnv(envMap) @@ -250,7 +254,10 @@ WAIT: continue } - tm.config.Lifecycle.Kill(consulTemplateSourceName, err.Error(), true) + tm.config.Lifecycle.Kill(context.Background(), + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Template failed: %v", err))) case <-tm.runner.TemplateRenderedCh(): // A template has been rendered, figure out what to do events := tm.runner.RenderEvents() @@ -328,7 +335,7 @@ WAIT: } missingStr := strings.Join(missingSlice, ", ") - tm.config.Events.EmitEvent(consulTemplateSourceName, fmt.Sprintf("Missing: %s", missingStr)) + tm.config.Events.EmitEvent(structs.NewTaskEvent(consulTemplateSourceName).SetDisplayMessage(fmt.Sprintf("Missing: %s", missingStr))) } } } @@ -350,7 +357,10 @@ func (tm *TaskTemplateManager) handleTemplateRerenders(allRenderedTime time.Time continue } - tm.config.Lifecycle.Kill(consulTemplateSourceName, err.Error(), true) + tm.config.Lifecycle.Kill(context.Background(), + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Template failed: %v", err))) case <-tm.runner.TemplateRenderedCh(): // A template has been rendered, figure out what to do var handling []string @@ -375,14 +385,20 @@ func (tm *TaskTemplateManager) handleTemplateRerenders(allRenderedTime time.Time // Lookup the template and determine what to do tmpls, ok := tm.lookup[id] if !ok { - tm.config.Lifecycle.Kill(consulTemplateSourceName, fmt.Sprintf("template runner returned unknown template id %q", id), true) + tm.config.Lifecycle.Kill(context.Background(), + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Template runner returned unknown template id %q", id))) return } // Read environment variables from templates envMap, err := loadTemplateEnv(tm.config.Templates, tm.config.TaskDir) if err != nil { - tm.config.Lifecycle.Kill(consulTemplateSourceName, err.Error(), true) + tm.config.Lifecycle.Kill(context.Background(), + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Template failed to read environment variables: %v", err))) return } tm.config.EnvBuilder.SetTemplateEnv(envMap) @@ -424,13 +440,15 @@ func (tm *TaskTemplateManager) handleTemplateRerenders(allRenderedTime time.Time } if restart { - const failure = false - tm.config.Lifecycle.Restart(consulTemplateSourceName, "template with change_mode restart re-rendered", failure) + tm.config.Lifecycle.Restart(context.Background(), + structs.NewTaskEvent(structs.TaskRestarting). + SetDisplayMessage("Template with change_mode restart re-rendered"), false) } else if len(signals) != 0 { var mErr multierror.Error for signal := range signals { - err := tm.config.Lifecycle.Signal(consulTemplateSourceName, "template re-rendered", tm.signals[signal]) - if err != nil { + s := tm.signals[signal] + event := structs.NewTaskEvent(structs.TaskSignaling).SetTaskSignal(s).SetDisplayMessage("Template re-rendered") + if err := tm.config.Lifecycle.Signal(event, s); err != nil { multierror.Append(&mErr, err) } } @@ -440,7 +458,11 @@ func (tm *TaskTemplateManager) handleTemplateRerenders(allRenderedTime time.Time for signal := range signals { flat = append(flat, tm.signals[signal]) } - tm.config.Lifecycle.Kill(consulTemplateSourceName, fmt.Sprintf("Sending signals %v failed: %v", flat, err), true) + + tm.config.Lifecycle.Kill(context.Background(), + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Template failed to send signals %v: %v", flat, err))) } } } diff --git a/client/allocrunnerv2/taskrunner/template/template_test.go b/client/allocrunnerv2/taskrunner/template/template_test.go index 102f6e745..f4fefeb8f 100644 --- a/client/allocrunnerv2/taskrunner/template/template_test.go +++ b/client/allocrunnerv2/taskrunner/template/template_test.go @@ -1,6 +1,7 @@ package template import ( + "context" "fmt" "io" "io/ioutil" @@ -41,10 +42,10 @@ type MockTaskHooks struct { UnblockCh chan struct{} - KillReason string - KillCh chan struct{} + KillEvent *structs.TaskEvent + KillCh chan struct{} - Events []string + Events []*structs.TaskEvent EmitEventCh chan struct{} } @@ -57,15 +58,16 @@ func NewMockTaskHooks() *MockTaskHooks { EmitEventCh: make(chan struct{}, 1), } } -func (m *MockTaskHooks) Restart(source, reason string, failure bool) { +func (m *MockTaskHooks) Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error { m.Restarts++ select { case m.RestartCh <- struct{}{}: default: } + return nil } -func (m *MockTaskHooks) Signal(source, reason string, s os.Signal) error { +func (m *MockTaskHooks) Signal(event *structs.TaskEvent, s os.Signal) error { m.Signals = append(m.Signals, s) select { case m.SignalCh <- struct{}{}: @@ -75,16 +77,17 @@ func (m *MockTaskHooks) Signal(source, reason string, s os.Signal) error { return m.SignalError } -func (m *MockTaskHooks) Kill(source, reason string, fail bool) { - m.KillReason = reason +func (m *MockTaskHooks) Kill(ctx context.Context, event *structs.TaskEvent) error { + m.KillEvent = event select { case m.KillCh <- struct{}{}: default: } + return nil } -func (m *MockTaskHooks) EmitEvent(source, message string) { - m.Events = append(m.Events, message) +func (m *MockTaskHooks) EmitEvent(event *structs.TaskEvent) { + m.Events = append(m.Events, event) select { case m.EmitEventCh <- struct{}{}: default: @@ -966,6 +969,8 @@ func TestTaskTemplateManager_Interpolate_Destination(t *testing.T) { func TestTaskTemplateManager_Signal_Error(t *testing.T) { t.Parallel() + require := require.New(t) + // Make a template that renders based on a key in Consul and sends SIGALRM key1 := "foo" content1 := "bar" @@ -1006,9 +1011,8 @@ func TestTaskTemplateManager_Signal_Error(t *testing.T) { t.Fatalf("Should have received a signals: %+v", harness.mockHooks) } - if !strings.Contains(harness.mockHooks.KillReason, "Sending signals") { - t.Fatalf("Unexpected error: %v", harness.mockHooks.KillReason) - } + require.NotNil(harness.mockHooks.KillEvent) + require.Contains(harness.mockHooks.KillEvent.DisplayMessage, "failed to send signals") } // TestTaskTemplateManager_Env asserts templates with the env flag set are read @@ -1287,6 +1291,8 @@ func TestTaskTemplateManager_Config_VaultGrace(t *testing.T) { func TestTaskTemplateManager_BlockedEvents(t *testing.T) { t.Parallel() + require := require.New(t) + // Make a template that will render based on a key in Consul var embedded string for i := 0; i < 5; i++ { @@ -1315,10 +1321,8 @@ func TestTaskTemplateManager_BlockedEvents(t *testing.T) { } // Check to see we got a correct message - event := harness.mockHooks.Events[0] - if !strings.Contains(event, "and 2 more") { - t.Fatalf("bad event: %q", event) - } + require.Len(harness.mockHooks.Events, 1) + require.Contains(harness.mockHooks.Events[0].DisplayMessage, "and 2 more") // Write 3 keys to Consul for i := 0; i < 3; i++ { @@ -1334,9 +1338,10 @@ func TestTaskTemplateManager_BlockedEvents(t *testing.T) { t.Fatalf("timeout") } + // TODO // Check to see we got a correct message - event = harness.mockHooks.Events[len(harness.mockHooks.Events)-1] - if !strings.Contains(event, "Missing") || strings.Contains(event, "more") { - t.Fatalf("bad event: %q", event) + eventMsg := harness.mockHooks.Events[len(harness.mockHooks.Events)-1].DisplayMessage + if !strings.Contains(eventMsg, "Missing") || strings.Contains(eventMsg, "more") { + t.Fatalf("bad event: %q", eventMsg) } } diff --git a/client/allocrunnerv2/taskrunner/template_hook.go b/client/allocrunnerv2/taskrunner/template_hook.go index 56957aa6b..4b3073d7c 100644 --- a/client/allocrunnerv2/taskrunner/template_hook.go +++ b/client/allocrunnerv2/taskrunner/template_hook.go @@ -148,7 +148,10 @@ func (h *templateHook) Update(ctx context.Context, req *interfaces.TaskUpdateReq if _, err := h.newManager(); err != nil { err := fmt.Errorf("failed to build template manager: %v", err) h.logger.Error("failed to build template manager", "error", err) - h.config.lifecycle.Kill(h.Name(), err.Error(), true) + h.config.lifecycle.Kill(context.Background(), + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Template update %v", err))) } return nil diff --git a/client/allocrunnerv2/taskrunner/vault_hook.go b/client/allocrunnerv2/taskrunner/vault_hook.go index 4fa0b6896..7b52e8b10 100644 --- a/client/allocrunnerv2/taskrunner/vault_hook.go +++ b/client/allocrunnerv2/taskrunner/vault_hook.go @@ -208,7 +208,10 @@ OUTER: if err := h.writeToken(token); err != nil { errorString := "failed to write Vault token to disk" h.logger.Error(errorString, "error", err) - h.lifecycle.Kill("vault", errorString, true) + h.lifecycle.Kill(h.ctx, + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Vault %v", errorString))) return } } @@ -232,18 +235,27 @@ OUTER: s, err := signals.Parse(h.vaultStanza.ChangeSignal) if err != nil { h.logger.Error("failed to parse signal", "error", err) - h.lifecycle.Kill("vault", fmt.Sprintf("failed to parse signal: %v", err), true) + h.lifecycle.Kill(h.ctx, + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Vault: failed to parse signal: %v", err))) return } - if err := h.lifecycle.Signal("vault", "new Vault token acquired", s); err != nil { + event := structs.NewTaskEvent(structs.TaskSignaling).SetTaskSignal(s).SetDisplayMessage("Vault: new Vault token acquired") + if err := h.lifecycle.Signal(event, s); err != nil { h.logger.Error("failed to send signal", "error", err) - h.lifecycle.Kill("vault", fmt.Sprintf("failed to send signal: %v", err), true) + h.lifecycle.Kill(h.ctx, + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Vault: failed to send signal: %v", err))) return } case structs.VaultChangeModeRestart: const noFailure = false - h.lifecycle.Restart("vault", "new Vault token acquired", noFailure) + h.lifecycle.Restart(h.ctx, + structs.NewTaskEvent(structs.TaskRestarting). + SetDisplayMessage("Vault: new Vault token acquired"), false) case structs.VaultChangeModeNoop: fallthrough default: @@ -289,14 +301,20 @@ func (h *vaultHook) deriveVaultToken() (token string, exit bool) { // Check if this is a server side error if structs.IsServerSide(err) { h.logger.Error("failed to derive Vault token", "error", err, "server_side", true) - h.lifecycle.Kill("vault", fmt.Sprintf("server error deriving vault token: %v", err), true) + h.lifecycle.Kill(h.ctx, + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Vault: server failed to derive vault token: %v", err))) return "", true } // Check if we can't recover from the error if !structs.IsRecoverable(err) { h.logger.Error("failed to derive Vault token", "error", err, "recoverable", false) - h.lifecycle.Kill("vault", fmt.Sprintf("failed to derive token: %v", err), true) + h.lifecycle.Kill(h.ctx, + structs.NewTaskEvent(structs.TaskKilling). + SetFailsTask(). + SetDisplayMessage(fmt.Sprintf("Vault: failed to derive vault token: %v", err))) return "", true } diff --git a/nomad/structs/structs.go b/nomad/structs/structs.go index c5cd06dde..ec0ec4b66 100644 --- a/nomad/structs/structs.go +++ b/nomad/structs/structs.go @@ -5900,6 +5900,12 @@ func (te *TaskEvent) GoString() string { return fmt.Sprintf("%v - %v", te.Time, te.Type) } +// SetDisplayMessage sets the display message of TaskEvent +func (te *TaskEvent) SetDisplayMessage(msg string) *TaskEvent { + te.DisplayMessage = msg + return te +} + // SetMessage sets the message of TaskEvent func (te *TaskEvent) SetMessage(msg string) *TaskEvent { te.Message = msg From 516d641db0d561810b5d051623b57d1fa647475e Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 12 Jul 2018 17:56:52 -0700 Subject: [PATCH 024/152] client: implement all-or-nothing alloc restoration Restoring calls NewAR -> Restore -> Run NewAR now calls NewTR AR.Restore calls TR.Restore AR.Run calls TR.Run --- client/allocrunnerv2/alloc_runner.go | 122 ++++++++++-------- .../allocrunnerv2/taskrunner/state/state.go | 7 + .../allocrunnerv2/taskrunner/task_runner.go | 28 +++- .../taskrunner/task_runner_hooks.go | 3 +- client/client.go | 121 ++++++++--------- client/state/state_database.go | 55 +++++++- 6 files changed, 214 insertions(+), 122 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index f7d2aeabf..e1c554c10 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -13,6 +13,7 @@ import ( "github.com/hashicorp/nomad/client/allocrunnerv2/state" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" "github.com/hashicorp/nomad/client/config" + clientstate "github.com/hashicorp/nomad/client/state" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" @@ -57,12 +58,18 @@ type allocRunner struct { } // NewAllocRunner returns a new allocation runner. -func NewAllocRunner(config *Config) *allocRunner { +func NewAllocRunner(config *Config) (*allocRunner, error) { + alloc := config.Alloc + tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup) + if tg == nil { + return nil, fmt.Errorf("failed to lookup task group %q", alloc.TaskGroup) + } + ar := &allocRunner{ + alloc: alloc, clientConfig: config.ClientConfig, vaultClient: config.Vault, - alloc: config.Alloc, - tasks: make(map[string]*taskrunner.TaskRunner), + tasks: make(map[string]*taskrunner.TaskRunner, len(tg.Tasks)), waitCh: make(chan struct{}), updateCh: make(chan *structs.Allocation), stateDB: config.StateDB, @@ -70,15 +77,44 @@ func NewAllocRunner(config *Config) *allocRunner { // Create alloc dir //XXX update AllocDir to hc log - ar.allocDir = allocdir.NewAllocDir(nil, filepath.Join(config.ClientConfig.AllocDir, config.Alloc.ID)) + ar.allocDir = allocdir.NewAllocDir(nil, filepath.Join(config.ClientConfig.AllocDir, alloc.ID)) // Create the logger based on the allocation ID - ar.logger = config.Logger.With("alloc_id", config.Alloc.ID) + ar.logger = config.Logger.With("alloc_id", alloc.ID) // Initialize the runners hooks. ar.initRunnerHooks() - return ar + // Create the TaskRunners + if err := ar.initTaskRunners(tg.Tasks); err != nil { + return nil, err + } + + return ar, nil +} + +// initTaskRunners creates task runners but does *not* run them. +func (ar *allocRunner) initTaskRunners(tasks []*structs.Task) error { + for _, task := range tasks { + config := &taskrunner.Config{ + Alloc: ar.alloc, + ClientConfig: ar.clientConfig, + Task: task, + TaskDir: ar.allocDir.NewTaskDir(task.Name), + Logger: ar.logger, + StateDB: ar.stateDB, + VaultClient: ar.vaultClient, + } + + // Create, but do not Run, the task runner + tr, err := taskrunner.NewTaskRunner(config) + if err != nil { + return fmt.Errorf("failed creating runner for task %q: %v", task.Name, err) + } + + ar.tasks[task.Name] = tr + } + return nil } func (ar *allocRunner) WaitCh() <-chan struct{} { @@ -91,7 +127,6 @@ func (ar *allocRunner) Run() { // Close the wait channel defer close(ar.waitCh) - var err error var taskWaitCh <-chan struct{} // Run the prestart hooks @@ -102,10 +137,7 @@ func (ar *allocRunner) Run() { } // Run the runners - taskWaitCh, err = ar.runImpl() - if err != nil { - ar.logger.Error("starting tasks failed", "error", err) - } + taskWaitCh = ar.runImpl() for { select { @@ -130,20 +162,9 @@ POST: } // runImpl is used to run the runners. -func (ar *allocRunner) runImpl() (<-chan struct{}, error) { - // Grab the task group - alloc := ar.Alloc() - tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup) - if tg == nil { - // XXX Fail and exit - ar.logger.Error("failed to lookup task group", "task_group", alloc.TaskGroup) - return nil, fmt.Errorf("failed to lookup task group %q", alloc.TaskGroup) - } - - for _, task := range tg.Tasks { - if err := ar.runTask(alloc, task); err != nil { - return nil, err - } +func (ar *allocRunner) runImpl() <-chan struct{} { + for _, task := range ar.tasks { + go task.Run() } // Return a combined WaitCh that is closed when all task runners have @@ -156,32 +177,7 @@ func (ar *allocRunner) runImpl() (<-chan struct{}, error) { } }() - return waitCh, nil -} - -// runTask is used to run a task. -func (ar *allocRunner) runTask(alloc *structs.Allocation, task *structs.Task) error { - // Create the runner - config := &taskrunner.Config{ - Alloc: alloc, - ClientConfig: ar.clientConfig, - Task: task, - TaskDir: ar.allocDir.NewTaskDir(task.Name), - Logger: ar.logger, - StateDB: ar.stateDB, - VaultClient: ar.vaultClient, - } - tr, err := taskrunner.NewTaskRunner(config) - if err != nil { - return err - } - - // Start the runner - go tr.Run() - - // Store the runner - ar.tasks[task.Name] = tr - return nil + return waitCh } // Alloc returns the current allocation being run by this runner. @@ -193,9 +189,29 @@ func (ar *allocRunner) Alloc() *structs.Allocation { } // SaveState does all the state related stuff. Who knows. FIXME -//XXX +//XXX do we need to do periodic syncing? if Saving is only called *before* Run +// *and* within Run -- *and* Updates are applid within Run -- we may be able to +// skip quite a bit of locking? maybe? func (ar *allocRunner) SaveState() error { - return nil + return ar.stateDB.Update(func(tx *bolt.Tx) error { + //XXX Track EvalID to only write alloc on change? + // Write the allocation + return clientstate.PutAllocation(tx, ar.Alloc()) + }) +} + +// Restore state from database. Must be called after NewAllocRunner but before +// Run. +func (ar *allocRunner) Restore() error { + return ar.stateDB.View(func(tx *bolt.Tx) error { + // Restore task runners + for _, tr := range ar.tasks { + if err := tr.Restore(tx); err != nil { + return err + } + } + return nil + }) } // Update the running allocation with a new version received from the server. diff --git a/client/allocrunnerv2/taskrunner/state/state.go b/client/allocrunnerv2/taskrunner/state/state.go index 287fa5a8b..9e3b4d245 100644 --- a/client/allocrunnerv2/taskrunner/state/state.go +++ b/client/allocrunnerv2/taskrunner/state/state.go @@ -5,6 +5,13 @@ import ( "github.com/hashicorp/nomad/helper" ) +var ( + // taskRunnerStateAllKey holds all the task runners state. At the moment + // there is no need to split it + //XXX refactor out of client/state and taskrunner + taskRunnerStateAllKey = []byte("simple-all") +) + // LocalState is Task state which is persisted for use when restarting Nomad // agents. type LocalState struct { diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index e5fd74838..4160eea74 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -18,7 +18,7 @@ import ( "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" - oldstate "github.com/hashicorp/nomad/client/state" + clientstate "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" "github.com/ugorji/go/codec" @@ -42,6 +42,7 @@ const ( var ( // taskRunnerStateAllKey holds all the task runners state. At the moment // there is no need to split it + //XXX refactor out of clientstate and new state taskRunnerStateAllKey = []byte("simple-all") ) @@ -446,16 +447,15 @@ func (tr *TaskRunner) persistLocalState() error { return nil } - // Start the transaction. - return tr.stateDB.Batch(func(tx *bolt.Tx) error { + return tr.stateDB.Update(func(tx *bolt.Tx) error { // Grab the task bucket //XXX move into new state pkg - taskBkt, err := oldstate.GetTaskBucket(tx, tr.allocID, tr.taskName) + taskBkt, err := clientstate.GetTaskBucket(tx, tr.allocID, tr.taskName) if err != nil { return fmt.Errorf("failed to retrieve allocation bucket: %v", err) } - if err := oldstate.PutData(taskBkt, taskRunnerStateAllKey, buf.Bytes()); err != nil { + if err := clientstate.PutData(taskBkt, taskRunnerStateAllKey, buf.Bytes()); err != nil { return fmt.Errorf("failed to write task_runner state: %v", err) } @@ -468,6 +468,24 @@ func (tr *TaskRunner) persistLocalState() error { }) } +// Restore task runner state. Called by AllocRunner.Restore after NewTaskRunner +// but before Run. +func (tr *TaskRunner) Restore(tx *bolt.Tx) error { + bkt, err := clientstate.GetTaskBucket(tx, tr.allocID, tr.taskName) + if err != nil { + return fmt.Errorf("failed to get task %q bucket: %v", tr.taskName, err) + } + + //XXX set persisted hash to avoid immediate write on first use? + var ls state.LocalState + if err := clientstate.GetObject(bkt, taskRunnerStateAllKey, &ls); err != nil { + return fmt.Errorf("failed to read task runner state: %v", err) + } + tr.localState = &ls + + return nil +} + // SetState sets the task runners allocation state. func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { // Ensure the event is populated with human readable strings diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 9323e683e..9affd2231 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -124,8 +124,9 @@ func (tr *TaskRunner) prestart() error { } tr.localStateLock.Unlock() - // Persist local state if the hook state has changed + // Store and persist local state if the hook state has changed if !hookState.Equal(origHookState) { + tr.localState.Hooks[name] = hookState if err := tr.persistLocalState(); err != nil { return err } diff --git a/client/client.go b/client/client.go index be4c1dee7..f08a5b425 100644 --- a/client/client.go +++ b/client/client.go @@ -32,7 +32,7 @@ import ( "github.com/hashicorp/nomad/client/allocrunnerv2" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/servers" - "github.com/hashicorp/nomad/client/state" + clientstate "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/client/stats" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/command/agent/consul" @@ -110,6 +110,8 @@ type AllocRunner interface { SaveState() error Update(*structs.Allocation) Alloc() *structs.Allocation + Restore() error + Run() } // Client is used to implement the client interaction with Nomad. Clients @@ -725,6 +727,7 @@ func (c *Client) restoreState() error { return nil } + //XXX REMOVED! make a note in backward compat / upgrading doc // COMPAT: Remove in 0.7.0 // 0.6.0 transitioned from individual state files to a single bolt-db. // The upgrade path is to: @@ -732,74 +735,72 @@ func (c *Client) restoreState() error { // If so, restore from that and delete old state // Restore using state database - // Allocs holds the IDs of the allocations being restored - var allocs []string - - // Upgrading tracks whether this is a pre 0.6.0 upgrade path - var upgrading bool - - // Scan the directory - allocDir := filepath.Join(c.config.StateDir, "alloc") - list, err := ioutil.ReadDir(allocDir) - if err != nil && !os.IsNotExist(err) { - return fmt.Errorf("failed to list alloc state: %v", err) - } else if err == nil && len(list) != 0 { - upgrading = true - for _, entry := range list { - allocs = append(allocs, entry.Name()) - } - } else { - // Normal path - err := c.stateDB.View(func(tx *bolt.Tx) error { - allocs, err = state.GetAllAllocationIDs(tx) - if err != nil { - return fmt.Errorf("failed to list allocations: %v", err) - } - return nil - }) + // Restore allocations + var allocs []*structs.Allocation + var err error + err = c.stateDB.View(func(tx *bolt.Tx) error { + allocs, err = clientstate.GetAllAllocations(tx) if err != nil { - return err + return fmt.Errorf("failed to list allocations: %v", err) } + return nil + }) + if err != nil { + return err } // Load each alloc back var mErr multierror.Error - for _, id := range allocs { - alloc := &structs.Allocation{ID: id} - - // don't worry about blocking/migrating when restoring - watcher := allocrunner.NoopPrevAlloc{} + for _, alloc := range allocs { + //XXX FIXME create a root logger + logger := hclog.New(&hclog.LoggerOptions{ + Name: "nomad", + Level: hclog.LevelFromString(c.configCopy.LogLevel), + TimeFormat: time.RFC3339, + }) c.configLock.RLock() - ar := allocrunner.NewAllocRunner(c.logger, c.configCopy.Copy(), c.stateDB, c.updateAllocStatus, alloc, c.vaultClient, c.consulService, watcher) + arConf := &allocrunnerv2.Config{ + Alloc: alloc, + Logger: logger, + ClientConfig: c.config, + StateDB: c.stateDB, + } c.configLock.RUnlock() + ar, err := allocrunnerv2.NewAllocRunner(arConf) + if err != nil { + c.logger.Printf("[ERR] client: failed to create alloc %q: %v", alloc.ID, err) + mErr.Errors = append(mErr.Errors, err) + continue + } + + // Restore state + if err := ar.Restore(); err != nil { + c.logger.Printf("[ERR] client: failed to restore alloc %q: %v", alloc.ID, err) + mErr.Errors = append(mErr.Errors, err) + continue + } + + //XXX is this locking necessary? c.allocLock.Lock() - c.allocs[id] = ar + c.allocs[alloc.ID] = ar c.allocLock.Unlock() - - if err := ar.RestoreState(); err != nil { - c.logger.Printf("[ERR] client: failed to restore state for alloc %q: %v", id, err) - mErr.Errors = append(mErr.Errors, err) - } else { - go ar.Run() - - if upgrading { - if err := ar.SaveState(); err != nil { - c.logger.Printf("[WARN] client: initial save state for alloc %q failed: %v", id, err) - } - } - } } - // Delete all the entries - if upgrading { - if err := os.RemoveAll(allocDir); err != nil { - mErr.Errors = append(mErr.Errors, err) - } + // Don't run any allocs if there were any failures + //XXX removing this check would switch from all-or-nothing restores to + // best-effort. went with all-or-nothing for now + if err := mErr.ErrorOrNil(); err != nil { + return err } - return mErr.ErrorOrNil() + // All allocs restored successfully, run them! + for _, ar := range c.allocs { + go ar.Run() + } + + return nil } // saveState is used to snapshot our state into the data dir. @@ -1918,10 +1919,11 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error // The long term fix is to pass in the config and node separately and then // we don't have to do a copy. //ar := allocrunner.NewAllocRunner(c.logger, c.configCopy.Copy(), c.stateDB, c.updateAllocStatus, alloc, c.vaultClient, c.consulService, prevAlloc) - //XXX FIXME + //XXX FIXME create a root logger logger := hclog.New(&hclog.LoggerOptions{ - Name: "nomad", - Level: hclog.LevelFromString(c.configCopy.LogLevel), + Name: "nomad", + Level: hclog.LevelFromString(c.configCopy.LogLevel), + TimeFormat: time.RFC3339, }) c.configLock.RLock() @@ -1934,12 +1936,15 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error } c.configLock.RUnlock() - ar := allocrunnerv2.NewAllocRunner(arConf) + ar, err := allocrunnerv2.NewAllocRunner(arConf) + if err != nil { + return err + } // Store the alloc runner. c.allocs[alloc.ID] = ar - //XXX(schmichael) Why do we do this? + // Initialize local state if err := ar.SaveState(); err != nil { c.logger.Printf("[WARN] client: initial save state for alloc %q failed: %v", alloc.ID, err) } diff --git a/client/state/state_database.go b/client/state/state_database.go index 4c3d9c2ed..0a170b9a3 100644 --- a/client/state/state_database.go +++ b/client/state/state_database.go @@ -184,20 +184,65 @@ func DeleteTaskBucket(tx *bolt.Tx, allocID, taskName string) error { return alloc.DeleteBucket(key) } -func GetAllAllocationIDs(tx *bolt.Tx) ([]string, error) { +//XXX duplicated in arv2?! +var ( + // The following are the key paths written to the state database + allocRunnerStateAllocKey = []byte("alloc") +) + +type allocRunnerAllocState struct { + Alloc *structs.Allocation +} + +func GetAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, error) { allocationsBkt := tx.Bucket(allocationsBucket) if allocationsBkt == nil { + // No allocs return nil, nil } + var allocs []*structs.Allocation + // Create a cursor for iteration. - var allocIDs []string c := allocationsBkt.Cursor() - // Iterate over all the buckets + // Iterate over all the allocation buckets for k, _ := c.First(); k != nil; k, _ = c.Next() { - allocIDs = append(allocIDs, string(k)) + allocID := string(k) + allocBkt := allocationsBkt.Bucket(k) + if allocBkt == nil { + //XXX merr? + return nil, fmt.Errorf("alloc %q missing", allocID) + } + + var allocState allocRunnerAllocState + if err := GetObject(allocBkt, allocRunnerStateAllocKey, &allocState); err != nil { + //XXX merr? + return nil, fmt.Errorf("failed to restore alloc %q: %v", allocID, err) + } + allocs = append(allocs, allocState.Alloc) } - return allocIDs, nil + return allocs, nil +} + +// PutAllocation stores an allocation given a writable transaction. +func PutAllocation(tx *bolt.Tx, alloc *structs.Allocation) error { + // Retrieve the root allocations bucket + allocsBkt, err := tx.CreateBucketIfNotExists(allocationsBucket) + if err != nil { + return err + } + + // Retrieve the specific allocations bucket + key := []byte(alloc.ID) + allocBkt, err := allocsBkt.CreateBucketIfNotExists(key) + if err != nil { + return err + } + + allocState := allocRunnerAllocState{ + Alloc: alloc, + } + return PutObject(allocBkt, allocRunnerStateAllocKey, &allocState) } From 6ebdf532ea721b1535c68fa86c6149062dc2eeaf Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 16 Jul 2018 16:32:37 -0700 Subject: [PATCH 025/152] wip split event emitting and state transitions --- .../allocrunnerv2/taskrunner/task_runner.go | 148 +++++++++++++----- 1 file changed, 106 insertions(+), 42 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 4160eea74..36fc11cfa 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -43,12 +43,15 @@ var ( // taskRunnerStateAllKey holds all the task runners state. At the moment // there is no need to split it //XXX refactor out of clientstate and new state - taskRunnerStateAllKey = []byte("simple-all") + //XXX Old key - going to need to migrate + //taskRunnerStateAllKey = []byte("simple-all") + taskLocalStateKey = []byte("local_state") + taskStateKey = []byte("task_state") ) type TaskRunner struct { - // allocID and taskName are immutable so store a copy to access without - // locks + // allocID and taskName are immutable so these fields may be accessed + // without locks allocID string taskName string @@ -58,7 +61,8 @@ type TaskRunner struct { clientConfig *config.Config // state captures the state of the task for updating the allocation - state *structs.TaskState + state *structs.TaskState + stateLock sync.Mutex // localState captures the node-local state of the task for when the // Nomad agent restarts @@ -368,10 +372,11 @@ func (tr *TaskRunner) runDriver() error { func (tr *TaskRunner) initDriver() error { // Create a task-specific event emitter callback to expose minimal // state to drivers + //XXX Replace with EmitEvent -- no need for a shim eventEmitter := func(m string, args ...interface{}) { msg := fmt.Sprintf(m, args...) tr.logger.Debug("driver event", "event", msg) - tr.SetState("", structs.NewTaskEvent(structs.TaskDriverMessage).SetDriverMessage(msg)) + tr.EmitEvent(structs.NewTaskEvent(structs.TaskDriverMessage).SetDriverMessage(msg)) } alloc := tr.Alloc() @@ -455,7 +460,7 @@ func (tr *TaskRunner) persistLocalState() error { return fmt.Errorf("failed to retrieve allocation bucket: %v", err) } - if err := clientstate.PutData(taskBkt, taskRunnerStateAllKey, buf.Bytes()); err != nil { + if err := clientstate.PutData(taskBkt, taskLocalStateKey, buf.Bytes()); err != nil { return fmt.Errorf("failed to write task_runner state: %v", err) } @@ -469,60 +474,52 @@ func (tr *TaskRunner) persistLocalState() error { } // Restore task runner state. Called by AllocRunner.Restore after NewTaskRunner -// but before Run. +// but before Run so no locks need to be acquired. func (tr *TaskRunner) Restore(tx *bolt.Tx) error { bkt, err := clientstate.GetTaskBucket(tx, tr.allocID, tr.taskName) if err != nil { return fmt.Errorf("failed to get task %q bucket: %v", tr.taskName, err) } + // Restore Local State //XXX set persisted hash to avoid immediate write on first use? var ls state.LocalState - if err := clientstate.GetObject(bkt, taskRunnerStateAllKey, &ls); err != nil { - return fmt.Errorf("failed to read task runner state: %v", err) + if err := clientstate.GetObject(bkt, taskLocalStateKey, &ls); err != nil { + return fmt.Errorf("failed to read local task runner state: %v", err) } tr.localState = &ls + // Restore Task State + var ts structs.TaskState + if err := clientstate.GetObject(bkt, taskStateKey, &ts); err != nil { + return fmt.Errorf("failed to read task state: %v", err) + } + tr.state = &tr + return nil } // SetState sets the task runners allocation state. func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { - // Ensure the event is populated with human readable strings - event.PopulateEventDisplayMessage() + tr.stateLock.Lock() + defer tr.stateLock.Unlock() - task := tr.state + taskState := tr.state - // Update the state of the task + //XXX REMOVE ME AFTER TESTING if state != "" { - task.State = state + panic("SetState must not be called with an empty state") } - // Handle the event - if event == nil { - if event.FailsTask { - task.Failed = true - } + // Append the event + tr.emitEventImpl(tx, event) - if event.Type == structs.TaskRestarting { - if !tr.clientConfig.DisableTaggedMetrics { - metrics.IncrCounterWithLabels([]string{"client", "allocs", "restart"}, 1, tr.baseLabels) - } - //if r.config.BackwardsCompatibleMetrics { - //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "restart"}, 1) - //} - task.Restarts++ - task.LastRestart = time.Unix(0, event.Time) - } - appendTaskEvent(task, event) - } - - // Handle the state transistion. + // Handle the state transition. switch state { case structs.TaskStateRunning: // Capture the start time if it is just starting - if task.State != structs.TaskStateRunning { - task.StartedAt = time.Now().UTC() + if taskState.State != structs.TaskStateRunning { + taskState.StartedAt = time.Now().UTC() if !tr.clientConfig.DisableTaggedMetrics { metrics.IncrCounterWithLabels([]string{"client", "allocs", "running"}, 1, tr.baseLabels) } @@ -532,12 +529,12 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { } case structs.TaskStateDead: // Capture the finished time if not already set - if task.FinishedAt.IsZero() { - task.FinishedAt = time.Now().UTC() + if taskState.FinishedAt.IsZero() { + taskState.FinishedAt = time.Now().UTC() } // Emitting metrics to indicate task complete and failures - if task.Failed { + if taskState.Failed { if !tr.clientConfig.DisableTaggedMetrics { metrics.IncrCounterWithLabels([]string{"client", "allocs", "failed"}, 1, tr.baseLabels) } @@ -554,15 +551,80 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { } } + // Persist the state and event + err := tr.stateDB.Update(func(tx *bolt.Tx) error { + bkt, err := clientstate.GetTaskBucket(tx, tr.allocID, tr.taskName) + if err != nil { + return err + } + + return bkt.PutObject(bkt, taskStateKey, tr.state) + }) + if err != nil { + // Only a warning because the next event/state-transition will + // try to persist it again. + tr.logger.Error("error persisting task state", "error", err, "event", event, "state", state) + } + // Create a copy and notify the alloc runner of the transition - //FIXME + //FIXME <-------START HERE //if err := tr.allocRunner.StateUpdated(tr.state.Copy()); err != nil { //tr.logger.Error("failed to save state", "error", err) //} } +// EmitEvent appends a new TaskEvent to this task's TaskState. The actual +// TaskState.State (pending, running, dead) is *not* updated. Use SetState to +// transition states. +// Events are persisted locally but errors are simply logged. func (tr *TaskRunner) EmitEvent(event *structs.TaskEvent) { - tr.SetState("", event) + tr.stateLock.Lock() + defer tr.stateLock.Unlock() + + tr.emitEventImpl(event) + + // Events that do *not* change task state can be batched. + //XXX Seems like this clamps the maximum transaction latency to 10ms. + err := tr.stateDB.Batch(func(tx *bolt.Tx) error { + bkt, err := clientstate.GetTaskBucket(tx, tr.allocID, tr.taskName) + if err != nil { + return err + } + + return bkt.PutObject(bkt, taskStateKey, tr.state) + }) + if err != nil { + // Only a warning because the next event/state-transition will + // try to persist it again. + tr.logger.Warn("error persisting event", "error", err, "event", event) + } +} + +// emitEventImpl is the implementation of EmitEvent without the locking so it +// can be used from SetState. +func (tr *TaskRunner) emitEventImpl(event *structs.TaskEvent) error { + // Ensure the event is populated with human readable strings + event.PopulateEventDisplayMessage() + + // Propogate failure from event to task state + if event.FailsTask { + tr.state.Failed = true + } + + // Update restart metrics + if event.Type == structs.TaskRestarting { + if !tr.clientConfig.DisableTaggedMetrics { + metrics.IncrCounterWithLabels([]string{"client", "allocs", "restart"}, 1, tr.baseLabels) + } + //if r.config.BackwardsCompatibleMetrics { + //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "restart"}, 1) + //} + tr.state.Restarts++ + tr.state.LastRestart = time.Unix(0, event.Time) + } + + // Append event to slice + appendTaskEvent(tr.state, event) } // WaitCh is closed when TaskRunner.Run exits. @@ -586,9 +648,11 @@ func (tr *TaskRunner) Update(update *structs.Allocation) { // appendTaskEvent updates the task status by appending the new event. func appendTaskEvent(state *structs.TaskState, event *structs.TaskEvent) { - capacity := 10 + const capacity = 10 if state.Events == nil { - state.Events = make([]*structs.TaskEvent, 0, capacity) + state.Events = make([]*structs.TaskEvent, 1, capacity) + state.Events[0] = event + return } // If we hit capacity, then shift it. From e1ba73b515f5d0a9359d2f614db1e8b3c1b1e388 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Tue, 17 Jul 2018 13:42:45 -0700 Subject: [PATCH 026/152] compile --- client/allocrunnerv2/taskrunner/task_runner.go | 11 +++++++---- 1 file changed, 7 insertions(+), 4 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 36fc11cfa..718f62ae6 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -494,7 +494,7 @@ func (tr *TaskRunner) Restore(tx *bolt.Tx) error { if err := clientstate.GetObject(bkt, taskStateKey, &ts); err != nil { return fmt.Errorf("failed to read task state: %v", err) } - tr.state = &tr + tr.state = &ts return nil } @@ -512,7 +512,7 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { } // Append the event - tr.emitEventImpl(tx, event) + tr.emitEventImpl(event) // Handle the state transition. switch state { @@ -558,7 +558,7 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { return err } - return bkt.PutObject(bkt, taskStateKey, tr.state) + return clientstate.PutObject(bkt, taskStateKey, tr.state) }) if err != nil { // Only a warning because the next event/state-transition will @@ -591,8 +591,9 @@ func (tr *TaskRunner) EmitEvent(event *structs.TaskEvent) { return err } - return bkt.PutObject(bkt, taskStateKey, tr.state) + return clientstate.PutObject(bkt, taskStateKey, tr.state) }) + if err != nil { // Only a warning because the next event/state-transition will // try to persist it again. @@ -625,6 +626,8 @@ func (tr *TaskRunner) emitEventImpl(event *structs.TaskEvent) error { // Append event to slice appendTaskEvent(tr.state, event) + + return nil } // WaitCh is closed when TaskRunner.Run exits. From 37790770527785737998e61448b636e739f05c3a Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Tue, 17 Jul 2018 13:48:53 -0700 Subject: [PATCH 027/152] Remove SetState from interface --- client/allocrunnerv2/taskrunner/artifact_hook.go | 5 ++--- client/allocrunnerv2/taskrunner/interfaces/events.go | 1 - client/allocrunnerv2/taskrunner/task_dir_hook.go | 2 +- client/allocrunnerv2/taskrunner/task_runner.go | 2 +- 4 files changed, 4 insertions(+), 6 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/artifact_hook.go b/client/allocrunnerv2/taskrunner/artifact_hook.go index cceea6544..ba3612f82 100644 --- a/client/allocrunnerv2/taskrunner/artifact_hook.go +++ b/client/allocrunnerv2/taskrunner/artifact_hook.go @@ -30,15 +30,14 @@ func (*artifactHook) Name() string { } func (h *artifactHook) Prestart(ctx context.Context, req *interfaces.TaskPrestartRequest, resp *interfaces.TaskPrestartResponse) error { - h.eventEmitter.SetState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskDownloadingArtifacts)) + h.eventEmitter.EmitEvent(structs.NewTaskEvent(structs.TaskDownloadingArtifacts)) for _, artifact := range req.Task.Artifacts { //XXX add ctx to GetArtifact to allow cancelling long downloads if err := getter.GetArtifact(req.TaskEnv, artifact, req.TaskDir); err != nil { wrapped := fmt.Errorf("failed to download artifact %q: %v", artifact.GetterSource, err) h.logger.Debug(wrapped.Error()) - h.eventEmitter.SetState(structs.TaskStatePending, - structs.NewTaskEvent(structs.TaskArtifactDownloadFailed).SetDownloadError(wrapped)) + h.eventEmitter.EmitEvent(structs.NewTaskEvent(structs.TaskArtifactDownloadFailed).SetDownloadError(wrapped)) return wrapped } } diff --git a/client/allocrunnerv2/taskrunner/interfaces/events.go b/client/allocrunnerv2/taskrunner/interfaces/events.go index b9b623dff..dd2fc3d07 100644 --- a/client/allocrunnerv2/taskrunner/interfaces/events.go +++ b/client/allocrunnerv2/taskrunner/interfaces/events.go @@ -3,6 +3,5 @@ package interfaces import "github.com/hashicorp/nomad/nomad/structs" type EventEmitter interface { - SetState(state string, event *structs.TaskEvent) EmitEvent(event *structs.TaskEvent) } diff --git a/client/allocrunnerv2/taskrunner/task_dir_hook.go b/client/allocrunnerv2/taskrunner/task_dir_hook.go index 4d2840227..0f843eec0 100644 --- a/client/allocrunnerv2/taskrunner/task_dir_hook.go +++ b/client/allocrunnerv2/taskrunner/task_dir_hook.go @@ -35,7 +35,7 @@ func (h *taskDirHook) Prestart(ctx context.Context, req *interfaces.TaskPrestart } // Emit the event that we are going to be building the task directory - h.runner.SetState("", structs.NewTaskEvent(structs.TaskSetup).SetMessage(structs.TaskBuildingTaskDir)) + h.runner.EmitEvent(structs.NewTaskEvent(structs.TaskSetup).SetMessage(structs.TaskBuildingTaskDir)) // Build the task directory structure fsi := h.runner.driver.FSIsolation() diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 718f62ae6..0b2ce71c5 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -507,7 +507,7 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { taskState := tr.state //XXX REMOVE ME AFTER TESTING - if state != "" { + if state == "" { panic("SetState must not be called with an empty state") } From 7ddc0eb65c3c5695901299d6d03afb7ffa3b305a Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Tue, 17 Jul 2018 13:57:57 -0700 Subject: [PATCH 028/152] Fix deadlock --- client/allocrunnerv2/alloc_runner.go | 7 +++++++ 1 file changed, 7 insertions(+) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index e1c554c10..ee57269aa 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -1,6 +1,7 @@ package allocrunnerv2 import ( + "context" "fmt" "path/filepath" "sync" @@ -139,10 +140,12 @@ func (ar *allocRunner) Run() { // Run the runners taskWaitCh = ar.runImpl() +MAIN: for { select { case <-taskWaitCh: // TaskRunners have all exited + break MAIN case updated := <-ar.updateCh: // Updated alloc received //XXX Update hooks @@ -231,6 +234,10 @@ func (ar *allocRunner) Update(update *structs.Allocation) { //XXX TODO func (ar *allocRunner) Destroy() { //TODO + + for _, tr := range ar.tasks { + tr.Kill(context.Background(), structs.NewTaskEvent(structs.TaskKilled)) + } } // IsDestroyed returns true if the alloc runner has been destroyed (stopped and From c93cfc89c0233dc38de39120ad6cbb47a3c8c11d Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Wed, 18 Jul 2018 11:43:08 -0700 Subject: [PATCH 029/152] wip --- client/allocrunnerv2/alloc_runner.go | 3 ++- client/allocrunnerv2/alloc_runner_test.go | 3 +++ client/allocrunnerv2/taskrunner/state/state.go | 9 +-------- client/allocrunnerv2/taskrunner/task_runner.go | 8 ++++++++ 4 files changed, 14 insertions(+), 9 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index ee57269aa..04c091267 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -196,8 +196,9 @@ func (ar *allocRunner) Alloc() *structs.Allocation { // *and* within Run -- *and* Updates are applid within Run -- we may be able to // skip quite a bit of locking? maybe? func (ar *allocRunner) SaveState() error { + // XXX Do we move this to the client return ar.stateDB.Update(func(tx *bolt.Tx) error { - //XXX Track EvalID to only write alloc on change? + //XXX Track AllocModifyIndex to only write alloc on change? // Write the allocation return clientstate.PutAllocation(tx, ar.Alloc()) }) diff --git a/client/allocrunnerv2/alloc_runner_test.go b/client/allocrunnerv2/alloc_runner_test.go index af2e5fe10..4ec222614 100644 --- a/client/allocrunnerv2/alloc_runner_test.go +++ b/client/allocrunnerv2/alloc_runner_test.go @@ -1,5 +1,7 @@ package allocrunnerv2 +/* + import ( "testing" @@ -82,3 +84,4 @@ func TestAllocRunner_Postrun_Basic(t *testing.T) { require.True(post.run) require.False(pre.run) } +*/ diff --git a/client/allocrunnerv2/taskrunner/state/state.go b/client/allocrunnerv2/taskrunner/state/state.go index 9e3b4d245..43a73d4bf 100644 --- a/client/allocrunnerv2/taskrunner/state/state.go +++ b/client/allocrunnerv2/taskrunner/state/state.go @@ -5,13 +5,6 @@ import ( "github.com/hashicorp/nomad/helper" ) -var ( - // taskRunnerStateAllKey holds all the task runners state. At the moment - // there is no need to split it - //XXX refactor out of client/state and taskrunner - taskRunnerStateAllKey = []byte("simple-all") -) - // LocalState is Task state which is persisted for use when restarting Nomad // agents. type LocalState struct { @@ -48,7 +41,7 @@ type HookState struct { // Prestart is true if the hook has run Prestart successfully and does // not need to run again PrestartDone bool - Data map[string]string + Data map[string]string } func (h *HookState) Copy() *HookState { diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 0b2ce71c5..e69bb31dd 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -473,6 +473,10 @@ func (tr *TaskRunner) persistLocalState() error { }) } +// XXX If the objects don't exists since the client shutdown before the task +// runner ever saved state, then we should treat it as a new task runner and not +// return an error +// // Restore task runner state. Called by AllocRunner.Restore after NewTaskRunner // but before Run so no locks need to be acquired. func (tr *TaskRunner) Restore(tx *bolt.Tx) error { @@ -496,6 +500,10 @@ func (tr *TaskRunner) Restore(tx *bolt.Tx) error { } tr.state = &ts + // XXX if driver has task { + // tr.restoreDriver() + // } + return nil } From 7f4ec509060ec6baa894d6d9b6a62987e65e0e24 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 18 Jul 2018 17:04:36 -0700 Subject: [PATCH 030/152] missed locking around c.allocs access --- client/client.go | 2 ++ 1 file changed, 2 insertions(+) diff --git a/client/client.go b/client/client.go index f08a5b425..0f48450e3 100644 --- a/client/client.go +++ b/client/client.go @@ -796,9 +796,11 @@ func (c *Client) restoreState() error { } // All allocs restored successfully, run them! + c.allocLock.Lock() for _, ar := range c.allocs { go ar.Run() } + c.allocLock.Unlock() return nil } From 9a63d6103d2449e34ba4b50ab11bfe27fcec441d Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 18 Jul 2018 16:51:50 -0700 Subject: [PATCH 031/152] tr: add validate task hook --- .../taskrunner/task_runner_test.go | 1 + .../taskrunner/task_runner_hooks.go | 1 + .../allocrunnerv2/taskrunner/validate_hook.go | 66 +++++++++++++++++++ .../taskrunner/validate_hook_test.go | 63 ++++++++++++++++++ client/driver/env/env.go | 3 +- 5 files changed, 133 insertions(+), 1 deletion(-) create mode 100644 client/allocrunnerv2/taskrunner/validate_hook.go create mode 100644 client/allocrunnerv2/taskrunner/validate_hook_test.go diff --git a/client/allocrunner/taskrunner/task_runner_test.go b/client/allocrunner/taskrunner/task_runner_test.go index 3503d3ee5..68e082616 100644 --- a/client/allocrunner/taskrunner/task_runner_test.go +++ b/client/allocrunner/taskrunner/task_runner_test.go @@ -664,6 +664,7 @@ func TestTaskRunner_UnregisterConsul_Retries(t *testing.T) { } } +//XXX Ported to allocrunnerv2/task_runner/validate_hook_test.go func TestTaskRunner_Validate_UserEnforcement(t *testing.T) { t.Parallel() ctx := testTaskRunner(t, false) diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 9affd2231..35ebfd474 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -18,6 +18,7 @@ func (tr *TaskRunner) initHooks() { // Create the task directory hook. This is run first to ensure the // directoy path exists for other hooks. tr.runnerHooks = []interfaces.TaskHook{ + newValidateHook(tr.clientConfig, hookLogger), newTaskDirHook(tr, hookLogger), newArtifactHook(tr, hookLogger), newShutdownDelayHook(task.ShutdownDelay, hookLogger), diff --git a/client/allocrunnerv2/taskrunner/validate_hook.go b/client/allocrunnerv2/taskrunner/validate_hook.go new file mode 100644 index 000000000..c17544875 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/validate_hook.go @@ -0,0 +1,66 @@ +package taskrunner + +import ( + "context" + "fmt" + + log "github.com/hashicorp/go-hclog" + multierror "github.com/hashicorp/go-multierror" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/driver/env" + "github.com/hashicorp/nomad/nomad/structs" +) + +// validateHook validates the task is able to be run. +type validateHook struct { + config *config.Config + logger log.Logger +} + +func newValidateHook(config *config.Config, logger log.Logger) *validateHook { + h := &validateHook{ + config: config, + } + h.logger = logger.Named(h.Name()) + return h +} + +func (*validateHook) Name() string { + return "validate" +} + +func (h *validateHook) Prestart(ctx context.Context, req *interfaces.TaskPrestartRequest, resp *interfaces.TaskPrestartResponse) error { + if err := validateTask(req.Task, req.TaskEnv, h.config); err != nil { + return err + } + + resp.Done = true + return nil +} + +func validateTask(task *structs.Task, taskEnv *env.TaskEnv, conf *config.Config) error { + var mErr multierror.Error + + // Validate the user + unallowedUsers := conf.ReadStringListToMapDefault("user.blacklist", config.DefaultUserBlacklist) + checkDrivers := conf.ReadStringListToMapDefault("user.checked_drivers", config.DefaultUserCheckedDrivers) + if _, driverMatch := checkDrivers[task.Driver]; driverMatch { + if _, unallowed := unallowedUsers[task.User]; unallowed { + mErr.Errors = append(mErr.Errors, fmt.Errorf("running as user %q is disallowed", task.User)) + } + } + + // Validate the Service names once they're interpolated + for i, service := range task.Services { + name := taskEnv.ReplaceEnv(service.Name) + if err := service.ValidateName(name); err != nil { + mErr.Errors = append(mErr.Errors, fmt.Errorf("service (%d) failed validation: %v", i, err)) + } + } + + if len(mErr.Errors) == 1 { + return mErr.Errors[0] + } + return mErr.ErrorOrNil() +} diff --git a/client/allocrunnerv2/taskrunner/validate_hook_test.go b/client/allocrunnerv2/taskrunner/validate_hook_test.go new file mode 100644 index 000000000..823f6cdcb --- /dev/null +++ b/client/allocrunnerv2/taskrunner/validate_hook_test.go @@ -0,0 +1,63 @@ +package taskrunner + +import ( + "testing" + + "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/driver/env" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/stretchr/testify/require" +) + +func TestTaskRunner_Validate_UserEnforcement(t *testing.T) { + t.Parallel() + + taskEnv := env.NewEmptyBuilder().Build() + conf := config.DefaultConfig() + + // Try to run as root with exec. + task := &structs.Task{ + Driver: "exec", + User: "root", + } + if err := validateTask(task, taskEnv, conf); err == nil { + t.Fatalf("expected error running as root with exec") + } + + // Try to run a non-blacklisted user with exec. + task.User = "foobar" + require.NoError(t, validateTask(task, taskEnv, conf)) + + // Try to run as root with docker. + task.Driver = "docker" + task.User = "root" + require.NoError(t, validateTask(task, taskEnv, conf)) +} + +func TestTaskRunner_Validate_ServiceName(t *testing.T) { + t.Parallel() + + builder := env.NewEmptyBuilder() + conf := config.DefaultConfig() + + // Create a task with a service for validation + task := &structs.Task{ + Services: []*structs.Service{ + { + Name: "ok", + }, + }, + } + + require.NoError(t, validateTask(task, builder.Build(), conf)) + + // Add an env var that should validate + builder.SetGenericEnv(map[string]string{"FOO": "bar"}) + task.Services[0].Name = "${FOO}" + require.NoError(t, validateTask(task, builder.Build(), conf)) + + // Add an env var that should *not* validate + builder.SetGenericEnv(map[string]string{"BAD": "invalid/in/consul"}) + task.Services[0].Name = "${BAD}" + require.Error(t, validateTask(task, builder.Build(), conf)) +} diff --git a/client/driver/env/env.go b/client/driver/env/env.go index ecd4ea2c8..013623168 100644 --- a/client/driver/env/env.go +++ b/client/driver/env/env.go @@ -247,7 +247,8 @@ func NewBuilder(node *structs.Node, alloc *structs.Allocation, task *structs.Tas // NewEmptyBuilder creates a new environment builder. func NewEmptyBuilder() *Builder { return &Builder{ - mu: &sync.RWMutex{}, + mu: &sync.RWMutex{}, + envvars: make(map[string]string), } } From bc905cc61dbd84d8270d28a85c42873c0106be85 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Wed, 18 Jul 2018 13:45:55 -0700 Subject: [PATCH 032/152] Define and thread through state updating interface --- client/allocrunnerv2/alloc_runner.go | 8 +++++ client/allocrunnerv2/interfaces/runner.go | 13 +++++++-- .../allocrunnerv2/taskrunner/task_runner.go | 29 ++++++++++++------- 3 files changed, 36 insertions(+), 14 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 04c091267..73cb82b60 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -104,6 +104,7 @@ func (ar *allocRunner) initTaskRunners(tasks []*structs.Task) error { TaskDir: ar.allocDir.NewTaskDir(task.Name), Logger: ar.logger, StateDB: ar.stateDB, + StateUpdater: ar, VaultClient: ar.vaultClient, } @@ -218,6 +219,13 @@ func (ar *allocRunner) Restore() error { }) } +// TaskStateUpdated is called when a task's state has been updated. This hook is +// used to compute changes to the alloc's ClientStatus and to update the server +// with the new state. +func (ar *allocRunner) TaskStateUpdated(task string, state *structs.TaskState) error { + return nil +} + // Update the running allocation with a new version received from the server. // // This method is safe for calling concurrently with Run() and does not modify diff --git a/client/allocrunnerv2/interfaces/runner.go b/client/allocrunnerv2/interfaces/runner.go index a888484f7..73f6e9f1f 100644 --- a/client/allocrunnerv2/interfaces/runner.go +++ b/client/allocrunnerv2/interfaces/runner.go @@ -1,6 +1,9 @@ package interfaces -import "github.com/hashicorp/nomad/client/allocrunnerv2/state" +import ( + "github.com/hashicorp/nomad/client/allocrunnerv2/state" + "github.com/hashicorp/nomad/nomad/structs" +) // AllocRunner is the interface for an allocation runner. type AllocRunner interface { @@ -13,8 +16,12 @@ type AllocRunner interface { // State returns a copy of the runners state object State() *state.State + + TaskStateHandler } -// TaskRunner is the interface for a task runner. -type TaskRunner interface { +// TaskStateHandler exposes a handler to be called when a task's state changes +type TaskStateHandler interface { + // TaskStateUpdated is used to emit updated task state + TaskStateUpdated(task string, state *structs.TaskState) error } diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index e69bb31dd..a419fb0f8 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -60,6 +60,9 @@ type TaskRunner struct { clientConfig *config.Config + // stateUpdater is used to emit updated task state + stateUpdater interfaces.TaskStateHandler + // state captures the state of the task for updating the allocation state *structs.TaskState stateLock sync.Mutex @@ -146,6 +149,9 @@ type Config struct { // StateDB is used to store and restore state. StateDB *bolt.DB + + // StateUpdater is used to emit updated task state + StateUpdater interfaces.TaskStateHandler } func NewTaskRunner(config *Config) (*TaskRunner, error) { @@ -170,13 +176,14 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { envBuilder: envBuilder, vaultClient: config.VaultClient, //XXX Make a Copy to avoid races? - state: config.Alloc.TaskStates[config.Task.Name], - localState: config.LocalState, - stateDB: config.StateDB, - ctx: trCtx, - ctxCancel: trCancel, - updateCh: make(chan *structs.Allocation), - waitCh: make(chan struct{}), + state: config.Alloc.TaskStates[config.Task.Name], + localState: config.LocalState, + stateDB: config.StateDB, + stateUpdater: config.StateUpdater, + ctx: trCtx, + ctxCancel: trCancel, + updateCh: make(chan *structs.Allocation), + waitCh: make(chan struct{}), } // Create the logger based on the allocation ID @@ -575,10 +582,9 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { } // Create a copy and notify the alloc runner of the transition - //FIXME <-------START HERE - //if err := tr.allocRunner.StateUpdated(tr.state.Copy()); err != nil { - //tr.logger.Error("failed to save state", "error", err) - //} + if err := tr.stateUpdater.TaskStateUpdated(tr.taskName, tr.state.Copy()); err != nil { + tr.logger.Error("failed to save state", "error", err) + } } // EmitEvent appends a new TaskEvent to this task's TaskState. The actual @@ -620,6 +626,7 @@ func (tr *TaskRunner) emitEventImpl(event *structs.TaskEvent) error { tr.state.Failed = true } + // XXX This seems like a super awkward spot for this? Why not shouldRestart? // Update restart metrics if event.Type == structs.TaskRestarting { if !tr.clientConfig.DisableTaggedMetrics { From fd3bc1bd39c4cbf1805bd7ba103fddf5a01536a0 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Wed, 18 Jul 2018 17:06:44 -0700 Subject: [PATCH 033/152] Update state with server --- client/allocrunnerv2/alloc_runner.go | 166 +++++++++++++++++- client/allocrunnerv2/config.go | 5 +- client/allocrunnerv2/state/state.go | 25 +-- .../allocrunnerv2/taskrunner/task_runner.go | 28 ++- .../taskrunner/task_runner_getters.go | 6 + client/client.go | 12 ++ client/gc_test.go | 3 + client/interfaces/client.go | 14 ++ client/util_test.go | 3 + 9 files changed, 228 insertions(+), 34 deletions(-) create mode 100644 client/interfaces/client.go diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 73cb82b60..4e67e428d 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -5,6 +5,7 @@ import ( "fmt" "path/filepath" "sync" + "time" "github.com/boltdb/bolt" log "github.com/hashicorp/go-hclog" @@ -14,19 +15,27 @@ import ( "github.com/hashicorp/nomad/client/allocrunnerv2/state" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" "github.com/hashicorp/nomad/client/config" + cinterfaces "github.com/hashicorp/nomad/client/interfaces" clientstate "github.com/hashicorp/nomad/client/state" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/client/vaultclient" + "github.com/hashicorp/nomad/helper" "github.com/hashicorp/nomad/nomad/structs" ) // allocRunner is used to run all the tasks in a given allocation type allocRunner struct { + // id is the ID of the allocation. Can be accessed without a lock + id string + // Logger is the logger for the alloc runner. logger log.Logger clientConfig *config.Config + // stateUpdater is used to emit updated task state + stateUpdater cinterfaces.AllocStateHandler + // vaultClient is the used to manage Vault tokens vaultClient vaultclient.VaultClient @@ -40,7 +49,8 @@ type allocRunner struct { //XXX implement for local state // state captures the state of the alloc runner - state *state.State + state *state.State + stateLock sync.RWMutex stateDB *bolt.DB @@ -52,7 +62,8 @@ type allocRunner struct { runnerHooks []interfaces.RunnerHook // tasks are the set of task runners - tasks map[string]*taskrunner.TaskRunner + tasks map[string]*taskrunner.TaskRunner + tasksLock sync.RWMutex // updateCh receives allocation updates via the Update method updateCh chan *structs.Allocation @@ -67,13 +78,16 @@ func NewAllocRunner(config *Config) (*allocRunner, error) { } ar := &allocRunner{ + id: alloc.ID, alloc: alloc, clientConfig: config.ClientConfig, vaultClient: config.Vault, tasks: make(map[string]*taskrunner.TaskRunner, len(tg.Tasks)), waitCh: make(chan struct{}), updateCh: make(chan *structs.Allocation), + state: &state.State{}, stateDB: config.StateDB, + stateUpdater: config.StateUpdater, } // Create alloc dir @@ -222,10 +236,156 @@ func (ar *allocRunner) Restore() error { // TaskStateUpdated is called when a task's state has been updated. This hook is // used to compute changes to the alloc's ClientStatus and to update the server // with the new state. -func (ar *allocRunner) TaskStateUpdated(task string, state *structs.TaskState) error { +func (ar *allocRunner) TaskStateUpdated(taskName string, state *structs.TaskState) error { + if state == nil { + return fmt.Errorf("nil state given to TaskStateUpdated") + } + + // If a task is dead, we potentially want to kill other tasks in the group + if state.State == structs.TaskStateDead { + // Find all tasks that are not the one that is dead and check if the one + // that is dead is a leader + var otherTaskRunners []*taskrunner.TaskRunner + var otherTaskNames []string + leader := false + for name, tr := range ar.tasks { + if name != taskName { + otherTaskRunners = append(otherTaskRunners, tr) + otherTaskNames = append(otherTaskNames, name) + } else if tr.Task().Leader { + leader = true + } + } + + // If the task failed, we should kill all the other tasks in the task group. + if state.Failed { + for _, tr := range otherTaskRunners { + tr.Kill(context.Background(), structs.NewTaskEvent(structs.TaskSiblingFailed).SetFailedSibling(taskName)) + } + if len(otherTaskRunners) > 0 { + ar.logger.Debug("task failure, destroying all tasks", "failed_task", taskName, "destroying", otherTaskNames) + } + } else if leader { + // If the task was a leader task we should kill all the other tasks. + for _, tr := range otherTaskRunners { + tr.Kill(context.Background(), structs.NewTaskEvent(structs.TaskLeaderDead)) + } + if len(otherTaskRunners) > 0 { + ar.logger.Debug("leader task dead, destroying all tasks", "leader_task", taskName, "destroying", otherTaskNames) + } + } + } + + // Gather the state of the other tasks + states := make(map[string]*structs.TaskState, len(ar.tasks)) + for name, tr := range ar.tasks { + if name == taskName { + states[name] = state + } else { + states[name] = tr.TaskState() + } + } + + // Get the client allocation + calloc := ar.clientAlloc(states) + + // Update the server + if err := ar.stateUpdater.AllocStateUpdated(calloc); err != nil { + ar.logger.Error("failed to update remote allocation state", "error", err) + } + return nil } +// clientAlloc takes in the task states and returns an Allocation populated +// with Client specific fields +func (ar *allocRunner) clientAlloc(taskStates map[string]*structs.TaskState) *structs.Allocation { + ar.stateLock.RLock() + defer ar.stateLock.RUnlock() + + a := &structs.Allocation{ + ID: ar.id, + TaskStates: taskStates, + } + + s := ar.state + if d := s.DeploymentStatus; d != nil { + a.DeploymentStatus = d.Copy() + } + + // Compute the ClientStatus + if s.ClientStatus != "" { + // The client status is being forced + a.ClientStatus, a.ClientDescription = s.ClientStatus, s.ClientDescription + } else { + a.ClientStatus, a.ClientDescription = getClientStatus(taskStates) + } + + // If the allocation is terminal, make sure all required fields are properly + // set. + if a.ClientTerminalStatus() { + alloc := ar.Alloc() + + // If we are part of a deployment and the task has failed, mark the + // alloc as unhealthy. This guards against the watcher not be started. + if a.ClientStatus == structs.AllocClientStatusFailed && + alloc.DeploymentID != "" && !a.DeploymentStatus.IsUnhealthy() { + a.DeploymentStatus = &structs.AllocDeploymentStatus{ + Healthy: helper.BoolToPtr(false), + } + } + + // Make sure we have marked the finished at for every task. This is used + // to calculate the reschedule time for failed allocations. + now := time.Now() + for _, task := range alloc.Job.LookupTaskGroup(alloc.TaskGroup).Tasks { + ts, ok := a.TaskStates[task.Name] + if !ok { + ts = &structs.TaskState{} + a.TaskStates[task.Name] = ts + } + if ts.FinishedAt.IsZero() { + ts.FinishedAt = now + } + } + } + + return a +} + +// getClientStatus takes in the task states for a given allocation and computes +// the client status and description +func getClientStatus(taskStates map[string]*structs.TaskState) (status, description string) { + var pending, running, dead, failed bool + for _, state := range taskStates { + switch state.State { + case structs.TaskStateRunning: + running = true + case structs.TaskStatePending: + pending = true + case structs.TaskStateDead: + if state.Failed { + failed = true + } else { + dead = true + } + } + } + + // Determine the alloc status + if failed { + return structs.AllocClientStatusFailed, "Failed tasks" + } else if running { + return structs.AllocClientStatusRunning, "Tasks are running" + } else if pending { + return structs.AllocClientStatusPending, "No tasks have started" + } else if dead { + return structs.AllocClientStatusComplete, "All tasks have completed" + } + + return "", "" +} + // Update the running allocation with a new version received from the server. // // This method is safe for calling concurrently with Run() and does not modify diff --git a/client/allocrunnerv2/config.go b/client/allocrunnerv2/config.go index 9e6395158..2a42cf94a 100644 --- a/client/allocrunnerv2/config.go +++ b/client/allocrunnerv2/config.go @@ -4,6 +4,7 @@ import ( "github.com/boltdb/bolt" log "github.com/hashicorp/go-hclog" clientconfig "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/interfaces" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" ) @@ -25,6 +26,6 @@ type Config struct { // Vault is the Vault client to use to retrieve Vault tokens Vault vaultclient.VaultClient - // XXX Can have a OnStateTransistion hook that we can use to update the - // server + // StateUpdater is used to emit updated task state + StateUpdater interfaces.AllocStateHandler } diff --git a/client/allocrunnerv2/state/state.go b/client/allocrunnerv2/state/state.go index eee301de6..73310b7c1 100644 --- a/client/allocrunnerv2/state/state.go +++ b/client/allocrunnerv2/state/state.go @@ -1,34 +1,19 @@ package state import ( - "sync" - "github.com/hashicorp/nomad/nomad/structs" ) +// XXX Why its own package? // State captures the state of the allocation runner. type State struct { - sync.RWMutex + // ClientStatus captures the overall state of the allocation + ClientStatus string - // ClientState captures the overall state of the allocation - ClientState string - - // ClientDesc is an optional human readable description of the + // ClientDescription is an optional human readable description of the // allocations client state - ClientDesc string + ClientDescription string // DeploymentStatus captures the status of the deployment DeploymentStatus *structs.AllocDeploymentStatus } - -type PersistentState struct { - ClientState string -} - -func (s *State) PersistentState() *PersistentState { - s.RLock() - defer s.RUnlock() - return &PersistentState{ - ClientState: s.ClientState, - } -} diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index a419fb0f8..425246397 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -516,16 +516,30 @@ func (tr *TaskRunner) Restore(tx *bolt.Tx) error { // SetState sets the task runners allocation state. func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { + // Update the local state + stateCopy := tr.setStateLocal(state, event) + + // Create a copy and notify the alloc runner of the transition + if err := tr.stateUpdater.TaskStateUpdated(tr.taskName, stateCopy); err != nil { + tr.logger.Error("failed to update remote state", "error", err) + } +} + +// setStateLocal updates the local in-memory state, persists a copy to disk and returns a +// copy of the task's state. +func (tr *TaskRunner) setStateLocal(state string, event *structs.TaskEvent) *structs.TaskState { tr.stateLock.Lock() defer tr.stateLock.Unlock() - taskState := tr.state - //XXX REMOVE ME AFTER TESTING if state == "" { panic("SetState must not be called with an empty state") } + // Update the task state + taskState := tr.state + taskState.State = state + // Append the event tr.emitEventImpl(event) @@ -567,24 +581,20 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { } // Persist the state and event - err := tr.stateDB.Update(func(tx *bolt.Tx) error { + if err := tr.stateDB.Update(func(tx *bolt.Tx) error { bkt, err := clientstate.GetTaskBucket(tx, tr.allocID, tr.taskName) if err != nil { return err } return clientstate.PutObject(bkt, taskStateKey, tr.state) - }) - if err != nil { + }); err != nil { // Only a warning because the next event/state-transition will // try to persist it again. tr.logger.Error("error persisting task state", "error", err, "event", event, "state", state) } - // Create a copy and notify the alloc runner of the transition - if err := tr.stateUpdater.TaskStateUpdated(tr.taskName, tr.state.Copy()); err != nil { - tr.logger.Error("failed to save state", "error", err) - } + return tr.state.Copy() } // EmitEvent appends a new TaskEvent to this task's TaskState. The actual diff --git a/client/allocrunnerv2/taskrunner/task_runner_getters.go b/client/allocrunnerv2/taskrunner/task_runner_getters.go index 3ee76dff8..d8bb78eba 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_getters.go +++ b/client/allocrunnerv2/taskrunner/task_runner_getters.go @@ -17,6 +17,12 @@ func (tr *TaskRunner) Task() *structs.Task { return tr.task } +func (tr *TaskRunner) TaskState() *structs.TaskState { + tr.stateLock.Lock() + defer tr.stateLock.Unlock() + return tr.state.Copy() +} + func (tr *TaskRunner) getVaultToken() string { tr.vaultTokenLock.Lock() defer tr.vaultTokenLock.Unlock() diff --git a/client/client.go b/client/client.go index 0f48450e3..0a198c24a 100644 --- a/client/client.go +++ b/client/client.go @@ -765,6 +765,7 @@ func (c *Client) restoreState() error { Logger: logger, ClientConfig: c.config, StateDB: c.stateDB, + StateUpdater: c, } c.configLock.RUnlock() @@ -1502,6 +1503,16 @@ func (c *Client) updateNodeStatus() error { return nil } +// XXX combine with below +func (c *Client) AllocStateUpdated(alloc *structs.Allocation) error { + if alloc == nil { + return fmt.Errorf("nil allocation provided to AllocStateUpdated") + } + + c.updateAllocStatus(alloc) + return nil +} + // updateAllocStatus is used to update the status of an allocation func (c *Client) updateAllocStatus(alloc *structs.Allocation) { if alloc.Terminated() { @@ -1935,6 +1946,7 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error ClientConfig: c.config, StateDB: c.stateDB, Vault: c.vaultClient, + StateUpdater: c, } c.configLock.RUnlock() diff --git a/client/gc_test.go b/client/gc_test.go index ab647b727..5ae118047 100644 --- a/client/gc_test.go +++ b/client/gc_test.go @@ -1,5 +1,7 @@ package client +/* +TODO(clientv2) import ( "fmt" "testing" @@ -520,3 +522,4 @@ func TestAllocGarbageCollector_UsedPercentThreshold(t *testing.T) { t.Fatalf("gcAlloc: %v", gcAlloc) } } +*/ diff --git a/client/interfaces/client.go b/client/interfaces/client.go new file mode 100644 index 000000000..6525df693 --- /dev/null +++ b/client/interfaces/client.go @@ -0,0 +1,14 @@ +package interfaces + +import "github.com/hashicorp/nomad/nomad/structs" + +type Client interface { + AllocStateHandler +} + +// AllocStateHandler exposes a handler to be called when a allocation's state changes +type AllocStateHandler interface { + // AllocStateUpdated is used to emit an updated allocation. This allocation + // is stripped to only include client settable fields. + AllocStateUpdated(alloc *structs.Allocation) error +} diff --git a/client/util_test.go b/client/util_test.go index 20e45f5e7..ca16bbeea 100644 --- a/client/util_test.go +++ b/client/util_test.go @@ -1,5 +1,7 @@ package client +/* +TODO(clientv2) import ( "reflect" "testing" @@ -73,3 +75,4 @@ func TestShuffleStrings(t *testing.T) { t.Fatalf("shuffle failed") } } +*/ From a2bf851805cce50c75dc04179d019ecfafb85091 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 19 Jul 2018 10:48:01 -0700 Subject: [PATCH 034/152] no need to TaskStateUpdated to return an error also updated comments --- client/allocrunnerv2/alloc_runner.go | 14 ++++---------- client/allocrunnerv2/interfaces/runner.go | 2 +- client/allocrunnerv2/taskrunner/task_runner.go | 6 ++---- 3 files changed, 7 insertions(+), 15 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 4e67e428d..56fcf12e3 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -233,14 +233,10 @@ func (ar *allocRunner) Restore() error { }) } -// TaskStateUpdated is called when a task's state has been updated. This hook is -// used to compute changes to the alloc's ClientStatus and to update the server -// with the new state. -func (ar *allocRunner) TaskStateUpdated(taskName string, state *structs.TaskState) error { - if state == nil { - return fmt.Errorf("nil state given to TaskStateUpdated") - } - +// TaskStateUpdated is called by TaskRunner when a task's state has been +// updated. This hook is used to compute changes to the alloc's ClientStatus +// and to update the server with the new state. +func (ar *allocRunner) TaskStateUpdated(taskName string, state *structs.TaskState) { // If a task is dead, we potentially want to kill other tasks in the group if state.State == structs.TaskStateDead { // Find all tasks that are not the one that is dead and check if the one @@ -293,8 +289,6 @@ func (ar *allocRunner) TaskStateUpdated(taskName string, state *structs.TaskStat if err := ar.stateUpdater.AllocStateUpdated(calloc); err != nil { ar.logger.Error("failed to update remote allocation state", "error", err) } - - return nil } // clientAlloc takes in the task states and returns an Allocation populated diff --git a/client/allocrunnerv2/interfaces/runner.go b/client/allocrunnerv2/interfaces/runner.go index 73f6e9f1f..3925e2384 100644 --- a/client/allocrunnerv2/interfaces/runner.go +++ b/client/allocrunnerv2/interfaces/runner.go @@ -23,5 +23,5 @@ type AllocRunner interface { // TaskStateHandler exposes a handler to be called when a task's state changes type TaskStateHandler interface { // TaskStateUpdated is used to emit updated task state - TaskStateUpdated(task string, state *structs.TaskState) error + TaskStateUpdated(task string, state *structs.TaskState) } diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 425246397..c22304995 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -519,10 +519,8 @@ func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { // Update the local state stateCopy := tr.setStateLocal(state, event) - // Create a copy and notify the alloc runner of the transition - if err := tr.stateUpdater.TaskStateUpdated(tr.taskName, stateCopy); err != nil { - tr.logger.Error("failed to update remote state", "error", err) - } + // Notify the alloc runner of the transition + tr.stateUpdater.TaskStateUpdated(tr.taskName, stateCopy) } // setStateLocal updates the local in-memory state, persists a copy to disk and returns a From ce04915c9fc9a8e727db66d121a2b39fadadf12e Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 19 Jul 2018 10:49:46 -0700 Subject: [PATCH 035/152] log before killing tasks --- client/allocrunnerv2/alloc_runner.go | 12 ++++++------ 1 file changed, 6 insertions(+), 6 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 56fcf12e3..1cd75eeaf 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -255,20 +255,20 @@ func (ar *allocRunner) TaskStateUpdated(taskName string, state *structs.TaskStat // If the task failed, we should kill all the other tasks in the task group. if state.Failed { - for _, tr := range otherTaskRunners { - tr.Kill(context.Background(), structs.NewTaskEvent(structs.TaskSiblingFailed).SetFailedSibling(taskName)) - } if len(otherTaskRunners) > 0 { ar.logger.Debug("task failure, destroying all tasks", "failed_task", taskName, "destroying", otherTaskNames) } + for _, tr := range otherTaskRunners { + tr.Kill(context.Background(), structs.NewTaskEvent(structs.TaskSiblingFailed).SetFailedSibling(taskName)) + } } else if leader { + if len(otherTaskRunners) > 0 { + ar.logger.Debug("leader task dead, destroying all tasks", "leader_task", taskName, "destroying", otherTaskNames) + } // If the task was a leader task we should kill all the other tasks. for _, tr := range otherTaskRunners { tr.Kill(context.Background(), structs.NewTaskEvent(structs.TaskLeaderDead)) } - if len(otherTaskRunners) > 0 { - ar.logger.Debug("leader task dead, destroying all tasks", "leader_task", taskName, "destroying", otherTaskNames) - } } } From 5be982e674990682626caa25230e4cfadfafb70b Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 19 Jul 2018 17:40:10 -0700 Subject: [PATCH 036/152] restore vault client --- client/client.go | 1 + 1 file changed, 1 insertion(+) diff --git a/client/client.go b/client/client.go index 0a198c24a..898b67cd8 100644 --- a/client/client.go +++ b/client/client.go @@ -766,6 +766,7 @@ func (c *Client) restoreState() error { ClientConfig: c.config, StateDB: c.stateDB, StateUpdater: c, + Vault: c.vaultClient, } c.configLock.RUnlock() From a4b4d7b2664206ae0c5a3583ab38dbe2804dfe54 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 19 Jul 2018 17:40:25 -0700 Subject: [PATCH 037/152] consul service hook Deregistration works but difficult to test due to terminal updates not being fully implemented in the new client/ar/tr. --- client/allocrunner/taskrunner/task_runner.go | 53 ++--- client/allocrunnerv2/alloc_runner.go | 7 + client/allocrunnerv2/config.go | 4 + .../interfaces/task_lifecycle.go | 18 +- client/allocrunnerv2/taskrunner/lifecycle.go | 6 +- .../allocrunnerv2/taskrunner/service_hook.go | 199 ++++++++++++++++++ .../allocrunnerv2/taskrunner/task_runner.go | 26 ++- .../taskrunner/task_runner_getters.go | 19 +- .../taskrunner/task_runner_hooks.go | 79 ++++++- client/client.go | 2 + command/agent/consul/check_watcher.go | 19 +- nomad/structs/structs.go | 21 +- 12 files changed, 399 insertions(+), 54 deletions(-) create mode 100644 client/allocrunnerv2/taskrunner/service_hook.go diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index 469f69109..9a4f6d360 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -28,7 +28,6 @@ import ( "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/client/vaultclient" - "github.com/hashicorp/nomad/command/agent/consul" "github.com/hashicorp/nomad/nomad/structs" "github.com/ugorji/go/codec" @@ -1495,14 +1494,15 @@ func (r *TaskRunner) startTask() error { // registerServices and checks with Consul. func (r *TaskRunner) registerServices(d driver.Driver, h driver.DriverHandle, n *cstructs.DriverNetwork) error { - var exec driver.ScriptExecutor - if d.Abilities().Exec { - // Allow set the script executor if the driver supports it - exec = h - } - interpolatedTask := interpolateServices(r.envBuilder.Build(), r.task) - taskServices := consul.NewTaskServices(r.alloc, interpolatedTask, r, exec, n) - return r.consul.RegisterTask(taskServices) + //var exec driver.ScriptExecutor + //if d.Abilities().Exec { + // // Allow set the script executor if the driver supports it + // exec = h + //} + //interpolatedTask := interpolateServices(r.envBuilder.Build(), r.task) + //taskServices := consul.NewTaskServices(r.alloc, interpolatedTask, r, exec, n) + panic("XXX broken during transition to allocrunnerv2") + return r.consul.RegisterTask(nil) } // interpolateServices interpolates tags in a service and checks with values from the @@ -1706,31 +1706,34 @@ func (r *TaskRunner) updateServices(d driver.Driver, h driver.ScriptExecutor, oldAlloc *structs.Allocation, oldTask *structs.Task, newAlloc *structs.Allocation, newTask *structs.Task) error { - var exec driver.ScriptExecutor - if d.Abilities().Exec { - // Allow set the script executor if the driver supports it - exec = h - } - r.driverNetLock.Lock() - net := r.driverNet.Copy() - r.driverNetLock.Unlock() - oldTaskServices := consul.NewTaskServices(oldAlloc, oldTask, r, exec, net) - newTaskServices := consul.NewTaskServices(newAlloc, newTask, r, exec, net) - return r.consul.UpdateTask(oldTaskServices, newTaskServices) + //var exec driver.ScriptExecutor + //if d.Abilities().Exec { + // // Allow set the script executor if the driver supports it + // exec = h + //} + //r.driverNetLock.Lock() + //net := r.driverNet.Copy() + //r.driverNetLock.Unlock() + //oldTaskServices := consul.NewTaskServices(oldAlloc, oldTask, r, exec, net) + //newTaskServices := consul.NewTaskServices(newAlloc, newTask, r, exec, net) + panic("XXX broken during transition to allocrunnerv2") + //return r.consul.UpdateTask(oldTaskServices, newTaskServices) + return r.consul.UpdateTask(nil, nil) } // removeServices and checks from Consul. Handles interpolation and deleting // Canary=true and Canary=false versions in case Canary=false is set at the // same time as the alloc is stopped. func (r *TaskRunner) removeServices() { - interpTask := interpolateServices(r.envBuilder.Build(), r.task) - taskServices := consul.NewTaskServices(r.alloc, interpTask, r, nil, nil) - r.consul.RemoveTask(taskServices) + panic("XXX broken during transition to allocrunnerv2") + //interpTask := interpolateServices(r.envBuilder.Build(), r.task) + //taskServices := consul.NewTaskServices(r.alloc, interpTask, r, nil, nil) + //r.consul.RemoveTask(taskServices) // Flip Canary and remove again in case canary is getting flipped at // the same time as the alloc is being destroyed - taskServices.Canary = !taskServices.Canary - r.consul.RemoveTask(taskServices) + //taskServices.Canary = !taskServices.Canary + //r.consul.RemoveTask(taskServices) } // handleDestroy kills the task handle. In the case that killing fails, diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 1cd75eeaf..a9cfbea1f 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -15,6 +15,7 @@ import ( "github.com/hashicorp/nomad/client/allocrunnerv2/state" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/consul" cinterfaces "github.com/hashicorp/nomad/client/interfaces" clientstate "github.com/hashicorp/nomad/client/state" cstructs "github.com/hashicorp/nomad/client/structs" @@ -36,6 +37,10 @@ type allocRunner struct { // stateUpdater is used to emit updated task state stateUpdater cinterfaces.AllocStateHandler + // consulClient is the client used by the consul service hook for + // registering services and checks + consulClient consul.ConsulServiceAPI + // vaultClient is the used to manage Vault tokens vaultClient vaultclient.VaultClient @@ -81,6 +86,7 @@ func NewAllocRunner(config *Config) (*allocRunner, error) { id: alloc.ID, alloc: alloc, clientConfig: config.ClientConfig, + consulClient: config.Consul, vaultClient: config.Vault, tasks: make(map[string]*taskrunner.TaskRunner, len(tg.Tasks)), waitCh: make(chan struct{}), @@ -119,6 +125,7 @@ func (ar *allocRunner) initTaskRunners(tasks []*structs.Task) error { Logger: ar.logger, StateDB: ar.stateDB, StateUpdater: ar, + Consul: ar.consulClient, VaultClient: ar.vaultClient, } diff --git a/client/allocrunnerv2/config.go b/client/allocrunnerv2/config.go index 2a42cf94a..ebf955a29 100644 --- a/client/allocrunnerv2/config.go +++ b/client/allocrunnerv2/config.go @@ -4,6 +4,7 @@ import ( "github.com/boltdb/bolt" log "github.com/hashicorp/go-hclog" clientconfig "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/interfaces" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" @@ -23,6 +24,9 @@ type Config struct { // StateDB is used to store and restore state. StateDB *bolt.DB + // Consul is the Consul client used to register task services and checks + Consul consul.ConsulServiceAPI + // Vault is the Vault client to use to retrieve Vault tokens Vault vaultclient.VaultClient diff --git a/client/allocrunnerv2/interfaces/task_lifecycle.go b/client/allocrunnerv2/interfaces/task_lifecycle.go index 01d098f8a..4de78c37a 100644 --- a/client/allocrunnerv2/interfaces/task_lifecycle.go +++ b/client/allocrunnerv2/interfaces/task_lifecycle.go @@ -3,7 +3,9 @@ package interfaces import ( "context" + "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" + cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" ) @@ -74,7 +76,14 @@ type TaskPrestartHook interface { } type TaskPoststartRequest struct { - // Network info + // Exec hook (may be nil) + DriverExec driver.ScriptExecutor + + // Network info (may be nil) + DriverNetwork *cstructs.DriverNetwork + + // TaskEnv is the task's environment + TaskEnv *env.TaskEnv } type TaskPoststartResponse struct{} @@ -107,6 +116,13 @@ type TaskExitedHook interface { type TaskUpdateRequest struct { VaultToken string + + // Alloc is the current version of the allocation (may have been + // updated since the hook was created) + Alloc *structs.Allocation + + // TaskEnv is the task's environment + TaskEnv *env.TaskEnv } type TaskUpdateResponse struct{} diff --git a/client/allocrunnerv2/taskrunner/lifecycle.go b/client/allocrunnerv2/taskrunner/lifecycle.go index 9de87f528..db4bc4d6e 100644 --- a/client/allocrunnerv2/taskrunner/lifecycle.go +++ b/client/allocrunnerv2/taskrunner/lifecycle.go @@ -9,7 +9,7 @@ import ( func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error { // Grab the handle - handle := tr.getDriverHandle() + handle, _ := tr.getDriverHandle() // Check it is running if handle == nil { @@ -41,7 +41,7 @@ func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, fai func (tr *TaskRunner) Signal(event *structs.TaskEvent, s os.Signal) error { // Grab the handle - handle := tr.getDriverHandle() + handle, _ := tr.getDriverHandle() // Check it is running if handle == nil { @@ -57,7 +57,7 @@ func (tr *TaskRunner) Signal(event *structs.TaskEvent, s os.Signal) error { func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error { // Grab the handle - handle := tr.getDriverHandle() + handle, _ := tr.getDriverHandle() // Check if the handle is running if handle == nil { diff --git a/client/allocrunnerv2/taskrunner/service_hook.go b/client/allocrunnerv2/taskrunner/service_hook.go new file mode 100644 index 000000000..c1097e36e --- /dev/null +++ b/client/allocrunnerv2/taskrunner/service_hook.go @@ -0,0 +1,199 @@ +package taskrunner + +import ( + "context" + "fmt" + "sync" + + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/consul" + "github.com/hashicorp/nomad/client/driver" + "github.com/hashicorp/nomad/client/driver/env" + cstructs "github.com/hashicorp/nomad/client/structs" + agentconsul "github.com/hashicorp/nomad/command/agent/consul" + "github.com/hashicorp/nomad/nomad/structs" +) + +type serviceHookConfig struct { + alloc *structs.Allocation + task *structs.Task + consul consul.ConsulServiceAPI + + // Restarter is a subset of the TaskLifecycle interface + restarter agentconsul.TaskRestarter + + logger log.Logger +} + +type serviceHook struct { + consul consul.ConsulServiceAPI + allocID string + taskName string + restarter agentconsul.TaskRestarter + logger log.Logger + + // The following fields may be updated + driverExec driver.ScriptExecutor + driverNet *cstructs.DriverNetwork + canary bool + services []*structs.Service + networks structs.Networks + taskEnv *env.TaskEnv + + // Since Update() may be called concurrently with any other hook all + // hook methods must be fully serialized + mu sync.Mutex +} + +func newServiceHook(c serviceHookConfig) *serviceHook { + h := &serviceHook{ + consul: c.consul, + allocID: c.alloc.ID, + taskName: c.task.Name, + services: c.task.Services, + restarter: c.restarter, + } + + if res := c.alloc.TaskResources[c.task.Name]; res != nil { + h.networks = res.Networks + } + + if c.alloc.DeploymentStatus != nil && c.alloc.DeploymentStatus.Canary { + h.canary = true + } + + h.logger = c.logger.Named(h.Name()) + return h +} + +func (h *serviceHook) Name() string { + return "consul_services" +} + +func (h *serviceHook) Poststart(ctx context.Context, req *interfaces.TaskPoststartRequest, _ *interfaces.TaskPoststartResponse) error { + h.mu.Lock() + defer h.mu.Unlock() + + // Store the TaskEnv for interpolating now and when Updating + h.driverExec = req.DriverExec + h.driverNet = req.DriverNetwork + h.taskEnv = req.TaskEnv + + // Create task services struct with request's driver metadata + taskServices := h.getTaskServices() + + return h.consul.RegisterTask(taskServices) +} + +func (h *serviceHook) Update(ctx context.Context, req *interfaces.TaskUpdateRequest, _ *interfaces.TaskUpdateResponse) error { + h.mu.Lock() + defer h.mu.Unlock() + + // Create old task services struct with request's driver metadata as it + // can't change due to Updates + oldTaskServices := h.getTaskServices() + + // Store new updated values out of request + canary := false + if req.Alloc.DeploymentStatus != nil { + canary = req.Alloc.DeploymentStatus.Canary + } + + var networks structs.Networks + if res := req.Alloc.TaskResources[h.taskName]; res != nil { + networks = res.Networks + } + + task := req.Alloc.LookupTask(h.taskName) + if task == nil { + return fmt.Errorf("task %q not found in updated alloc", h.taskName) + } + + // Update service hook fields + h.taskEnv = req.TaskEnv + h.services = task.Services + h.networks = networks + h.canary = canary + + // Create new task services struct with those new values + newTaskServices := h.getTaskServices() + + return h.consul.UpdateTask(oldTaskServices, newTaskServices) +} + +func (h *serviceHook) Exited(ctx context.Context, req *interfaces.TaskExitedRequest, _ *interfaces.TaskExitedResponse) error { + h.mu.Lock() + defer h.mu.Unlock() + + taskServices := h.getTaskServices() + h.consul.RemoveTask(taskServices) + + // Canary flag may be getting flipped when the alloc is being + // destroyed, so remove both variations of the service + taskServices.Canary = !taskServices.Canary + h.consul.RemoveTask(taskServices) + + return nil +} + +func (h *serviceHook) getTaskServices() *agentconsul.TaskServices { + // Interpolate with the task's environment + interpolatedServices := interpolateServices(h.taskEnv, h.services) + + // Create task services struct with request's driver metadata + return &agentconsul.TaskServices{ + AllocID: h.allocID, + Name: h.taskName, + Restarter: h.restarter, + Services: interpolatedServices, + DriverExec: h.driverExec, + DriverNetwork: h.driverNet, + Networks: h.networks, + Canary: h.canary, + } +} + +// interpolateServices returns an interpolated copy of services and checks with +// values from the task's environment. +func interpolateServices(taskEnv *env.TaskEnv, services []*structs.Service) []*structs.Service { + interpolated := make([]*structs.Service, len(services)) + + for i, origService := range services { + // Create a copy as we need to reinterpolate every time the + // environment changes + service := origService.Copy() + + for _, check := range service.Checks { + check.Name = taskEnv.ReplaceEnv(check.Name) + check.Type = taskEnv.ReplaceEnv(check.Type) + check.Command = taskEnv.ReplaceEnv(check.Command) + check.Args = taskEnv.ParseAndReplace(check.Args) + check.Path = taskEnv.ReplaceEnv(check.Path) + check.Protocol = taskEnv.ReplaceEnv(check.Protocol) + check.PortLabel = taskEnv.ReplaceEnv(check.PortLabel) + check.InitialStatus = taskEnv.ReplaceEnv(check.InitialStatus) + check.Method = taskEnv.ReplaceEnv(check.Method) + check.GRPCService = taskEnv.ReplaceEnv(check.GRPCService) + if len(check.Header) > 0 { + header := make(map[string][]string, len(check.Header)) + for k, vs := range check.Header { + newVals := make([]string, len(vs)) + for i, v := range vs { + newVals[i] = taskEnv.ReplaceEnv(v) + } + header[taskEnv.ReplaceEnv(k)] = newVals + } + check.Header = header + } + } + + service.Name = taskEnv.ReplaceEnv(service.Name) + service.PortLabel = taskEnv.ReplaceEnv(service.PortLabel) + service.Tags = taskEnv.ParseAndReplace(service.Tags) + service.CanaryTags = taskEnv.ParseAndReplace(service.CanaryTags) + interpolated[i] = service + } + + return interpolated +} diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index c22304995..7cafde9b8 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -16,9 +16,11 @@ import ( "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" clientstate "github.com/hashicorp/nomad/client/state" + cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" "github.com/ugorji/go/codec" @@ -100,8 +102,8 @@ type TaskRunner struct { // driver is the driver for the task. driver driver.Driver - // handle is the handle to the currently running driver - handle driver.DriverHandle + handle driver.DriverHandle // the handle to the running driver + driverNet *cstructs.DriverNetwork // driver network if one exists handleLock sync.Mutex // task is the task being run @@ -121,6 +123,10 @@ type TaskRunner struct { // transistions. runnerHooks []interfaces.TaskHook + // consulClient is the client used by the consul service hook for + // registering services and checks + consulClient consul.ConsulServiceAPI + // vaultClient is the client to use to derive and renew Vault tokens vaultClient vaultclient.VaultClient @@ -137,6 +143,7 @@ type TaskRunner struct { type Config struct { Alloc *structs.Allocation ClientConfig *config.Config + Consul consul.ConsulServiceAPI Task *structs.Task TaskDir *allocdir.TaskDir Logger log.Logger @@ -174,6 +181,7 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { taskDir: config.TaskDir, taskName: config.Task.Name, envBuilder: envBuilder, + consulClient: config.Consul, vaultClient: config.VaultClient, //XXX Make a Copy to avoid races? state: config.Alloc.TaskStates[config.Task.Name], @@ -278,12 +286,12 @@ MAIN: } // Grab the handle - handle = tr.getDriverHandle() + handle, _ = tr.getDriverHandle() select { case waitRes := <-handle.WaitCh(): // Clear the handle - tr.setDriverHandle(nil) + tr.clearDriverHandle() // Store the wait result on the restart tracker tr.restartTracker.SetWaitResult(waitRes) @@ -291,7 +299,9 @@ MAIN: tr.logger.Debug("task killed") } - // TODO Need to run exited hooks + if err := tr.exited(); err != nil { + tr.logger.Error("exited hooks failed", "error", err) + } RESTART: // Actually restart by sleeping and also watching for destroy events @@ -365,10 +375,8 @@ func (tr *TaskRunner) runDriver() error { return err } - // Grab the handle - tr.setDriverHandle(sresp.Handle) - - //XXX need to capture the driver network + // Store the driver handle and associated metadata + tr.setDriverHandle(sresp.Handle, sresp.Network) // Emit an event that we started tr.SetState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) diff --git a/client/allocrunnerv2/taskrunner/task_runner_getters.go b/client/allocrunnerv2/taskrunner/task_runner_getters.go index d8bb78eba..8a0b6a9e7 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_getters.go +++ b/client/allocrunnerv2/taskrunner/task_runner_getters.go @@ -2,6 +2,7 @@ package taskrunner import ( "github.com/hashicorp/nomad/client/driver" + cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" ) @@ -35,14 +36,26 @@ func (tr *TaskRunner) setVaultToken(token string) { tr.vaultToken = token } -func (tr *TaskRunner) getDriverHandle() driver.DriverHandle { +// getDriverHandle returns the DriverHandle and associated driver metadata (at +// this point just the network) if it exists. +func (tr *TaskRunner) getDriverHandle() (driver.DriverHandle, *cstructs.DriverNetwork) { tr.handleLock.Lock() defer tr.handleLock.Unlock() - return tr.handle + return tr.handle, tr.driverNet } -func (tr *TaskRunner) setDriverHandle(handle driver.DriverHandle) { +func (tr *TaskRunner) setDriverHandle(handle driver.DriverHandle, net *cstructs.DriverNetwork) { tr.handleLock.Lock() defer tr.handleLock.Unlock() tr.handle = handle + tr.driverNet = net +} + +// clearDriverHandle clears the driver handle and associated driver metadata +// (driver network). +func (tr *TaskRunner) clearDriverHandle() { + tr.handleLock.Lock() + defer tr.handleLock.Unlock() + tr.handle = nil + tr.driverNet = nil } diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 35ebfd474..802831e9d 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -5,6 +5,7 @@ import ( "fmt" "time" + multierror "github.com/hashicorp/go-multierror" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" "github.com/hashicorp/nomad/nomad/structs" @@ -49,6 +50,17 @@ func (tr *TaskRunner) initHooks() { envBuilder: tr.envBuilder, })) } + + // If there are any services, add the hook + if len(task.Services) != 0 { + tr.runnerHooks = append(tr.runnerHooks, newServiceHook(serviceHookConfig{ + alloc: tr.Alloc(), + task: tr.Task(), + consul: tr.consulClient, + restarter: tr, + logger: hookLogger, + })) + } } // prestart is used to run the runners prestart hooks. @@ -159,6 +171,9 @@ func (tr *TaskRunner) poststart() error { }() } + handle, net := tr.getDriverHandle() + + var merr multierror.Error for _, hook := range tr.runnerHooks { post, ok := hook.(interfaces.TaskPoststartHook) if !ok { @@ -172,11 +187,14 @@ func (tr *TaskRunner) poststart() error { tr.logger.Trace("running poststart hook", "name", name, "start", start) } - req := interfaces.TaskPoststartRequest{} + req := interfaces.TaskPoststartRequest{ + DriverExec: handle, + DriverNetwork: net, + TaskEnv: tr.envBuilder.Build(), + } var resp interfaces.TaskPoststartResponse - // XXX We shouldn't exit on the first one if err := post.Poststart(tr.ctx, &req, &resp); err != nil { - return fmt.Errorf("poststart hook %q failed: %v", name, err) + merr.Errors = append(merr.Errors, fmt.Errorf("poststart hook %q failed: %v", name, err)) } if tr.logger.IsTrace() { @@ -185,7 +203,48 @@ func (tr *TaskRunner) poststart() error { } } - return nil + return merr.ErrorOrNil() +} + +// exited is used to run the exited hooks before a task is stopped. +func (tr *TaskRunner) exited() error { + if tr.logger.IsTrace() { + start := time.Now() + tr.logger.Trace("running exited hooks", "start", start) + defer func() { + end := time.Now() + tr.logger.Trace("finished exited hooks", "end", end, "duration", end.Sub(start)) + }() + } + + var merr multierror.Error + for _, hook := range tr.runnerHooks { + post, ok := hook.(interfaces.TaskExitedHook) + if !ok { + continue + } + + name := post.Name() + var start time.Time + if tr.logger.IsTrace() { + start = time.Now() + tr.logger.Trace("running exited hook", "name", name, "start", start) + } + + req := interfaces.TaskExitedRequest{} + var resp interfaces.TaskExitedResponse + if err := post.Exited(tr.ctx, &req, &resp); err != nil { + merr.Errors = append(merr.Errors, fmt.Errorf("exited hook %q failed: %v", name, err)) + } + + if tr.logger.IsTrace() { + end := time.Now() + tr.logger.Trace("finished exited hooks", "name", name, "end", end, "duration", end.Sub(start)) + } + } + + return merr.ErrorOrNil() + } // stop is used to run the stop hooks. @@ -199,6 +258,7 @@ func (tr *TaskRunner) stop() error { }() } + var merr multierror.Error for _, hook := range tr.runnerHooks { post, ok := hook.(interfaces.TaskStopHook) if !ok { @@ -214,9 +274,8 @@ func (tr *TaskRunner) stop() error { req := interfaces.TaskStopRequest{} var resp interfaces.TaskStopResponse - // XXX We shouldn't exit on the first one if err := post.Stop(tr.ctx, &req, &resp); err != nil { - return fmt.Errorf("stop hook %q failed: %v", name, err) + merr.Errors = append(merr.Errors, fmt.Errorf("stop hook %q failed: %v", name, err)) } if tr.logger.IsTrace() { @@ -225,7 +284,7 @@ func (tr *TaskRunner) stop() error { } } - return nil + return merr.ErrorOrNil() } // update is used to run the runners update hooks. @@ -239,6 +298,10 @@ func (tr *TaskRunner) updateHooks() { }() } + // Prepare state needed by Update hooks + alloc := tr.Alloc() + + // Execute Update hooks for _, hook := range tr.runnerHooks { upd, ok := hook.(interfaces.TaskUpdateHook) if !ok { @@ -251,6 +314,8 @@ func (tr *TaskRunner) updateHooks() { // Build the request req := interfaces.TaskUpdateRequest{ VaultToken: tr.getVaultToken(), + Alloc: alloc, + TaskEnv: tr.envBuilder.Build(), } // Time the update hook diff --git a/client/client.go b/client/client.go index 898b67cd8..7d1d7537e 100644 --- a/client/client.go +++ b/client/client.go @@ -766,6 +766,7 @@ func (c *Client) restoreState() error { ClientConfig: c.config, StateDB: c.stateDB, StateUpdater: c, + Consul: c.consulService, Vault: c.vaultClient, } c.configLock.RUnlock() @@ -1946,6 +1947,7 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error Logger: logger, ClientConfig: c.config, StateDB: c.stateDB, + Consul: c.consulService, Vault: c.vaultClient, StateUpdater: c, } diff --git a/command/agent/consul/check_watcher.go b/command/agent/consul/check_watcher.go index d1f93be4e..b6e8610fc 100644 --- a/command/agent/consul/check_watcher.go +++ b/command/agent/consul/check_watcher.go @@ -24,7 +24,7 @@ type ChecksAPI interface { // TaskRestarter allows the checkWatcher to restart tasks. type TaskRestarter interface { - Restart(source, reason string, failure bool) + Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error } // checkRestart handles restarting a task if a check is unhealthy. @@ -60,7 +60,7 @@ type checkRestart struct { // // Returns true if a restart was triggered in which case this check should be // removed (checks are added on task startup). -func (c *checkRestart) apply(now time.Time, status string) bool { +func (c *checkRestart) apply(ctx context.Context, now time.Time, status string) bool { healthy := func() { if !c.unhealthyState.IsZero() { c.logger.Debug("canceling restart because check became healthy") @@ -104,7 +104,13 @@ func (c *checkRestart) apply(now time.Time, status string) bool { // Tell TaskRunner to restart due to failure const failure = true - c.task.Restart("healthcheck", fmt.Sprintf("check %q unhealthy", c.checkName), failure) + reason := fmt.Sprintf("healthcheck: check %q unhealthy", c.checkName) + event := structs.NewTaskEvent(structs.TaskRestartSignal).SetRestartReason(reason) + err := c.task.Restart(ctx, event, failure) + if err != nil { + // Error restarting + return false + } return true } @@ -228,6 +234,11 @@ func (w *checkWatcher) Run(ctx context.Context) { // Loop over watched checks and update their status from results for cid, check := range checks { + // Shortcircuit if told to exit + if ctx.Err() != nil { + return + } + if _, ok := restartedTasks[check.taskKey]; ok { // Check for this task already restarted; remove and skip check delete(checks, cid) @@ -243,7 +254,7 @@ func (w *checkWatcher) Run(ctx context.Context) { continue } - restarted := check.apply(now, result.Status) + restarted := check.apply(ctx, now, result.Status) if restarted { // Checks are registered+watched on // startup, so it's safe to remove them diff --git a/nomad/structs/structs.go b/nomad/structs/structs.go index ec0ec4b66..9b2b534a1 100644 --- a/nomad/structs/structs.go +++ b/nomad/structs/structs.go @@ -6964,7 +6964,8 @@ type Allocation struct { // COMPAT(0.11): Remove in 0.11 // Resources is the total set of resources allocated as part - // of this allocation of the task group. + // of this allocation of the task group. Dynamic ports will be set by + // the scheduler. Resources *Resources // COMPAT(0.11): Remove in 0.11 @@ -6974,7 +6975,8 @@ type Allocation struct { // COMPAT(0.11): Remove in 0.11 // TaskResources is the set of resources allocated to each - // task. These should sum to the total Resources. + // task. These should sum to the total Resources. Dynamic ports will be + // set by the scheduler. TaskResources map[string]*Resources // AllocatedResources is the total resources allocated for the task group. @@ -7375,6 +7377,21 @@ func (a *Allocation) ComparableResources() *ComparableResources { } } +// LookupTask by name from the Allocation. Returns nil if the Job is not set, the +// TaskGroup does not exist, or the task name cannot be found. +func (a *Allocation) LookupTask(name string) *Task { + if a.Job == nil { + return nil + } + + tg := a.Job.LookupTaskGroup(a.TaskGroup) + if tg == nil { + return nil + } + + return tg.LookupTask(name) +} + // Stub returns a list stub for the allocation func (a *Allocation) Stub() *AllocListStub { return &AllocListStub{ From 7132b67c1e42c3f33400b5880697936cbe45e295 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Sun, 29 Jul 2018 19:27:06 -0700 Subject: [PATCH 038/152] Add Network method to Handle interface Should probably be moved to an Inspect method in the Driver Plugin world --- client/driver/docker.go | 28 +++++++++++++++++++--------- client/driver/driver.go | 4 ++++ client/driver/exec.go | 4 ++++ client/driver/java.go | 4 ++++ client/driver/mock_driver.go | 5 +++++ client/driver/qemu.go | 4 ++++ client/driver/raw_exec.go | 4 ++++ client/driver/rkt.go | 5 +++++ 8 files changed, 49 insertions(+), 9 deletions(-) diff --git a/client/driver/docker.go b/client/driver/docker.go index 711d5c075..a35f3f648 100644 --- a/client/driver/docker.go +++ b/client/driver/docker.go @@ -500,6 +500,7 @@ type DockerHandle struct { waitCh chan *dstructs.WaitResult doneCh chan bool removeContainerOnExit bool + net *cstructs.DriverNetwork } func NewDockerDriver(ctx *DriverContext) Driver { @@ -908,6 +909,15 @@ func (d *DockerDriver) Start(ctx *ExecContext, task *structs.Task) (*StartRespon container.ID, container.State.String()) } + // Detect container address + ip, autoUse := d.detectIP(container) + + net := &cstructs.DriverNetwork{ + PortMap: d.driverConfig.PortMap, + IP: ip, + AutoAdvertise: autoUse, + } + // Return a driver handle maxKill := d.DriverContext.config.MaxKillTimeout h := &DockerHandle{ @@ -928,22 +938,17 @@ func (d *DockerDriver) Start(ctx *ExecContext, task *structs.Task) (*StartRespon doneCh: make(chan bool), waitCh: make(chan *dstructs.WaitResult, 1), removeContainerOnExit: d.config.ReadBoolDefault(dockerCleanupContainerConfigOption, dockerCleanupContainerConfigDefault), + net: net, } go h.collectStats() go h.run() - // Detect container address - ip, autoUse := d.detectIP(container) - // Create a response with the driver handle and container network metadata resp := &StartResponse{ - Handle: h, - Network: &cstructs.DriverNetwork{ - PortMap: d.driverConfig.PortMap, - IP: ip, - AutoAdvertise: autoUse, - }, + Handle: h, + Network: net, } + return resp, nil } @@ -1811,6 +1816,7 @@ func (d *DockerDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, er maxKillTimeout: pid.MaxKillTimeout, doneCh: make(chan bool), waitCh: make(chan *dstructs.WaitResult, 1), + net: nil, //FIXME Need to get driver network } go h.collectStats() go h.run() @@ -1843,6 +1849,10 @@ func (h *DockerHandle) WaitCh() chan *dstructs.WaitResult { return h.waitCh } +func (h *DockerHandle) Network() *cstructs.DriverNetwork { + return h.net +} + func (h *DockerHandle) Update(task *structs.Task) error { // Store the updated kill timeout. h.killTimeout = GetKillTimeout(task.KillTimeout, h.maxKillTimeout) diff --git a/client/driver/driver.go b/client/driver/driver.go index 042d4b3d7..a667efe85 100644 --- a/client/driver/driver.go +++ b/client/driver/driver.go @@ -320,6 +320,10 @@ type DriverHandle interface { // ScriptExecutor is an interface used to execute commands such as // health check scripts in the a DriverHandle's context. ScriptExecutor + + // Network returns the driver's network or nil if the driver did not + // create a network. + Network() *cstructs.DriverNetwork } // ScriptExecutor is an interface that supports Exec()ing commands in the diff --git a/client/driver/exec.go b/client/driver/exec.go index 41ab5237b..edab4143f 100644 --- a/client/driver/exec.go +++ b/client/driver/exec.go @@ -270,6 +270,10 @@ func (h *execHandle) Signal(s os.Signal) error { return h.executor.Signal(s) } +func (d *execHandle) Network() *cstructs.DriverNetwork { + return nil +} + func (h *execHandle) Kill() error { if err := h.executor.ShutDown(); err != nil { if h.pluginClient.Exited() { diff --git a/client/driver/java.go b/client/driver/java.go index 3c4b31958..e8cbe4813 100644 --- a/client/driver/java.go +++ b/client/driver/java.go @@ -408,6 +408,10 @@ func (h *javaHandle) Signal(s os.Signal) error { return h.executor.Signal(s) } +func (d *javaHandle) Network() *cstructs.DriverNetwork { + return nil +} + func (h *javaHandle) Kill() error { if err := h.executor.ShutDown(); err != nil { if h.pluginClient.Exited() { diff --git a/client/driver/mock_driver.go b/client/driver/mock_driver.go index ffa6b0977..f3b1f2ede 100644 --- a/client/driver/mock_driver.go +++ b/client/driver/mock_driver.go @@ -372,6 +372,11 @@ func (h *mockDriverHandle) Update(task *structs.Task) error { return nil } +// TODO Implement when we need it. +func (d *mockDriverHandle) Network() *cstructs.DriverNetwork { + return nil +} + // TODO Implement when we need it. func (h *mockDriverHandle) Signal(s os.Signal) error { return h.signalErr diff --git a/client/driver/qemu.go b/client/driver/qemu.go index 158628672..dcbc0d0ec 100644 --- a/client/driver/qemu.go +++ b/client/driver/qemu.go @@ -451,6 +451,10 @@ func (h *qemuHandle) Signal(s os.Signal) error { return fmt.Errorf("Qemu driver can't send signals") } +func (d *qemuHandle) Network() *cstructs.DriverNetwork { + return nil +} + func (h *qemuHandle) Kill() error { gracefulShutdownSent := false // Attempt a graceful shutdown only if it was configured in the job diff --git a/client/driver/raw_exec.go b/client/driver/raw_exec.go index 9f6033528..68aff51f8 100644 --- a/client/driver/raw_exec.go +++ b/client/driver/raw_exec.go @@ -299,6 +299,10 @@ func (h *rawExecHandle) Signal(s os.Signal) error { return h.executor.Signal(s) } +func (d *rawExecHandle) Network() *cstructs.DriverNetwork { + return nil +} + func (h *rawExecHandle) Kill() error { if err := h.executor.ShutDown(); err != nil { if h.pluginClient.Exited() { diff --git a/client/driver/rkt.go b/client/driver/rkt.go index f417fdcfc..1d297000f 100644 --- a/client/driver/rkt.go +++ b/client/driver/rkt.go @@ -815,6 +815,11 @@ func (h *rktHandle) Signal(s os.Signal) error { return fmt.Errorf("Rkt does not support signals") } +//FIXME implement +func (d *rktHandle) Network() *cstructs.DriverNetwork { + return nil +} + // Kill is used to terminate the task. We send an Interrupt // and then provide a 5 second grace period before doing a Kill. func (h *rktHandle) Kill() error { From 39b3f3a85ba0e2b69fdd3025721ae87d170c0fb4 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 30 Jul 2018 16:37:26 -0700 Subject: [PATCH 039/152] call handle.Network() instead of storing it --- client/allocrunner/taskrunner/task_runner.go | 1 + client/allocrunnerv2/taskrunner/lifecycle.go | 6 +++--- client/allocrunnerv2/taskrunner/task_runner.go | 8 +++----- .../allocrunnerv2/taskrunner/task_runner_getters.go | 13 +++---------- .../allocrunnerv2/taskrunner/task_runner_hooks.go | 3 ++- 5 files changed, 12 insertions(+), 19 deletions(-) diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index 9a4f6d360..fb14943b9 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -115,6 +115,7 @@ type TaskRunner struct { // updateCh is used to receive updated versions of the allocation updateCh chan *structs.Allocation + // handle is returned when Starting or Opening a driver handle driver.DriverHandle handleLock sync.Mutex diff --git a/client/allocrunnerv2/taskrunner/lifecycle.go b/client/allocrunnerv2/taskrunner/lifecycle.go index db4bc4d6e..9de87f528 100644 --- a/client/allocrunnerv2/taskrunner/lifecycle.go +++ b/client/allocrunnerv2/taskrunner/lifecycle.go @@ -9,7 +9,7 @@ import ( func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error { // Grab the handle - handle, _ := tr.getDriverHandle() + handle := tr.getDriverHandle() // Check it is running if handle == nil { @@ -41,7 +41,7 @@ func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, fai func (tr *TaskRunner) Signal(event *structs.TaskEvent, s os.Signal) error { // Grab the handle - handle, _ := tr.getDriverHandle() + handle := tr.getDriverHandle() // Check it is running if handle == nil { @@ -57,7 +57,7 @@ func (tr *TaskRunner) Signal(event *structs.TaskEvent, s os.Signal) error { func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error { // Grab the handle - handle, _ := tr.getDriverHandle() + handle := tr.getDriverHandle() // Check if the handle is running if handle == nil { diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 7cafde9b8..d9705711f 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -20,7 +20,6 @@ import ( "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" clientstate "github.com/hashicorp/nomad/client/state" - cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" "github.com/ugorji/go/codec" @@ -102,8 +101,7 @@ type TaskRunner struct { // driver is the driver for the task. driver driver.Driver - handle driver.DriverHandle // the handle to the running driver - driverNet *cstructs.DriverNetwork // driver network if one exists + handle driver.DriverHandle // the handle to the running driver handleLock sync.Mutex // task is the task being run @@ -286,7 +284,7 @@ MAIN: } // Grab the handle - handle, _ = tr.getDriverHandle() + handle = tr.getDriverHandle() select { case waitRes := <-handle.WaitCh(): @@ -376,7 +374,7 @@ func (tr *TaskRunner) runDriver() error { } // Store the driver handle and associated metadata - tr.setDriverHandle(sresp.Handle, sresp.Network) + tr.setDriverHandle(sresp.Handle) // Emit an event that we started tr.SetState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) diff --git a/client/allocrunnerv2/taskrunner/task_runner_getters.go b/client/allocrunnerv2/taskrunner/task_runner_getters.go index 8a0b6a9e7..013f55849 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_getters.go +++ b/client/allocrunnerv2/taskrunner/task_runner_getters.go @@ -2,7 +2,6 @@ package taskrunner import ( "github.com/hashicorp/nomad/client/driver" - cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" ) @@ -36,26 +35,20 @@ func (tr *TaskRunner) setVaultToken(token string) { tr.vaultToken = token } -// getDriverHandle returns the DriverHandle and associated driver metadata (at -// this point just the network) if it exists. -func (tr *TaskRunner) getDriverHandle() (driver.DriverHandle, *cstructs.DriverNetwork) { +func (tr *TaskRunner) getDriverHandle() driver.DriverHandle { tr.handleLock.Lock() defer tr.handleLock.Unlock() - return tr.handle, tr.driverNet + return tr.handle } -func (tr *TaskRunner) setDriverHandle(handle driver.DriverHandle, net *cstructs.DriverNetwork) { +func (tr *TaskRunner) setDriverHandle(handle driver.DriverHandle) { tr.handleLock.Lock() defer tr.handleLock.Unlock() tr.handle = handle - tr.driverNet = net } -// clearDriverHandle clears the driver handle and associated driver metadata -// (driver network). func (tr *TaskRunner) clearDriverHandle() { tr.handleLock.Lock() defer tr.handleLock.Unlock() tr.handle = nil - tr.driverNet = nil } diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 802831e9d..ac44c1451 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -171,7 +171,8 @@ func (tr *TaskRunner) poststart() error { }() } - handle, net := tr.getDriverHandle() + handle := tr.getDriverHandle() + net := handle.Network() var merr multierror.Error for _, hook := range tr.runnerHooks { From a5d3e3fb0a17979cfb9c40c35524b7fc5a9cb9a7 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 1 Aug 2018 11:03:52 -0700 Subject: [PATCH 040/152] Implement alloc updates in arv2 Updates are applied asynchronously but sequentially --- client/allocrunnerv2/alloc_runner.go | 45 +++++++++--- .../allocrunnerv2/taskrunner/task_runner.go | 69 +++++++++++++++---- .../taskrunner/task_runner_getters.go | 14 ++++ .../taskrunner/task_runner_hooks.go | 4 +- client/allocrunnerv2/taskrunner/vault_hook.go | 9 +-- client/client.go | 11 ++- 6 files changed, 114 insertions(+), 38 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index a9cfbea1f..4d48a29d3 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -24,6 +24,14 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) +const ( + // updateChCap is the capacity of AllocRunner's updateCh. It must be 1 + // as we only want to process the latest update, so if there's already + // a pending update it will be removed from the chan before adding the + // newer update. + updateChCap = 1 +) + // allocRunner is used to run all the tasks in a given allocation type allocRunner struct { // id is the ID of the allocation. Can be accessed without a lock @@ -70,7 +78,9 @@ type allocRunner struct { tasks map[string]*taskrunner.TaskRunner tasksLock sync.RWMutex - // updateCh receives allocation updates via the Update method + // updateCh receives allocation updates via the Update method. Must + // have buffer size 1 in order to support dropping pending updates when + // a newer allocation is received. updateCh chan *structs.Allocation } @@ -90,7 +100,7 @@ func NewAllocRunner(config *Config) (*allocRunner, error) { vaultClient: config.Vault, tasks: make(map[string]*taskrunner.TaskRunner, len(tg.Tasks)), waitCh: make(chan struct{}), - updateCh: make(chan *structs.Allocation), + updateCh: make(chan *structs.Allocation, updateChCap), state: &state.State{}, stateDB: config.StateDB, stateUpdater: config.StateUpdater, @@ -169,9 +179,12 @@ MAIN: // TaskRunners have all exited break MAIN case updated := <-ar.updateCh: - // Updated alloc received - //XXX Update hooks - //XXX Update ar.alloc + // Update ar.alloc + ar.setAlloc(updated) + + //TODO Run AR Update hooks + + // Update task runners for _, tr := range ar.tasks { tr.Update(updated) } @@ -206,13 +219,18 @@ func (ar *allocRunner) runImpl() <-chan struct{} { } // Alloc returns the current allocation being run by this runner. -//XXX how do we handle mutate the state saving stuff func (ar *allocRunner) Alloc() *structs.Allocation { ar.allocLock.RLock() defer ar.allocLock.RUnlock() return ar.alloc } +func (ar *allocRunner) setAlloc(updated *structs.Allocation) { + ar.allocLock.Lock() + ar.alloc = updated + ar.allocLock.Unlock() +} + // SaveState does all the state related stuff. Who knows. FIXME //XXX do we need to do periodic syncing? if Saving is only called *before* Run // *and* within Run -- *and* Updates are applid within Run -- we may be able to @@ -389,10 +407,19 @@ func getClientStatus(taskStates map[string]*structs.TaskState) (status, descript // Update the running allocation with a new version received from the server. // -// This method is safe for calling concurrently with Run() and does not modify -// the passed in allocation. +// This method sends the updated alloc to Run for serially processing updates. +// If there is already a pending update it will be discarded and replaced by +// the latest update. func (ar *allocRunner) Update(update *structs.Allocation) { - ar.updateCh <- update + select { + case ar.updateCh <- update: + // Updated alloc sent + case <-ar.updateCh: + // There was a pending update; replace it with the new update. + // This also prevents Update() from blocking if its called + // concurrently with Run() exiting. + ar.updateCh <- update + } } // Destroy the alloc runner by stopping it if it is still running and cleaning diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index d9705711f..d73c46524 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -38,6 +38,12 @@ const ( // killFailureLimit is how many times we will attempt to kill a task before // giving up and potentially leaking resources. killFailureLimit = 5 + + // triggerUpdatechCap is the capacity for the triggerUpdateCh used for + // triggering updates. It should be exactly 1 as even if multiple + // updates have come in since the last one was handled, we only need to + // handle the last one. + triggerUpdateChCap = 1 ) var ( @@ -91,8 +97,10 @@ type TaskRunner struct { // Logger is the logger for the task runner. logger log.Logger - // updateCh receives Alloc updates - updateCh chan *structs.Allocation + // triggerUpdateCh is ticked whenever update hooks need to be run and + // must be created with cap=1 to signal a pending update and prevent + // callers from deadlocking if the receiver has exited. + triggerUpdateCh chan struct{} // waitCh is closed when the task runner has transitioned to a terminal // state @@ -182,14 +190,14 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { consulClient: config.Consul, vaultClient: config.VaultClient, //XXX Make a Copy to avoid races? - state: config.Alloc.TaskStates[config.Task.Name], - localState: config.LocalState, - stateDB: config.StateDB, - stateUpdater: config.StateUpdater, - ctx: trCtx, - ctxCancel: trCancel, - updateCh: make(chan *structs.Allocation), - waitCh: make(chan struct{}), + state: config.Alloc.TaskStates[config.Task.Name], + localState: config.LocalState, + stateDB: config.StateDB, + stateUpdater: config.StateUpdater, + ctx: trCtx, + ctxCancel: trCancel, + triggerUpdateCh: make(chan struct{}, triggerUpdateChCap), + waitCh: make(chan struct{}), } // Create the logger based on the allocation ID @@ -258,6 +266,11 @@ func (tr *TaskRunner) Run() { defer close(tr.waitCh) var handle driver.DriverHandle + // Updates are handled asynchronously with the other hooks but each + // triggered update - whether due to alloc updates or a new vault token + // - should be handled serially. + go tr.handleUpdates() + MAIN: for tr.ctx.Err() == nil { // Run the prestart hooks @@ -329,6 +342,20 @@ MAIN: tr.logger.Debug("task run loop exiting") } +// handleUpdates runs update hooks when triggerUpdateCh is ticked and exits +// when Run has returned. Should only be run in a goroutine from Run. +func (tr *TaskRunner) handleUpdates() { + for { + select { + case <-tr.triggerUpdateCh: + // Update triggered; run hooks + tr.updateHooks() + case <-tr.waitCh: + return + } + } +} + func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { // Determine if we should restart state, when := tr.restartTracker.GetState() @@ -665,16 +692,28 @@ func (tr *TaskRunner) WaitCh() <-chan struct{} { } // Update the running allocation with a new version received from the server. +// Calls Update hooks asynchronously with Run(). // // This method is safe for calling concurrently with Run() and does not modify // the passed in allocation. func (tr *TaskRunner) Update(update *structs.Allocation) { + // Update tr.alloc + tr.setAlloc(update) + + // Trigger update hooks + tr.triggerUpdateHooks() +} + +// triggerUpdate if there isn't already an update pending. Should be called +// instead of calling updateHooks directly to serialize runs of update hooks. +// TaskRunner state should be updated prior to triggering update hooks. +// +// Does not block. +func (tr *TaskRunner) triggerUpdateHooks() { select { - case tr.updateCh <- update: - case <-tr.WaitCh(): - //XXX Do we log here like we used to? If we're just - //shutting down it's not an error to drop the update as - //it will be applied on startup + case tr.triggerUpdateCh <- struct{}{}: + default: + // already an update hook pending } } diff --git a/client/allocrunnerv2/taskrunner/task_runner_getters.go b/client/allocrunnerv2/taskrunner/task_runner_getters.go index 013f55849..8a40c65bc 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_getters.go +++ b/client/allocrunnerv2/taskrunner/task_runner_getters.go @@ -11,6 +11,12 @@ func (tr *TaskRunner) Alloc() *structs.Allocation { return tr.alloc } +func (tr *TaskRunner) setAlloc(updated *structs.Allocation) { + tr.allocLock.Lock() + tr.alloc = updated + tr.allocLock.Unlock() +} + func (tr *TaskRunner) Task() *structs.Task { tr.taskLock.RLock() defer tr.taskLock.RUnlock() @@ -29,10 +35,18 @@ func (tr *TaskRunner) getVaultToken() string { return tr.vaultToken } +// setVaultToken updates the vault token on the task runner as well as in the +// task's environment. These two places must be set atomically to avoid a task +// seeing a different token on the task runner and in its environment. func (tr *TaskRunner) setVaultToken(token string) { tr.vaultTokenLock.Lock() defer tr.vaultTokenLock.Unlock() + + // Update the Vault token on the runner tr.vaultToken = token + + // Update the task's environment + tr.envBuilder.SetVaultToken(token, tr.task.Vault.Env) } func (tr *TaskRunner) getDriverHandle() driver.DriverHandle { diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index ac44c1451..4ff6bb0a4 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -288,7 +288,9 @@ func (tr *TaskRunner) stop() error { return merr.ErrorOrNil() } -// update is used to run the runners update hooks. +// update is used to run the runners update hooks. Should only be called from +// Run(). To trigger an update, update state on the TaskRunner and call +// triggerUpdateHooks. func (tr *TaskRunner) updateHooks() { if tr.logger.IsTrace() { start := time.Now() diff --git a/client/allocrunnerv2/taskrunner/vault_hook.go b/client/allocrunnerv2/taskrunner/vault_hook.go index 7b52e8b10..479479e71 100644 --- a/client/allocrunnerv2/taskrunner/vault_hook.go +++ b/client/allocrunnerv2/taskrunner/vault_hook.go @@ -38,14 +38,11 @@ type vaultTokenUpdateHandler interface { } func (tr *TaskRunner) updatedVaultToken(token string) { - // Update the Vault token on the runner + // Update the task runner and environment tr.setVaultToken(token) - // Update the tasks environment - tr.envBuilder.SetVaultToken(token, tr.task.Vault.Env) - - // Update the hooks with the new Vault token - tr.updateHooks() + // Trigger update hooks with the new Vault token + tr.triggerUpdateHooks() } type vaultHookConfig struct { diff --git a/client/client.go b/client/client.go index 7d1d7537e..82e734d14 100644 --- a/client/client.go +++ b/client/client.go @@ -1845,10 +1845,8 @@ func (c *Client) runAllocs(update *allocUpdates) { // Update the existing allocations for _, update := range diff.updated { - if err := c.updateAlloc(update); err != nil { - c.logger.Printf("[ERR] client: failed to update alloc %q: %v", - update.ID, err) - } + c.logger.Printf("[TRACE] client: updating alloc %q to index %d", update.ID, update.AllocModifyIndex) + c.updateAlloc(update) } // Make room for new allocations before running @@ -1893,17 +1891,16 @@ func (c *Client) removeAlloc(allocID string) { } // updateAlloc is invoked when we should update an allocation -func (c *Client) updateAlloc(update *structs.Allocation) error { +func (c *Client) updateAlloc(update *structs.Allocation) { c.allocLock.Lock() defer c.allocLock.Unlock() ar, ok := c.allocs[update.ID] if !ok { c.logger.Printf("[WARN] client: missing context for alloc %q", update.ID) - return nil + return } ar.Update(update) - return nil } // addAlloc is invoked when we should add an allocation From 5383d20505ae6508d9633336bf36cf4c74f17a5f Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 7 Aug 2018 17:37:39 -0700 Subject: [PATCH 041/152] removing old restoration path before api change --- client/allocrunner/alloc_runner.go | 206 ++++++++++--------- client/allocrunner/taskrunner/task_runner.go | 59 +++--- 2 files changed, 136 insertions(+), 129 deletions(-) diff --git a/client/allocrunner/alloc_runner.go b/client/allocrunner/alloc_runner.go index 590685751..9bf22149e 100644 --- a/client/allocrunner/alloc_runner.go +++ b/client/allocrunner/alloc_runner.go @@ -199,45 +199,46 @@ func (r *AllocRunner) pre060StateFilePath() string { // RestoreState is used to restore the state of the alloc runner func (r *AllocRunner) RestoreState() error { - err := r.stateDB.View(func(tx *bolt.Tx) error { - bkt, err := state.GetAllocationBucket(tx, r.allocID) - if err != nil { - return fmt.Errorf("failed to get allocation bucket: %v", err) - } + //XXX Deprecated: see allocrunnerv2 + //err := r.stateDB.View(func(tx *bolt.Tx) error { + // bkt, err := state.GetAllocationBucket(tx, r.allocID) + // if err != nil { + // return fmt.Errorf("failed to get allocation bucket: %v", err) + // } - // Get the state objects - var mutable allocRunnerMutableState - var immutable allocRunnerImmutableState - var allocState allocRunnerAllocState - var allocDir allocdir.AllocDir + // // Get the state objects + // var mutable allocRunnerMutableState + // var immutable allocRunnerImmutableState + // var allocState allocRunnerAllocState + // var allocDir allocdir.AllocDir - if err := state.GetObject(bkt, allocRunnerStateAllocKey, &allocState); err != nil { - return fmt.Errorf("failed to read alloc runner alloc state: %v", err) - } - if err := state.GetObject(bkt, allocRunnerStateImmutableKey, &immutable); err != nil { - return fmt.Errorf("failed to read alloc runner immutable state: %v", err) - } - if err := state.GetObject(bkt, allocRunnerStateMutableKey, &mutable); err != nil { - return fmt.Errorf("failed to read alloc runner mutable state: %v", err) - } - if err := state.GetObject(bkt, allocRunnerStateAllocDirKey, &allocDir); err != nil { - return fmt.Errorf("failed to read alloc runner alloc_dir state: %v", err) - } + // if err := state.GetObject(bkt, allocRunnerStateAllocKey, &allocState); err != nil { + // return fmt.Errorf("failed to read alloc runner alloc state: %v", err) + // } + // if err := state.GetObject(bkt, allocRunnerStateImmutableKey, &immutable); err != nil { + // return fmt.Errorf("failed to read alloc runner immutable state: %v", err) + // } + // if err := state.GetObject(bkt, allocRunnerStateMutableKey, &mutable); err != nil { + // return fmt.Errorf("failed to read alloc runner mutable state: %v", err) + // } + // if err := state.GetObject(bkt, allocRunnerStateAllocDirKey, &allocDir); err != nil { + // return fmt.Errorf("failed to read alloc runner alloc_dir state: %v", err) + // } - // Populate the fields - r.alloc = allocState.Alloc - r.allocDir = &allocDir - r.allocClientStatus = mutable.AllocClientStatus - r.allocClientDescription = mutable.AllocClientDescription - r.taskStates = mutable.TaskStates - r.alloc.ClientStatus = getClientStatus(r.taskStates) - r.alloc.DeploymentStatus = mutable.DeploymentStatus - return nil - }) + // // Populate the fields + // r.alloc = allocState.Alloc + // r.allocDir = &allocDir + // r.allocClientStatus = mutable.AllocClientStatus + // r.allocClientDescription = mutable.AllocClientDescription + // r.taskStates = mutable.TaskStates + // r.alloc.ClientStatus = getClientStatus(r.taskStates) + // r.alloc.DeploymentStatus = mutable.DeploymentStatus + // return nil + //}) - if err != nil { - return fmt.Errorf("failed to read allocation state: %v", err) - } + //if err != nil { + // return fmt.Errorf("failed to read allocation state: %v", err) + //} var snapshotErrors multierror.Error if r.alloc == nil { @@ -343,87 +344,90 @@ func (r *AllocRunner) saveAllocRunnerState() error { return nil } - // Grab all the relevant data - alloc := r.Alloc() + //XXX Deprecated: see allocrunnerv2 + return nil - r.allocLock.Lock() - allocClientStatus := r.allocClientStatus - allocClientDescription := r.allocClientDescription - r.allocLock.Unlock() + //// Grab all the relevant data + //alloc := r.Alloc() - r.allocDirLock.Lock() - allocDir := r.allocDir.Copy() - r.allocDirLock.Unlock() + //r.allocLock.Lock() + //allocClientStatus := r.allocClientStatus + //allocClientDescription := r.allocClientDescription + //r.allocLock.Unlock() - // Start the transaction. - return r.stateDB.Batch(func(tx *bolt.Tx) error { + //r.allocDirLock.Lock() + //allocDir := r.allocDir.Copy() + //r.allocDirLock.Unlock() - // Grab the allocation bucket - allocBkt, err := state.GetAllocationBucket(tx, r.allocID) - if err != nil { - return fmt.Errorf("failed to retrieve allocation bucket: %v", err) - } + //// Start the transaction. + //return r.stateDB.Batch(func(tx *bolt.Tx) error { - // Write the allocation if the eval has changed - r.persistedEvalLock.Lock() - lastPersisted := r.persistedEval - r.persistedEvalLock.Unlock() - if alloc.EvalID != lastPersisted { - allocState := &allocRunnerAllocState{ - Alloc: alloc, - } + // // Grab the allocation bucket + // allocBkt, err := state.GetAllocationBucket(tx, r.allocID) + // if err != nil { + // return fmt.Errorf("failed to retrieve allocation bucket: %v", err) + // } - if err := state.PutObject(allocBkt, allocRunnerStateAllocKey, &allocState); err != nil { - return fmt.Errorf("failed to write alloc_runner alloc state: %v", err) - } + // // Write the allocation if the eval has changed + // r.persistedEvalLock.Lock() + // lastPersisted := r.persistedEval + // r.persistedEvalLock.Unlock() + // if alloc.EvalID != lastPersisted { + // allocState := &allocRunnerAllocState{ + // Alloc: alloc, + // } - tx.OnCommit(func() { - r.persistedEvalLock.Lock() - r.persistedEval = alloc.EvalID - r.persistedEvalLock.Unlock() - }) - } + // if err := state.PutObject(allocBkt, allocRunnerStateAllocKey, &allocState); err != nil { + // return fmt.Errorf("failed to write alloc_runner alloc state: %v", err) + // } - // Write immutable data iff it hasn't been written yet - if !r.immutablePersisted { - immutable := &allocRunnerImmutableState{ - Version: r.config.Version.VersionNumber(), - } + // tx.OnCommit(func() { + // r.persistedEvalLock.Lock() + // r.persistedEval = alloc.EvalID + // r.persistedEvalLock.Unlock() + // }) + // } - if err := state.PutObject(allocBkt, allocRunnerStateImmutableKey, &immutable); err != nil { - return fmt.Errorf("failed to write alloc_runner immutable state: %v", err) - } + // // Write immutable data iff it hasn't been written yet + // if !r.immutablePersisted { + // immutable := &allocRunnerImmutableState{ + // Version: r.config.Version.VersionNumber(), + // } - tx.OnCommit(func() { - r.immutablePersisted = true - }) - } + // if err := state.PutObject(allocBkt, allocRunnerStateImmutableKey, &immutable); err != nil { + // return fmt.Errorf("failed to write alloc_runner immutable state: %v", err) + // } - // Write the alloc dir data if it hasn't been written before and it exists. - if !r.allocDirPersisted && allocDir != nil { - if err := state.PutObject(allocBkt, allocRunnerStateAllocDirKey, allocDir); err != nil { - return fmt.Errorf("failed to write alloc_runner allocDir state: %v", err) - } + // tx.OnCommit(func() { + // r.immutablePersisted = true + // }) + // } - tx.OnCommit(func() { - r.allocDirPersisted = true - }) - } + // // Write the alloc dir data if it hasn't been written before and it exists. + // if !r.allocDirPersisted && allocDir != nil { + // if err := state.PutObject(allocBkt, allocRunnerStateAllocDirKey, allocDir); err != nil { + // return fmt.Errorf("failed to write alloc_runner allocDir state: %v", err) + // } - // Write the mutable state every time - mutable := &allocRunnerMutableState{ - AllocClientStatus: allocClientStatus, - AllocClientDescription: allocClientDescription, - TaskStates: alloc.TaskStates, - DeploymentStatus: alloc.DeploymentStatus, - } + // tx.OnCommit(func() { + // r.allocDirPersisted = true + // }) + // } - if err := state.PutObject(allocBkt, allocRunnerStateMutableKey, &mutable); err != nil { - return fmt.Errorf("failed to write alloc_runner mutable state: %v", err) - } + // // Write the mutable state every time + // mutable := &allocRunnerMutableState{ + // AllocClientStatus: allocClientStatus, + // AllocClientDescription: allocClientDescription, + // TaskStates: alloc.TaskStates, + // DeploymentStatus: alloc.DeploymentStatus, + // } - return nil - }) + // if err := state.PutObject(allocBkt, allocRunnerStateMutableKey, &mutable); err != nil { + // return fmt.Errorf("failed to write alloc_runner mutable state: %v", err) + // } + + // return nil + //}) } // DestroyState is used to cleanup after ourselves diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index fb14943b9..60032e60d 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -353,20 +353,21 @@ func (r *TaskRunner) pre060StateFilePath() string { // executor. func (r *TaskRunner) RestoreState() (string, error) { var snap taskRunnerState - err := r.stateDB.View(func(tx *bolt.Tx) error { - bkt, err := state.GetTaskBucket(tx, r.alloc.ID, r.task.Name) - if err != nil { - return fmt.Errorf("failed to get task bucket: %v", err) - } + //XXX Deprecated: see allocrunnerv2 + //err := r.stateDB.View(func(tx *bolt.Tx) error { + // bkt, err := state.GetTaskBucket(tx, r.alloc.ID, r.task.Name) + // if err != nil { + // return fmt.Errorf("failed to get task bucket: %v", err) + // } - if err := state.GetObject(bkt, taskRunnerStateAllKey, &snap); err != nil { - return fmt.Errorf("failed to read task runner state: %v", err) - } - return nil - }) - if err != nil { - return "", err - } + // if err := state.GetObject(bkt, taskRunnerStateAllKey, &snap); err != nil { + // return fmt.Errorf("failed to read task runner state: %v", err) + // } + // return nil + //}) + //if err != nil { + // return "", err + //} // Restore fields from the snapshot r.artifactsDownloaded = snap.ArtifactDownloaded @@ -511,24 +512,26 @@ func (r *TaskRunner) SaveState() error { } // Start the transaction. - return r.stateDB.Batch(func(tx *bolt.Tx) error { - // Grab the task bucket - taskBkt, err := state.GetTaskBucket(tx, r.alloc.ID, r.task.Name) - if err != nil { - return fmt.Errorf("failed to retrieve allocation bucket: %v", err) - } + //XXX Deprecated: see allocrunnerv2 + return nil + //return r.stateDB.Batch(func(tx *bolt.Tx) error { + // // Grab the task bucket + // taskBkt, err := state.GetTaskBucket(tx, r.alloc.ID, r.task.Name) + // if err != nil { + // return fmt.Errorf("failed to retrieve allocation bucket: %v", err) + // } - if err := state.PutData(taskBkt, taskRunnerStateAllKey, buf.Bytes()); err != nil { - return fmt.Errorf("failed to write task_runner state: %v", err) - } + // if err := state.PutData(taskBkt, taskRunnerStateAllKey, buf.Bytes()); err != nil { + // return fmt.Errorf("failed to write task_runner state: %v", err) + // } - // Store the hash that was persisted - tx.OnCommit(func() { - r.persistedHash = h - }) + // // Store the hash that was persisted + // tx.OnCommit(func() { + // r.persistedHash = h + // }) - return nil - }) + // return nil + //}) } // DestroyState is used to cleanup after ourselves From fae5e89a0e3faa0eef7004a4e3c004d0b6019f81 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 7 Aug 2018 17:38:13 -0700 Subject: [PATCH 042/152] artifacts: don't emit event when there's no artifacts --- client/allocrunnerv2/taskrunner/artifact_hook.go | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/client/allocrunnerv2/taskrunner/artifact_hook.go b/client/allocrunnerv2/taskrunner/artifact_hook.go index ba3612f82..2253fdfe3 100644 --- a/client/allocrunnerv2/taskrunner/artifact_hook.go +++ b/client/allocrunnerv2/taskrunner/artifact_hook.go @@ -30,6 +30,11 @@ func (*artifactHook) Name() string { } func (h *artifactHook) Prestart(ctx context.Context, req *interfaces.TaskPrestartRequest, resp *interfaces.TaskPrestartResponse) error { + if len(req.Task.Artifacts) == 0 { + resp.Done = true + return nil + } + h.eventEmitter.EmitEvent(structs.NewTaskEvent(structs.TaskDownloadingArtifacts)) for _, artifact := range req.Task.Artifacts { From d890de036a188156dcacd7fe2c010254d1e9f26d Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 7 Aug 2018 17:38:40 -0700 Subject: [PATCH 043/152] tr: persist hook state whenever it changes --- .../taskrunner/task_runner_hooks.go | 27 +++++++++++-------- 1 file changed, 16 insertions(+), 11 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 4ff6bb0a4..bad2c84a3 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -124,22 +124,17 @@ func (tr *TaskRunner) prestart() error { // Store the hook state { - tr.localStateLock.Lock() - hookState, ok := tr.localState.Hooks[name] - if !ok { - hookState = &state.HookState{} - tr.localState.Hooks[name] = hookState + hookState := &state.HookState{ + Data: resp.HookData, + PrestartDone: resp.Done, } - if resp.HookData != nil { - hookState.Data = resp.HookData - hookState.PrestartDone = resp.Done - } - tr.localStateLock.Unlock() - // Store and persist local state if the hook state has changed if !hookState.Equal(origHookState) { + tr.localStateLock.Lock() tr.localState.Hooks[name] = hookState + tr.localStateLock.Unlock() + if err := tr.persistLocalState(); err != nil { return err } @@ -198,6 +193,8 @@ func (tr *TaskRunner) poststart() error { merr.Errors = append(merr.Errors, fmt.Errorf("poststart hook %q failed: %v", name, err)) } + // No need to persist as PoststartResponse is currently empty + if tr.logger.IsTrace() { end := time.Now() tr.logger.Trace("finished poststart hooks", "name", name, "end", end, "duration", end.Sub(start)) @@ -238,6 +235,8 @@ func (tr *TaskRunner) exited() error { merr.Errors = append(merr.Errors, fmt.Errorf("exited hook %q failed: %v", name, err)) } + // No need to persist as TaskExitedResponse is currently empty + if tr.logger.IsTrace() { end := time.Now() tr.logger.Trace("finished exited hooks", "name", name, "end", end, "duration", end.Sub(start)) @@ -279,6 +278,8 @@ func (tr *TaskRunner) stop() error { merr.Errors = append(merr.Errors, fmt.Errorf("stop hook %q failed: %v", name, err)) } + // No need to persist as TaskStopResponse is currently empty + if tr.logger.IsTrace() { end := time.Now() tr.logger.Trace("finished stop hooks", "name", name, "end", end, "duration", end.Sub(start)) @@ -334,6 +335,8 @@ func (tr *TaskRunner) updateHooks() { tr.logger.Error("update hook failed", "name", name, "error", err) } + // No need to persist as TaskUpdateResponse is currently empty + if tr.logger.IsTrace() { end := time.Now() tr.logger.Trace("finished update hooks", "name", name, "end", end, "duration", end.Sub(start)) @@ -375,6 +378,8 @@ func (tr *TaskRunner) kill() { tr.logger.Error("kill hook failed", "name", name, "error", err) } + // No need to persist as TaskKillResponse is currently empty + if tr.logger.IsTrace() { end := time.Now() tr.logger.Trace("finished kill hooks", "name", name, "end", end, "duration", end.Sub(start)) From 533bc93b3a7da64c0ccdd685a411f4b602400ab3 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 7 Aug 2018 17:46:37 -0700 Subject: [PATCH 044/152] implement all boltdb interactions behind StateDB --- client/allocrunnerv2/alloc_runner.go | 31 +- client/allocrunnerv2/config.go | 4 +- .../allocrunnerv2/taskrunner/task_runner.go | 104 ++--- client/client.go | 29 +- client/state/interface.go | 16 + client/state/noopdb.go | 32 ++ client/state/state_database.go | 410 ++++++++++++------ 7 files changed, 376 insertions(+), 250 deletions(-) create mode 100644 client/state/interface.go create mode 100644 client/state/noopdb.go diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 4d48a29d3..68e1f90ee 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -7,7 +7,6 @@ import ( "sync" "time" - "github.com/boltdb/bolt" log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner" @@ -17,7 +16,7 @@ import ( "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/consul" cinterfaces "github.com/hashicorp/nomad/client/interfaces" - clientstate "github.com/hashicorp/nomad/client/state" + cstate "github.com/hashicorp/nomad/client/state" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/helper" @@ -65,7 +64,7 @@ type allocRunner struct { state *state.State stateLock sync.RWMutex - stateDB *bolt.DB + stateDB cstate.StateDB // allocDir is used to build the allocations directory structure. allocDir *allocdir.AllocDir @@ -236,26 +235,24 @@ func (ar *allocRunner) setAlloc(updated *structs.Allocation) { // *and* within Run -- *and* Updates are applid within Run -- we may be able to // skip quite a bit of locking? maybe? func (ar *allocRunner) SaveState() error { - // XXX Do we move this to the client - return ar.stateDB.Update(func(tx *bolt.Tx) error { - //XXX Track AllocModifyIndex to only write alloc on change? - // Write the allocation - return clientstate.PutAllocation(tx, ar.Alloc()) - }) + //XXX Do we move this to the client + //XXX Track AllocModifyIndex to only write alloc on change? + + // Write the allocation + return ar.stateDB.PutAllocation(ar.Alloc()) } // Restore state from database. Must be called after NewAllocRunner but before // Run. func (ar *allocRunner) Restore() error { - return ar.stateDB.View(func(tx *bolt.Tx) error { - // Restore task runners - for _, tr := range ar.tasks { - if err := tr.Restore(tx); err != nil { - return err - } + // Restore task runners + for _, tr := range ar.tasks { + if err := tr.Restore(); err != nil { + return err } - return nil - }) + } + + return nil } // TaskStateUpdated is called by TaskRunner when a task's state has been diff --git a/client/allocrunnerv2/config.go b/client/allocrunnerv2/config.go index ebf955a29..6d5865299 100644 --- a/client/allocrunnerv2/config.go +++ b/client/allocrunnerv2/config.go @@ -1,11 +1,11 @@ package allocrunnerv2 import ( - "github.com/boltdb/bolt" log "github.com/hashicorp/go-hclog" clientconfig "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/interfaces" + cstate "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" ) @@ -22,7 +22,7 @@ type Config struct { Alloc *structs.Allocation // StateDB is used to store and restore state. - StateDB *bolt.DB + StateDB cstate.StateDB // Consul is the Consul client used to register task services and checks Consul consul.ConsulServiceAPI diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index d73c46524..45ecb158e 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -9,7 +9,6 @@ import ( "time" metrics "github.com/armon/go-metrics" - "github.com/boltdb/bolt" log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" @@ -19,7 +18,7 @@ import ( "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" - clientstate "github.com/hashicorp/nomad/client/state" + cstate "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" "github.com/ugorji/go/codec" @@ -46,16 +45,6 @@ const ( triggerUpdateChCap = 1 ) -var ( - // taskRunnerStateAllKey holds all the task runners state. At the moment - // there is no need to split it - //XXX refactor out of clientstate and new state - //XXX Old key - going to need to migrate - //taskRunnerStateAllKey = []byte("simple-all") - taskLocalStateKey = []byte("local_state") - taskStateKey = []byte("task_state") -) - type TaskRunner struct { // allocID and taskName are immutable so these fields may be accessed // without locks @@ -79,8 +68,8 @@ type TaskRunner struct { localState *state.LocalState localStateLock sync.RWMutex - // stateDB is for persisting localState - stateDB *bolt.DB + // stateDB is for persisting localState and taskState + stateDB cstate.StateDB // persistedHash is the hash of the last persisted state for skipping // unnecessary writes @@ -161,7 +150,7 @@ type Config struct { LocalState *state.LocalState // StateDB is used to store and restore state. - StateDB *bolt.DB + StateDB cstate.StateDB // StateUpdater is used to emit updated task state StateUpdater interfaces.TaskStateHandler @@ -464,8 +453,13 @@ func (tr *TaskRunner) handleDestroy(handle driver.DriverHandle) (destroyed bool, } // persistLocalState persists local state to disk synchronously. +// +//XXX Not safe for concurrent calls. Should it be? func (tr *TaskRunner) persistLocalState() error { - // buffer for writing to boltdb + tr.localStateLock.Lock() + defer tr.localStateLock.Unlock() + + // buffer for writing serialized state to var buf bytes.Buffer // Hash for skipping unnecessary writes @@ -479,10 +473,7 @@ func (tr *TaskRunner) persistLocalState() error { w := io.MultiWriter(h, &buf) // Encode as msgpack value - tr.localStateLock.Lock() - err = codec.NewEncoder(w, structs.MsgpackHandle).Encode(&tr.localState) - tr.localStateLock.Unlock() - if err != nil { + if err := codec.NewEncoder(w, structs.MsgpackHandle).Encode(&tr.localState); err != nil { return fmt.Errorf("failed to serialize snapshot: %v", err) } @@ -492,25 +483,14 @@ func (tr *TaskRunner) persistLocalState() error { return nil } - return tr.stateDB.Update(func(tx *bolt.Tx) error { - // Grab the task bucket - //XXX move into new state pkg - taskBkt, err := clientstate.GetTaskBucket(tx, tr.allocID, tr.taskName) - if err != nil { - return fmt.Errorf("failed to retrieve allocation bucket: %v", err) - } + if err := tr.stateDB.PutTaskRunnerLocalState(tr.allocID, tr.taskName, buf.Bytes()); err != nil { + return err + } - if err := clientstate.PutData(taskBkt, taskLocalStateKey, buf.Bytes()); err != nil { - return fmt.Errorf("failed to write task_runner state: %v", err) - } + // State was persisted, set the hash + tr.persistedHash = hashVal - // Store the hash that was persisted - tx.OnCommit(func() { - tr.persistedHash = hashVal - }) - - return nil - }) + return nil } // XXX If the objects don't exists since the client shutdown before the task @@ -519,31 +499,14 @@ func (tr *TaskRunner) persistLocalState() error { // // Restore task runner state. Called by AllocRunner.Restore after NewTaskRunner // but before Run so no locks need to be acquired. -func (tr *TaskRunner) Restore(tx *bolt.Tx) error { - bkt, err := clientstate.GetTaskBucket(tx, tr.allocID, tr.taskName) +func (tr *TaskRunner) Restore() error { + ls, ts, err := tr.stateDB.GetTaskRunnerState(tr.allocID, tr.taskName) if err != nil { - return fmt.Errorf("failed to get task %q bucket: %v", tr.taskName, err) + return err } - // Restore Local State - //XXX set persisted hash to avoid immediate write on first use? - var ls state.LocalState - if err := clientstate.GetObject(bkt, taskLocalStateKey, &ls); err != nil { - return fmt.Errorf("failed to read local task runner state: %v", err) - } - tr.localState = &ls - - // Restore Task State - var ts structs.TaskState - if err := clientstate.GetObject(bkt, taskStateKey, &ts); err != nil { - return fmt.Errorf("failed to read task state: %v", err) - } - tr.state = &ts - - // XXX if driver has task { - // tr.restoreDriver() - // } - + tr.localState = ls + tr.state = ts return nil } @@ -612,14 +575,7 @@ func (tr *TaskRunner) setStateLocal(state string, event *structs.TaskEvent) *str } // Persist the state and event - if err := tr.stateDB.Update(func(tx *bolt.Tx) error { - bkt, err := clientstate.GetTaskBucket(tx, tr.allocID, tr.taskName) - if err != nil { - return err - } - - return clientstate.PutObject(bkt, taskStateKey, tr.state) - }); err != nil { + if err := tr.stateDB.PutTaskState(tr.allocID, tr.taskName, taskState); err != nil { // Only a warning because the next event/state-transition will // try to persist it again. tr.logger.Error("error persisting task state", "error", err, "event", event, "state", state) @@ -638,18 +594,8 @@ func (tr *TaskRunner) EmitEvent(event *structs.TaskEvent) { tr.emitEventImpl(event) - // Events that do *not* change task state can be batched. - //XXX Seems like this clamps the maximum transaction latency to 10ms. - err := tr.stateDB.Batch(func(tx *bolt.Tx) error { - bkt, err := clientstate.GetTaskBucket(tx, tr.allocID, tr.taskName) - if err != nil { - return err - } - - return clientstate.PutObject(bkt, taskStateKey, tr.state) - }) - - if err != nil { + //XXX EmitEvents do not change TaskState.State so batch? + if err := tr.stateDB.PutTaskState(tr.allocID, tr.taskName, tr.state); err != nil { // Only a warning because the next event/state-transition will // try to persist it again. tr.logger.Warn("error persisting event", "error", err, "event", event) diff --git a/client/client.go b/client/client.go index 82e734d14..e52d5b917 100644 --- a/client/client.go +++ b/client/client.go @@ -32,7 +32,7 @@ import ( "github.com/hashicorp/nomad/client/allocrunnerv2" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/servers" - clientstate "github.com/hashicorp/nomad/client/state" + "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/client/stats" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/command/agent/consul" @@ -122,7 +122,7 @@ type Client struct { start time.Time // stateDB is used to efficiently store client state. - stateDB *bolt.DB + stateDB state.StateDB // configCopy is a copy that should be passed to alloc-runners. configCopy *config.Config @@ -380,10 +380,10 @@ func (c *Client) init() error { } c.logger.Printf("[INFO] client: using state directory %v", c.config.StateDir) - // Create or open the state database - db, err := bolt.Open(filepath.Join(c.config.StateDir, "state.db"), 0600, nil) + // Open the state database + db, err := state.NewStateDB(c.config.StateDir, c.config.DevMode) if err != nil { - return fmt.Errorf("failed to create state database: %v", err) + return fmt.Errorf("failed to open state database: %v", err) } c.stateDB = db @@ -736,19 +736,19 @@ func (c *Client) restoreState() error { // Restore using state database // Restore allocations - var allocs []*structs.Allocation - var err error - err = c.stateDB.View(func(tx *bolt.Tx) error { - allocs, err = clientstate.GetAllAllocations(tx) - if err != nil { - return fmt.Errorf("failed to list allocations: %v", err) - } - return nil - }) + allocs, allocErrs, err := c.stateDB.GetAllAllocations() if err != nil { return err } + for allocID, err := range allocErrs { + c.logger.Printf("[ERR] client: failed to restore alloc %q: %v", allocID, err) + //XXX Cleanup? + // Try to clean up alloc dir + // Remove boltdb entries? + // Send to server with clientstatus=failed + } + // Load each alloc back var mErr multierror.Error for _, alloc := range allocs { @@ -782,6 +782,7 @@ func (c *Client) restoreState() error { if err := ar.Restore(); err != nil { c.logger.Printf("[ERR] client: failed to restore alloc %q: %v", alloc.ID, err) mErr.Errors = append(mErr.Errors, err) + //TODO Cleanup allocrunner continue } diff --git a/client/state/interface.go b/client/state/interface.go new file mode 100644 index 000000000..e4d60f211 --- /dev/null +++ b/client/state/interface.go @@ -0,0 +1,16 @@ +package state + +import ( + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + "github.com/hashicorp/nomad/nomad/structs" +) + +// StateDB implementations store and load Nomad client state. +type StateDB interface { + GetAllAllocations() ([]*structs.Allocation, map[string]error, error) + PutAllocation(*structs.Allocation) error + GetTaskRunnerState(allocID, taskName string) (*state.LocalState, *structs.TaskState, error) + PutTaskRunnerLocalState(allocID, taskName string, buf []byte) error + PutTaskState(allocID, taskName string, state *structs.TaskState) error + Close() error +} diff --git a/client/state/noopdb.go b/client/state/noopdb.go new file mode 100644 index 000000000..67f5f1760 --- /dev/null +++ b/client/state/noopdb.go @@ -0,0 +1,32 @@ +package state + +import ( + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + "github.com/hashicorp/nomad/nomad/structs" +) + +type noopDB struct{} + +func (n noopDB) GetAllAllocations() ([]*structs.Allocation, map[string]error, error) { + return nil, nil, nil +} + +func (n noopDB) PutAllocation(*structs.Allocation) error { + return nil +} + +func (n noopDB) GetTaskRunnerState(allocID string, taskName string) (*state.LocalState, *structs.TaskState, error) { + return nil, nil, nil +} + +func (n noopDB) PutTaskRunnerLocalState(allocID string, taskName string, buf []byte) error { + return nil +} + +func (n noopDB) PutTaskState(allocID string, taskName string, state *structs.TaskState) error { + return nil +} + +func (n noopDB) Close() error { + return nil +} diff --git a/client/state/state_database.go b/client/state/state_database.go index 0a170b9a3..1828f62a2 100644 --- a/client/state/state_database.go +++ b/client/state/state_database.go @@ -3,8 +3,10 @@ package state import ( "bytes" "fmt" + "path/filepath" "github.com/boltdb/bolt" + trstate "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" "github.com/hashicorp/nomad/nomad/structs" "github.com/ugorji/go/codec" ) @@ -14,6 +16,7 @@ The client has a boltDB backed state store. The schema as of 0.6 looks as follow allocations/ (bucket) |--> / (bucket) + |--> alloc -> *structs.Allocation |--> alloc_runner persisted objects (k/v) |--> / (bucket) |--> task_runner persisted objects (k/v) @@ -23,119 +26,20 @@ var ( // allocationsBucket is the bucket name containing all allocation related // data allocationsBucket = []byte("allocations") + + // allocKey is the key serialized Allocations are stored under + allocKey = []byte("alloc") + + // taskRunnerStateAllKey holds all the task runners state. At the moment + // there is no need to split it + //XXX Old key - going to need to migrate + //taskRunnerStateAllKey = []byte("simple-all") + + // allocations -> $allocid -> $taskname -> the keys below + taskLocalStateKey = []byte("local_state") + taskStateKey = []byte("task_state") ) -func PutObject(bkt *bolt.Bucket, key []byte, obj interface{}) error { - if !bkt.Writable() { - return fmt.Errorf("bucket must be writable") - } - - // Serialize the object - var buf bytes.Buffer - if err := codec.NewEncoder(&buf, structs.MsgpackHandle).Encode(obj); err != nil { - return fmt.Errorf("failed to encode passed object: %v", err) - } - - if err := bkt.Put(key, buf.Bytes()); err != nil { - return fmt.Errorf("failed to write data at key %v: %v", string(key), err) - } - - return nil -} - -func PutData(bkt *bolt.Bucket, key, value []byte) error { - if !bkt.Writable() { - return fmt.Errorf("bucket must be writable") - } - - if err := bkt.Put(key, value); err != nil { - return fmt.Errorf("failed to write data at key %v: %v", string(key), err) - } - - return nil -} - -func GetObject(bkt *bolt.Bucket, key []byte, obj interface{}) error { - // Get the data - data := bkt.Get(key) - if data == nil { - return fmt.Errorf("no data at key %v", string(key)) - } - - // Deserialize the object - if err := codec.NewDecoderBytes(data, structs.MsgpackHandle).Decode(obj); err != nil { - return fmt.Errorf("failed to decode data into passed object: %v", err) - } - - return nil -} - -// GetAllocationBucket returns the bucket used to persist state about a -// particular allocation. If the root allocation bucket or the specific -// allocation bucket doesn't exist, it will be created as long as the -// transaction is writable. -func GetAllocationBucket(tx *bolt.Tx, allocID string) (*bolt.Bucket, error) { - var err error - w := tx.Writable() - - // Retrieve the root allocations bucket - allocations := tx.Bucket(allocationsBucket) - if allocations == nil { - if !w { - return nil, fmt.Errorf("Allocations bucket doesn't exist and transaction is not writable") - } - - allocations, err = tx.CreateBucket(allocationsBucket) - if err != nil { - return nil, err - } - } - - // Retrieve the specific allocations bucket - key := []byte(allocID) - alloc := allocations.Bucket(key) - if alloc == nil { - if !w { - return nil, fmt.Errorf("Allocation bucket doesn't exist and transaction is not writable") - } - - alloc, err = allocations.CreateBucket(key) - if err != nil { - return nil, err - } - } - - return alloc, nil -} - -// GetTaskBucket returns the bucket used to persist state about a -// particular task. If the root allocation bucket, the specific -// allocation or task bucket doesn't exist, they will be created as long as the -// transaction is writable. -func GetTaskBucket(tx *bolt.Tx, allocID, taskName string) (*bolt.Bucket, error) { - alloc, err := GetAllocationBucket(tx, allocID) - if err != nil { - return nil, err - } - - // Retrieve the specific task bucket - w := tx.Writable() - key := []byte(taskName) - task := alloc.Bucket(key) - if task == nil { - if !w { - return nil, fmt.Errorf("Task bucket doesn't exist and transaction is not writable") - } - - task, err = alloc.CreateBucket(key) - if err != nil { - return nil, err - } - } - - return task, nil -} - // DeleteAllocationBucket is used to delete an allocation bucket if it exists. func DeleteAllocationBucket(tx *bolt.Tx, allocID string) error { if !tx.Writable() { @@ -184,17 +88,13 @@ func DeleteTaskBucket(tx *bolt.Tx, allocID, taskName string) error { return alloc.DeleteBucket(key) } -//XXX duplicated in arv2?! -var ( - // The following are the key paths written to the state database - allocRunnerStateAllocKey = []byte("alloc") -) +var () -type allocRunnerAllocState struct { +type allocEntry struct { Alloc *structs.Allocation } -func GetAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, error) { +func getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map[string]error) { allocationsBkt := tx.Bucket(allocationsBucket) if allocationsBkt == nil { // No allocs @@ -202,6 +102,7 @@ func GetAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, error) { } var allocs []*structs.Allocation + errs := map[string]error{} // Create a cursor for iteration. c := allocationsBkt.Cursor() @@ -211,38 +112,271 @@ func GetAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, error) { allocID := string(k) allocBkt := allocationsBkt.Bucket(k) if allocBkt == nil { - //XXX merr? - return nil, fmt.Errorf("alloc %q missing", allocID) + errs[allocID] = fmt.Errorf("missing alloc bucket") + continue } - var allocState allocRunnerAllocState - if err := GetObject(allocBkt, allocRunnerStateAllocKey, &allocState); err != nil { - //XXX merr? - return nil, fmt.Errorf("failed to restore alloc %q: %v", allocID, err) + var allocState allocEntry + if err := getObject(allocBkt, allocKey, &allocState); err != nil { + errs[allocID] = fmt.Errorf("failed to decode alloc %v", err) + continue } + allocs = append(allocs, allocState.Alloc) } - return allocs, nil + return allocs, errs } -// PutAllocation stores an allocation given a writable transaction. -func PutAllocation(tx *bolt.Tx, alloc *structs.Allocation) error { - // Retrieve the root allocations bucket - allocsBkt, err := tx.CreateBucketIfNotExists(allocationsBucket) +// BoltStateDB persists and restores Nomad client state in a boltdb. All +// methods are safe for concurrent access. Create via NewStateDB by setting +// devMode=false. +type BoltStateDB struct { + db *bolt.DB +} + +func NewStateDB(stateDir string, devMode bool) (StateDB, error) { + // Return a noop state db implementation when in debug mode + if devMode { + return noopDB{}, nil + } + + // Create or open the boltdb state database + db, err := bolt.Open(filepath.Join(stateDir, "state.db"), 0600, nil) if err != nil { - return err + return nil, fmt.Errorf("failed to create state database: %v", err) + } + + sdb := &BoltStateDB{ + db: db, + } + return sdb, nil +} + +// GetAllAllocations gets all allocations persisted by this client and returns +// a map of alloc ids to errors for any allocations that could not be restored. +// +// If a fatal error was encountered it will be returned and the other two +// values will be nil. +func (s *BoltStateDB) GetAllAllocations() ([]*structs.Allocation, map[string]error, error) { + var allocs []*structs.Allocation + var errs map[string]error + err := s.db.View(func(tx *bolt.Tx) error { + allocs, errs = getAllAllocations(tx) + return nil + }) + + // db.View itself may return an error, so still check + if err != nil { + return nil, nil, err + } + + return allocs, errs, nil +} + +// PutAllocation stores an allocation or returns an error. +func (s *BoltStateDB) PutAllocation(alloc *structs.Allocation) error { + return s.db.Update(func(tx *bolt.Tx) error { + // Retrieve the root allocations bucket + allocsBkt, err := tx.CreateBucketIfNotExists(allocationsBucket) + if err != nil { + return err + } + + // Retrieve the specific allocations bucket + key := []byte(alloc.ID) + allocBkt, err := allocsBkt.CreateBucketIfNotExists(key) + if err != nil { + return err + } + + allocState := allocEntry{ + Alloc: alloc, + } + return putObject(allocBkt, allocKey, &allocState) + }) +} + +// GetTaskRunnerState restores TaskRunner specific state. +//TODO wrap in a single struct? +func (s *BoltStateDB) GetTaskRunnerState(allocID, taskName string) (*trstate.LocalState, *structs.TaskState, error) { + var ls trstate.LocalState + var ts structs.TaskState + + err := s.db.View(func(tx *bolt.Tx) error { + bkt, err := getTaskBucket(tx, allocID, taskName) + if err != nil { + return fmt.Errorf("failed to get task %q bucket: %v", taskName, err) + } + + // Restore Local State + //XXX set persisted hash to avoid immediate write on first use? + if err := getObject(bkt, taskLocalStateKey, &ls); err != nil { + return fmt.Errorf("failed to read local task runner state: %v", err) + } + + // Restore Task State + if err := getObject(bkt, taskStateKey, &ts); err != nil { + return fmt.Errorf("failed to read task state: %v", err) + } + + // XXX if driver has task { + // tr.restoreDriver() + // } + + return nil + }) + + if err != nil { + return nil, nil, err + } + + return &ls, &ts, nil + +} + +// PutTaskRunnerLocalState stores TaskRunner's LocalState or returns an error. +// It is up to the caller to serialize the state to bytes. +func (s *BoltStateDB) PutTaskRunnerLocalState(allocID, taskName string, buf []byte) error { + return s.db.Update(func(tx *bolt.Tx) error { + taskBkt, err := getTaskBucket(tx, allocID, taskName) + if err != nil { + return fmt.Errorf("failed to retrieve allocation bucket: %v", err) + } + + if err := putData(taskBkt, taskLocalStateKey, buf); err != nil { + return fmt.Errorf("failed to write task_runner state: %v", err) + } + + return nil + }) +} + +// PutTaskState stores a task's state or returns an error. +func (s *BoltStateDB) PutTaskState(allocID, taskName string, state *structs.TaskState) error { + return s.db.Update(func(tx *bolt.Tx) error { + taskBkt, err := getTaskBucket(tx, allocID, taskName) + if err != nil { + return fmt.Errorf("failed to retrieve allocation bucket: %v", err) + } + + return putObject(taskBkt, taskStateKey, state) + }) +} + +// Close releases all database resources and unlocks the database file on disk. +// All transactions must be closed before closing the database. +func (s *BoltStateDB) Close() error { + return s.db.Close() +} + +func putObject(bkt *bolt.Bucket, key []byte, obj interface{}) error { + if !bkt.Writable() { + return fmt.Errorf("bucket must be writable") + } + + // Serialize the object + var buf bytes.Buffer + if err := codec.NewEncoder(&buf, structs.MsgpackHandle).Encode(obj); err != nil { + return fmt.Errorf("failed to encode passed object: %v", err) + } + + if err := bkt.Put(key, buf.Bytes()); err != nil { + return fmt.Errorf("failed to write data at key %v: %v", string(key), err) + } + + return nil +} + +func putData(bkt *bolt.Bucket, key, value []byte) error { + if !bkt.Writable() { + return fmt.Errorf("bucket must be writable") + } + + if err := bkt.Put(key, value); err != nil { + return fmt.Errorf("failed to write data at key %v: %v", string(key), err) + } + + return nil +} + +func getObject(bkt *bolt.Bucket, key []byte, obj interface{}) error { + // Get the data + data := bkt.Get(key) + if data == nil { + return fmt.Errorf("no data at key %v", string(key)) + } + + // Deserialize the object + if err := codec.NewDecoderBytes(data, structs.MsgpackHandle).Decode(obj); err != nil { + return fmt.Errorf("failed to decode data into passed object: %v", err) + } + + return nil +} + +// getAllocationBucket returns the bucket used to persist state about a +// particular allocation. If the root allocation bucket or the specific +// allocation bucket doesn't exist, it will be created as long as the +// transaction is writable. +func getAllocationBucket(tx *bolt.Tx, allocID string) (*bolt.Bucket, error) { + var err error + w := tx.Writable() + + // Retrieve the root allocations bucket + allocations := tx.Bucket(allocationsBucket) + if allocations == nil { + if !w { + return nil, fmt.Errorf("Allocations bucket doesn't exist and transaction is not writable") + } + + allocations, err = tx.CreateBucket(allocationsBucket) + if err != nil { + return nil, err + } } // Retrieve the specific allocations bucket - key := []byte(alloc.ID) - allocBkt, err := allocsBkt.CreateBucketIfNotExists(key) - if err != nil { - return err + key := []byte(allocID) + alloc := allocations.Bucket(key) + if alloc == nil { + if !w { + return nil, fmt.Errorf("Allocation bucket doesn't exist and transaction is not writable") + } + + alloc, err = allocations.CreateBucket(key) + if err != nil { + return nil, err + } } - allocState := allocRunnerAllocState{ - Alloc: alloc, - } - return PutObject(allocBkt, allocRunnerStateAllocKey, &allocState) + return alloc, nil +} + +// getTaskBucket returns the bucket used to persist state about a +// particular task. If the root allocation bucket, the specific +// allocation or task bucket doesn't exist, they will be created as long as the +// transaction is writable. +func getTaskBucket(tx *bolt.Tx, allocID, taskName string) (*bolt.Bucket, error) { + alloc, err := getAllocationBucket(tx, allocID) + if err != nil { + return nil, err + } + + // Retrieve the specific task bucket + w := tx.Writable() + key := []byte(taskName) + task := alloc.Bucket(key) + if task == nil { + if !w { + return nil, fmt.Errorf("Task bucket doesn't exist and transaction is not writable") + } + + task, err = alloc.CreateBucket(key) + if err != nil { + return nil, err + } + } + + return task, nil } From a3fe0510d198664e524b00280186e5a05b4f2b96 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 8 Aug 2018 17:06:56 -0700 Subject: [PATCH 045/152] Move all encoding and put deduping into state db Still WIP as it does not handle deletions. --- .../allocrunnerv2/taskrunner/task_runner.go | 39 +--- client/client.go | 2 +- client/state/interface.go | 2 +- client/state/kvcodec.go | 104 +++++++++++ client/state/kvcodec_test.go | 70 +++++++ client/state/noopdb.go | 2 +- client/state/state_database.go | 173 +++++++----------- 7 files changed, 249 insertions(+), 143 deletions(-) create mode 100644 client/state/kvcodec.go create mode 100644 client/state/kvcodec_test.go diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 45ecb158e..9844bb28e 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -1,10 +1,8 @@ package taskrunner import ( - "bytes" "context" "fmt" - "io" "sync" "time" @@ -21,8 +19,6 @@ import ( cstate "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" - "github.com/ugorji/go/codec" - "golang.org/x/crypto/blake2b" ) const ( @@ -453,44 +449,11 @@ func (tr *TaskRunner) handleDestroy(handle driver.DriverHandle) (destroyed bool, } // persistLocalState persists local state to disk synchronously. -// -//XXX Not safe for concurrent calls. Should it be? func (tr *TaskRunner) persistLocalState() error { tr.localStateLock.Lock() defer tr.localStateLock.Unlock() - // buffer for writing serialized state to - var buf bytes.Buffer - - // Hash for skipping unnecessary writes - h, err := blake2b.New256(nil) - if err != nil { - // Programming error that should never happen! - return err - } - - // Multiplex writes to both - w := io.MultiWriter(h, &buf) - - // Encode as msgpack value - if err := codec.NewEncoder(w, structs.MsgpackHandle).Encode(&tr.localState); err != nil { - return fmt.Errorf("failed to serialize snapshot: %v", err) - } - - // If the hashes are equal, skip the write - hashVal := h.Sum(nil) - if bytes.Equal(hashVal, tr.persistedHash) { - return nil - } - - if err := tr.stateDB.PutTaskRunnerLocalState(tr.allocID, tr.taskName, buf.Bytes()); err != nil { - return err - } - - // State was persisted, set the hash - tr.persistedHash = hashVal - - return nil + return tr.stateDB.PutTaskRunnerLocalState(tr.allocID, tr.taskName, tr.localState) } // XXX If the objects don't exists since the client shutdown before the task diff --git a/client/client.go b/client/client.go index e52d5b917..13b3ec415 100644 --- a/client/client.go +++ b/client/client.go @@ -381,7 +381,7 @@ func (c *Client) init() error { c.logger.Printf("[INFO] client: using state directory %v", c.config.StateDir) // Open the state database - db, err := state.NewStateDB(c.config.StateDir, c.config.DevMode) + db, err := state.GetStateDBFactory(c.config.DevMode)(c.config.StateDir) if err != nil { return fmt.Errorf("failed to open state database: %v", err) } diff --git a/client/state/interface.go b/client/state/interface.go index e4d60f211..59d6bd08b 100644 --- a/client/state/interface.go +++ b/client/state/interface.go @@ -10,7 +10,7 @@ type StateDB interface { GetAllAllocations() ([]*structs.Allocation, map[string]error, error) PutAllocation(*structs.Allocation) error GetTaskRunnerState(allocID, taskName string) (*state.LocalState, *structs.TaskState, error) - PutTaskRunnerLocalState(allocID, taskName string, buf []byte) error + PutTaskRunnerLocalState(allocID, taskName string, val interface{}) error PutTaskState(allocID, taskName string, state *structs.TaskState) error Close() error } diff --git a/client/state/kvcodec.go b/client/state/kvcodec.go new file mode 100644 index 000000000..581ea528c --- /dev/null +++ b/client/state/kvcodec.go @@ -0,0 +1,104 @@ +package state + +import ( + "bytes" + "fmt" + "io" + "sync" + + "github.com/hashicorp/nomad/nomad/structs" + "github.com/ugorji/go/codec" + "golang.org/x/crypto/blake2b" +) + +type kvStore interface { + Get(key []byte) (val []byte) + Put(key, val []byte) error + Writable() bool +} + +// keyValueCodec handles encoding and decoding values from a key/value store +// such as boltdb. +type keyValueCodec struct { + // hashes maps keys to the hash of the last content written + hashes map[string][]byte + hashesLock sync.Mutex +} + +func newKeyValueCodec() *keyValueCodec { + return &keyValueCodec{ + hashes: make(map[string][]byte), + } +} + +// hashKey returns a unique key for each hashed boltdb value +func (c *keyValueCodec) hashKey(path string, key []byte) string { + return path + "-" + string(key) +} + +// Put into kv store iff it has changed since the last write. A globally +// unique key is constructed for each value by concatinating the path and key +// passed in. +func (c *keyValueCodec) Put(bkt kvStore, path string, key []byte, val interface{}) error { + if !bkt.Writable() { + return fmt.Errorf("bucket must be writable") + } + + // buffer for writing serialized state to + var buf bytes.Buffer + + // Hash for skipping unnecessary writes + h, err := blake2b.New256(nil) + if err != nil { + // Programming error that should never happen! + return err + } + + // Multiplex writes to both hasher and buffer + w := io.MultiWriter(h, &buf) + + // Serialize the object + if err := codec.NewEncoder(w, structs.MsgpackHandle).Encode(val); err != nil { + return fmt.Errorf("failed to encode passed object: %v", err) + } + + // If the hashes are equal, skip the write + hashVal := h.Sum(nil) + hashKey := c.hashKey(path, key) + + c.hashesLock.Lock() + persistedHash := c.hashes[hashKey] + c.hashesLock.Unlock() + + if bytes.Equal(hashVal, persistedHash) { + return nil + } + + if err := bkt.Put(key, buf.Bytes()); err != nil { + return fmt.Errorf("failed to write data at key %s: %v", key, err) + } + + // New value written, store hash + c.hashesLock.Lock() + c.hashes[hashKey] = hashVal + c.hashesLock.Unlock() + + return nil + +} + +// Get value by key from boltdb. +func (c *keyValueCodec) Get(bkt kvStore, key []byte, obj interface{}) error { + // Get the data + data := bkt.Get(key) + if data == nil { + return fmt.Errorf("no data at key %v", string(key)) + } + + // Deserialize the object + if err := codec.NewDecoderBytes(data, structs.MsgpackHandle).Decode(obj); err != nil { + return fmt.Errorf("failed to decode data into passed object: %v", err) + } + + return nil +} diff --git a/client/state/kvcodec_test.go b/client/state/kvcodec_test.go new file mode 100644 index 000000000..49c110b91 --- /dev/null +++ b/client/state/kvcodec_test.go @@ -0,0 +1,70 @@ +package state + +import ( + "testing" + + "github.com/stretchr/testify/require" +) + +// mockKVStore tracks puts and is useful for testing KVCodec's write-on-change +// code. +type mockKVStore struct { + puts int +} + +func (mockKVStore) Get(key []byte) (val []byte) { + return nil +} + +func (m *mockKVStore) Put(key, val []byte) error { + m.puts++ + return nil +} + +func (mockKVStore) Writable() bool { + return true +} + +// TestKVCodec_PutHash asserts that Puts on the underlying kvstore only occur +// when the data actually changes. +func TestKVCodec_PutHash(t *testing.T) { + require := require.New(t) + codec := newKeyValueCodec() + + // Create arguments for Put + kv := new(mockKVStore) + path := "path-path" + key := []byte("key1") + val := &struct { + Val int + }{ + Val: 1, + } + + // Initial Put should be written + require.NoError(codec.Put(kv, path, key, val)) + require.Equal(1, kv.puts) + + // Writing the same values again should be a noop + require.NoError(codec.Put(kv, path, key, val)) + require.Equal(1, kv.puts) + + // Changing the value should write again + val.Val++ + require.NoError(codec.Put(kv, path, key, val)) + require.Equal(2, kv.puts) + + // Changing the key should write again + key = []byte("key2") + require.NoError(codec.Put(kv, path, key, val)) + require.Equal(3, kv.puts) + + // Changing the path should write again + path = "new-path" + require.NoError(codec.Put(kv, path, key, val)) + require.Equal(4, kv.puts) + + // Writing the same values again should be a noop + require.NoError(codec.Put(kv, path, key, val)) + require.Equal(4, kv.puts) +} diff --git a/client/state/noopdb.go b/client/state/noopdb.go index 67f5f1760..7a31c461f 100644 --- a/client/state/noopdb.go +++ b/client/state/noopdb.go @@ -19,7 +19,7 @@ func (n noopDB) GetTaskRunnerState(allocID string, taskName string) (*state.Loca return nil, nil, nil } -func (n noopDB) PutTaskRunnerLocalState(allocID string, taskName string, buf []byte) error { +func (n noopDB) PutTaskRunnerLocalState(allocID string, taskName string, val interface{}) error { return nil } diff --git a/client/state/state_database.go b/client/state/state_database.go index 1828f62a2..8793d651e 100644 --- a/client/state/state_database.go +++ b/client/state/state_database.go @@ -1,14 +1,13 @@ package state import ( - "bytes" "fmt" "path/filepath" + "strings" "github.com/boltdb/bolt" trstate "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" "github.com/hashicorp/nomad/nomad/structs" - "github.com/ugorji/go/codec" ) /* @@ -40,6 +39,7 @@ var ( taskStateKey = []byte("task_state") ) +//TODO delete from kvcodec // DeleteAllocationBucket is used to delete an allocation bucket if it exists. func DeleteAllocationBucket(tx *bolt.Tx, allocID string) error { if !tx.Writable() { @@ -61,6 +61,7 @@ func DeleteAllocationBucket(tx *bolt.Tx, allocID string) error { return allocations.DeleteBucket(key) } +//TODO delete from kvcodec // DeleteTaskBucket is used to delete a task bucket if it exists. func DeleteTaskBucket(tx *bolt.Tx, allocID, taskName string) error { if !tx.Writable() { @@ -88,13 +89,70 @@ func DeleteTaskBucket(tx *bolt.Tx, allocID, taskName string) error { return alloc.DeleteBucket(key) } -var () +// NewStateDBFunc creates a StateDB given a state directory. +type NewStateDBFunc func(stateDir string) (StateDB, error) +// GetStateDBFactory returns a func for creating a StateDB +func GetStateDBFactory(devMode bool) NewStateDBFunc { + // Return a noop state db implementation when in debug mode + if devMode { + return func(string) (StateDB, error) { + return noopDB{}, nil + } + } + + return NewBoltStateDB +} + +// BoltStateDB persists and restores Nomad client state in a boltdb. All +// methods are safe for concurrent access. Create via NewStateDB by setting +// devMode=false. +type BoltStateDB struct { + db *bolt.DB + codec *keyValueCodec +} + +func NewBoltStateDB(stateDir string) (StateDB, error) { + // Create or open the boltdb state database + db, err := bolt.Open(filepath.Join(stateDir, "state.db"), 0600, nil) + if err != nil { + return nil, fmt.Errorf("failed to create state database: %v", err) + } + + sdb := &BoltStateDB{ + db: db, + codec: newKeyValueCodec(), + } + return sdb, nil +} + +// GetAllAllocations gets all allocations persisted by this client and returns +// a map of alloc ids to errors for any allocations that could not be restored. +// +// If a fatal error was encountered it will be returned and the other two +// values will be nil. +func (s *BoltStateDB) GetAllAllocations() ([]*structs.Allocation, map[string]error, error) { + var allocs []*structs.Allocation + var errs map[string]error + err := s.db.View(func(tx *bolt.Tx) error { + allocs, errs = s.getAllAllocations(tx) + return nil + }) + + // db.View itself may return an error, so still check + if err != nil { + return nil, nil, err + } + + return allocs, errs, nil +} + +// allocEntry wraps values in the Allocations buckets type allocEntry struct { Alloc *structs.Allocation } -func getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map[string]error) { +func (s *BoltStateDB) getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map[string]error) { allocationsBkt := tx.Bucket(allocationsBucket) if allocationsBkt == nil { // No allocs @@ -117,7 +175,7 @@ func getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map[string]error) { } var allocState allocEntry - if err := getObject(allocBkt, allocKey, &allocState); err != nil { + if err := s.codec.Get(allocBkt, allocKey, &allocState); err != nil { errs[allocID] = fmt.Errorf("failed to decode alloc %v", err) continue } @@ -128,52 +186,6 @@ func getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map[string]error) { return allocs, errs } -// BoltStateDB persists and restores Nomad client state in a boltdb. All -// methods are safe for concurrent access. Create via NewStateDB by setting -// devMode=false. -type BoltStateDB struct { - db *bolt.DB -} - -func NewStateDB(stateDir string, devMode bool) (StateDB, error) { - // Return a noop state db implementation when in debug mode - if devMode { - return noopDB{}, nil - } - - // Create or open the boltdb state database - db, err := bolt.Open(filepath.Join(stateDir, "state.db"), 0600, nil) - if err != nil { - return nil, fmt.Errorf("failed to create state database: %v", err) - } - - sdb := &BoltStateDB{ - db: db, - } - return sdb, nil -} - -// GetAllAllocations gets all allocations persisted by this client and returns -// a map of alloc ids to errors for any allocations that could not be restored. -// -// If a fatal error was encountered it will be returned and the other two -// values will be nil. -func (s *BoltStateDB) GetAllAllocations() ([]*structs.Allocation, map[string]error, error) { - var allocs []*structs.Allocation - var errs map[string]error - err := s.db.View(func(tx *bolt.Tx) error { - allocs, errs = getAllAllocations(tx) - return nil - }) - - // db.View itself may return an error, so still check - if err != nil { - return nil, nil, err - } - - return allocs, errs, nil -} - // PutAllocation stores an allocation or returns an error. func (s *BoltStateDB) PutAllocation(alloc *structs.Allocation) error { return s.db.Update(func(tx *bolt.Tx) error { @@ -193,7 +205,7 @@ func (s *BoltStateDB) PutAllocation(alloc *structs.Allocation) error { allocState := allocEntry{ Alloc: alloc, } - return putObject(allocBkt, allocKey, &allocState) + return s.codec.Put(allocBkt, alloc.ID, allocKey, &allocState) }) } @@ -211,12 +223,12 @@ func (s *BoltStateDB) GetTaskRunnerState(allocID, taskName string) (*trstate.Loc // Restore Local State //XXX set persisted hash to avoid immediate write on first use? - if err := getObject(bkt, taskLocalStateKey, &ls); err != nil { + if err := s.codec.Get(bkt, taskLocalStateKey, &ls); err != nil { return fmt.Errorf("failed to read local task runner state: %v", err) } // Restore Task State - if err := getObject(bkt, taskStateKey, &ts); err != nil { + if err := s.codec.Get(bkt, taskStateKey, &ts); err != nil { return fmt.Errorf("failed to read task state: %v", err) } @@ -237,14 +249,15 @@ func (s *BoltStateDB) GetTaskRunnerState(allocID, taskName string) (*trstate.Loc // PutTaskRunnerLocalState stores TaskRunner's LocalState or returns an error. // It is up to the caller to serialize the state to bytes. -func (s *BoltStateDB) PutTaskRunnerLocalState(allocID, taskName string, buf []byte) error { +func (s *BoltStateDB) PutTaskRunnerLocalState(allocID, taskName string, val interface{}) error { return s.db.Update(func(tx *bolt.Tx) error { taskBkt, err := getTaskBucket(tx, allocID, taskName) if err != nil { return fmt.Errorf("failed to retrieve allocation bucket: %v", err) } - if err := putData(taskBkt, taskLocalStateKey, buf); err != nil { + path := strings.Join([]string{allocID, taskName, string(taskLocalStateKey)}, "-") + if err := s.codec.Put(taskBkt, path, taskLocalStateKey, val); err != nil { return fmt.Errorf("failed to write task_runner state: %v", err) } @@ -260,7 +273,8 @@ func (s *BoltStateDB) PutTaskState(allocID, taskName string, state *structs.Task return fmt.Errorf("failed to retrieve allocation bucket: %v", err) } - return putObject(taskBkt, taskStateKey, state) + path := strings.Join([]string{allocID, taskName, string(taskStateKey)}, "-") + return s.codec.Put(taskBkt, path, taskStateKey, state) }) } @@ -270,51 +284,6 @@ func (s *BoltStateDB) Close() error { return s.db.Close() } -func putObject(bkt *bolt.Bucket, key []byte, obj interface{}) error { - if !bkt.Writable() { - return fmt.Errorf("bucket must be writable") - } - - // Serialize the object - var buf bytes.Buffer - if err := codec.NewEncoder(&buf, structs.MsgpackHandle).Encode(obj); err != nil { - return fmt.Errorf("failed to encode passed object: %v", err) - } - - if err := bkt.Put(key, buf.Bytes()); err != nil { - return fmt.Errorf("failed to write data at key %v: %v", string(key), err) - } - - return nil -} - -func putData(bkt *bolt.Bucket, key, value []byte) error { - if !bkt.Writable() { - return fmt.Errorf("bucket must be writable") - } - - if err := bkt.Put(key, value); err != nil { - return fmt.Errorf("failed to write data at key %v: %v", string(key), err) - } - - return nil -} - -func getObject(bkt *bolt.Bucket, key []byte, obj interface{}) error { - // Get the data - data := bkt.Get(key) - if data == nil { - return fmt.Errorf("no data at key %v", string(key)) - } - - // Deserialize the object - if err := codec.NewDecoderBytes(data, structs.MsgpackHandle).Decode(obj); err != nil { - return fmt.Errorf("failed to decode data into passed object: %v", err) - } - - return nil -} - // getAllocationBucket returns the bucket used to persist state about a // particular allocation. If the root allocation bucket or the specific // allocation bucket doesn't exist, it will be created as long as the From 990228a6e2dcf9591ba9c68ffca4d0428095521a Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 15 Aug 2018 14:03:18 -0700 Subject: [PATCH 046/152] wip wrap boltdb to get path information finished but doesn't handle deleting deeply nested buckets --- client/state/kvcodec.go | 66 +++++++++---- client/state/kvcodec_test.go | 42 +++++---- client/state/named_bucket.go | 123 ++++++++++++++++++++++++ client/state/named_bucket_test.go | 151 ++++++++++++++++++++++++++++++ client/state/state_database.go | 108 +++++++++------------ 5 files changed, 392 insertions(+), 98 deletions(-) create mode 100644 client/state/named_bucket.go create mode 100644 client/state/named_bucket_test.go diff --git a/client/state/kvcodec.go b/client/state/kvcodec.go index 581ea528c..61bca1a4f 100644 --- a/client/state/kvcodec.go +++ b/client/state/kvcodec.go @@ -12,38 +12,36 @@ import ( ) type kvStore interface { + Path() []byte + Bucket(name []byte) kvStore + CreateBucket(key []byte) (kvStore, error) + CreateBucketIfNotExists(key []byte) (kvStore, error) + DeleteBucket(key []byte) error Get(key []byte) (val []byte) Put(key, val []byte) error - Writable() bool } // keyValueCodec handles encoding and decoding values from a key/value store // such as boltdb. type keyValueCodec struct { - // hashes maps keys to the hash of the last content written - hashes map[string][]byte + // hashes maps buckets to keys to the hash of the last content written: + // bucket -> key -> hash for example: + // allocations/1234 -> alloc -> abcd + // allocations/1234/redis -> task_state -> efff + hashes map[string]map[string][]byte hashesLock sync.Mutex } func newKeyValueCodec() *keyValueCodec { return &keyValueCodec{ - hashes: make(map[string][]byte), + hashes: make(map[string]map[string][]byte), } } -// hashKey returns a unique key for each hashed boltdb value -func (c *keyValueCodec) hashKey(path string, key []byte) string { - return path + "-" + string(key) -} - // Put into kv store iff it has changed since the last write. A globally // unique key is constructed for each value by concatinating the path and key // passed in. -func (c *keyValueCodec) Put(bkt kvStore, path string, key []byte, val interface{}) error { - if !bkt.Writable() { - return fmt.Errorf("bucket must be writable") - } - +func (c *keyValueCodec) Put(bkt kvStore, key []byte, val interface{}) error { // buffer for writing serialized state to var buf bytes.Buffer @@ -63,24 +61,34 @@ func (c *keyValueCodec) Put(bkt kvStore, path string, key []byte, val interface{ } // If the hashes are equal, skip the write + hashPath := string(bkt.Path()) + hashKey := string(key) hashVal := h.Sum(nil) - hashKey := c.hashKey(path, key) + + // lastHash value or nil if it hasn't been hashed yet + var lastHash []byte c.hashesLock.Lock() - persistedHash := c.hashes[hashKey] + if hashBkt, ok := c.hashes[hashPath]; ok { + lastHash = hashBkt[hashKey] + } else { + // Create hash bucket + c.hashes[hashPath] = make(map[string][]byte, 2) + } c.hashesLock.Unlock() - if bytes.Equal(hashVal, persistedHash) { + if bytes.Equal(hashVal, lastHash) { return nil } + // New value: write it to the underlying store if err := bkt.Put(key, buf.Bytes()); err != nil { return fmt.Errorf("failed to write data at key %s: %v", key, err) } - // New value written, store hash + // New value written, store hash (bucket path map was created above) c.hashesLock.Lock() - c.hashes[hashKey] = hashVal + c.hashes[hashPath][hashKey] = hashVal c.hashesLock.Unlock() return nil @@ -102,3 +110,23 @@ func (c *keyValueCodec) Get(bkt kvStore, key []byte, obj interface{}) error { return nil } + +// DeleteBucket or do nothing if bucket doesn't exist. +func (c *keyValueCodec) DeleteBucket(parent kvStore, bktName []byte) error { + // Get the path of the bucket being deleted + bkt := parent.Bucket(bktName) + if bkt == nil { + // Doesn't exist! Nothing to delete + return nil + } + + // Delete the bucket + err := parent.DeleteBucket(bktName) + + // Always purge all corresponding hashes to prevent memory leaks + c.hashesLock.Lock() + delete(c.hashes, string(bkt.Path())) + c.hashesLock.Unlock() + + return err +} diff --git a/client/state/kvcodec_test.go b/client/state/kvcodec_test.go index 49c110b91..ee625d016 100644 --- a/client/state/kvcodec_test.go +++ b/client/state/kvcodec_test.go @@ -12,6 +12,26 @@ type mockKVStore struct { puts int } +func (mockKVStore) Path() []byte { + return []byte{} +} + +func (m *mockKVStore) Bucket(name []byte) kvStore { + return m +} + +func (m *mockKVStore) CreateBucket(key []byte) (kvStore, error) { + return m, nil +} + +func (m *mockKVStore) CreateBucketIfNotExists(key []byte) (kvStore, error) { + return m, nil +} + +func (m *mockKVStore) DeleteBucket(key []byte) error { + return nil +} + func (mockKVStore) Get(key []byte) (val []byte) { return nil } @@ -21,10 +41,6 @@ func (m *mockKVStore) Put(key, val []byte) error { return nil } -func (mockKVStore) Writable() bool { - return true -} - // TestKVCodec_PutHash asserts that Puts on the underlying kvstore only occur // when the data actually changes. func TestKVCodec_PutHash(t *testing.T) { @@ -33,7 +49,6 @@ func TestKVCodec_PutHash(t *testing.T) { // Create arguments for Put kv := new(mockKVStore) - path := "path-path" key := []byte("key1") val := &struct { Val int @@ -42,29 +57,24 @@ func TestKVCodec_PutHash(t *testing.T) { } // Initial Put should be written - require.NoError(codec.Put(kv, path, key, val)) + require.NoError(codec.Put(kv, key, val)) require.Equal(1, kv.puts) // Writing the same values again should be a noop - require.NoError(codec.Put(kv, path, key, val)) + require.NoError(codec.Put(kv, key, val)) require.Equal(1, kv.puts) // Changing the value should write again val.Val++ - require.NoError(codec.Put(kv, path, key, val)) + require.NoError(codec.Put(kv, key, val)) require.Equal(2, kv.puts) // Changing the key should write again key = []byte("key2") - require.NoError(codec.Put(kv, path, key, val)) + require.NoError(codec.Put(kv, key, val)) require.Equal(3, kv.puts) - // Changing the path should write again - path = "new-path" - require.NoError(codec.Put(kv, path, key, val)) - require.Equal(4, kv.puts) - // Writing the same values again should be a noop - require.NoError(codec.Put(kv, path, key, val)) - require.Equal(4, kv.puts) + require.NoError(codec.Put(kv, key, val)) + require.Equal(3, kv.puts) } diff --git a/client/state/named_bucket.go b/client/state/named_bucket.go new file mode 100644 index 000000000..b5383e838 --- /dev/null +++ b/client/state/named_bucket.go @@ -0,0 +1,123 @@ +package state + +import "github.com/boltdb/bolt" + +// namedBucket is a wrapper around bolt.Bucket's to preserve their path +// information and expose it via the Path() method. +// +// Knowing the full bucket path to a key is necessary for tracking accesses in +// another datastructure such as the hashing writer keyValueCodec. +type namedBucket struct { + path []byte + name []byte + bkt *bolt.Bucket +} + +// newNamedBucket from a bolt transaction. +func newNamedBucket(tx *bolt.Tx, root []byte) *namedBucket { + b := tx.Bucket(root) + if b == nil { + return nil + } + + return &namedBucket{ + path: root, + name: root, + bkt: b, + } +} + +// createNamedBucketIfNotExists from a bolt transaction. +func createNamedBucketIfNotExists(tx *bolt.Tx, root []byte) (*namedBucket, error) { + b, err := tx.CreateBucketIfNotExists(root) + if err != nil { + return nil, err + } + + return &namedBucket{ + path: root, + name: root, + bkt: b, + }, nil +} + +// Path to this bucket (including this bucket). +func (n *namedBucket) Path() []byte { + return n.path +} + +// Name of this bucket. +func (n *namedBucket) Name() []byte { + return n.name +} + +// Bucket returns a bucket inside the current one or nil if the bucket does not +// exist. +func (n *namedBucket) Bucket(name []byte) kvStore { + b := n.bkt.Bucket(name) + if b == nil { + return nil + } + + return &namedBucket{ + path: n.chBkt(name), + name: name, + bkt: b, + } +} + +// CreateBucketIfNotExists creates a bucket if it doesn't exist and returns it +// or an error. +func (n *namedBucket) CreateBucketIfNotExists(name []byte) (kvStore, error) { + b, err := n.bkt.CreateBucketIfNotExists(name) + if err != nil { + return nil, err + } + + return &namedBucket{ + path: n.chBkt(name), + name: name, + bkt: b, + }, nil +} + +// CreateBucket creates a bucket and returns it. +func (n *namedBucket) CreateBucket(name []byte) (kvStore, error) { + b, err := n.bkt.CreateBucket(name) + if err != nil { + return nil, err + } + + return &namedBucket{ + path: n.chBkt(name), + name: name, + bkt: b, + }, nil +} + +// DeleteBucket calls DeleteBucket on the underlying bolt.Bucket. +func (n *namedBucket) DeleteBucket(name []byte) error { + return n.bkt.DeleteBucket(name) +} + +// Get calls Get on the underlying bolt.Bucket. +func (n *namedBucket) Get(key []byte) []byte { + return n.bkt.Get(key) +} + +// Put calls Put on the underlying bolt.Bucket. +func (n *namedBucket) Put(key, value []byte) error { + return n.bkt.Put(key, value) +} + +// chBkt is like chdir but for buckets: it appends the new name to the end of +// a copy of the path and returns it. +func (n *namedBucket) chBkt(name []byte) []byte { + // existing path + new path element + path separator + path := make([]byte, len(n.path)+len(name)+1) + copy(path[0:len(n.path)], n.path) + path[len(n.path)] = '/' + copy(path[len(n.path)+1:], name) + + return path +} diff --git a/client/state/named_bucket_test.go b/client/state/named_bucket_test.go new file mode 100644 index 000000000..705483a16 --- /dev/null +++ b/client/state/named_bucket_test.go @@ -0,0 +1,151 @@ +package state + +import ( + "io/ioutil" + "os" + "path/filepath" + "testing" + + "github.com/boltdb/bolt" + "github.com/stretchr/testify/require" +) + +func setupBoltDB(t *testing.T) (*bolt.DB, func()) { + dir, err := ioutil.TempDir("", "nomadtest_") + require.NoError(t, err) + cleanup := func() { + if err := os.RemoveAll(dir); err != nil { + t.Logf("error removing test dir: %v", err) + } + } + + dbFilename := filepath.Join(dir, "nomadtest.db") + db, err := bolt.Open(dbFilename, 0600, nil) + if err != nil { + cleanup() + t.Fatalf("error creating boltdb: %v", err) + } + + return db, func() { + db.Close() + cleanup() + } +} + +// TestNamedBucket_Path asserts that creating and changing buckets are tracked +// properly by the namedBucket wrapper. +func TestNamedBucket_Path(t *testing.T) { + t.Parallel() + require := require.New(t) + + db, cleanup := setupBoltDB(t) + defer cleanup() + + parentBktName, childBktName := []byte("root"), []byte("child") + parentKey, parentVal := []byte("pkey"), []byte("pval") + childKey, childVal := []byte("ckey"), []byte("cval") + + require.NoError(db.Update(func(tx *bolt.Tx) error { + // Trying to open a named bucket from a nonexistent bucket + // should return nil. + require.Nil(newNamedBucket(tx, []byte("nonexistent"))) + + // Creating a named bucket from a bolt tx should work and set + // the path and name properly. + b, err := createNamedBucketIfNotExists(tx, parentBktName) + require.NoError(err) + require.Equal(parentBktName, b.Name()) + require.Equal(parentBktName, b.Path()) + + // Trying to descend into a nonexistent bucket should return + // nil. + require.Nil(b.Bucket([]byte("nonexistent"))) + + // Descending into a new bucket should update the path. + childBkt, err := b.CreateBucket(childBktName) + require.NoError(err) + require.Equal(childBktName, childBkt.(*namedBucket).Name()) + require.Equal([]byte("root/child"), childBkt.Path()) + + // Assert the parent bucket did not get changed. + require.Equal(parentBktName, b.Name()) + require.Equal(parentBktName, b.Path()) + + // Add entries to both buckets + require.NoError(b.Put(parentKey, parentVal)) + require.NoError(childBkt.Put(childKey, childVal)) + return nil + })) + + // Read buckets and values back out + require.NoError(db.View(func(tx *bolt.Tx) error { + b := newNamedBucket(tx, parentBktName) + require.NotNil(b) + require.Equal(parentVal, b.Get(parentKey)) + require.Nil(b.Get(childKey)) + + childBkt := b.Bucket(childBktName) + require.NotNil(childBkt) + require.Nil(childBkt.Get(parentKey)) + require.Equal(childVal, childBkt.Get(childKey)) + return nil + })) +} + +// TestNamedBucket_DeleteBucket asserts that deleting a bucket properly purges +// all related keys from the internal hashes map. +func TestNamedBucket_DeleteBucket(t *testing.T) { + t.Parallel() + require := require.New(t) + + db, cleanup := setupBoltDB(t) + defer cleanup() + + // Create some nested buckets and keys (key values will just be their names) + b1Name, c1Name, c2Name, c1c1Name := []byte("b1"), []byte("c1"), []byte("c2"), []byte("c1c1") + b1k1, c1k1, c2k1, c1c1k1 := []byte("b1k1"), []byte("c1k1"), []byte("c2k1"), []byte("c1c1k1") + + codec := newKeyValueCodec() + + // Create initial db state + require.NoError(db.Update(func(tx *bolt.Tx) error { + // Create bucket 1 and key + b1, err := createNamedBucketIfNotExists(tx, b1Name) + require.NoError(err) + require.NoError(codec.Put(b1, b1k1, b1k1)) + + // Create child bucket 1 and key + c1, err := b1.CreateBucketIfNotExists(c1Name) + require.NoError(err) + require.NoError(codec.Put(c1, c1k1, c1k1)) + + // Create child-child bucket 1 and key + c1c1, err := c1.CreateBucketIfNotExists(c1c1Name) + require.NoError(err) + require.NoError(codec.Put(c1c1, c1c1k1, c1c1k1)) + + // Create child bucket 2 and key + c2, err := b1.CreateBucketIfNotExists(c2Name) + require.NoError(err) + require.NoError(codec.Put(c2, c2k1, c2k1)) + return nil + })) + + // codec should be tracking 4 hash buckets (b1, c1, c2, c1c1) + require.Len(codec.hashes, 4) + + // Delete c1 + require.NoError(db.Update(func(tx *bolt.Tx) error { + b1 := newNamedBucket(tx, b1Name) + return codec.DeleteBucket(b1, c1Name) + })) + + START HERE // We don't appear to be properly deleting the sub-bucket + // codec should be tracking 2 hash buckets (b1, c2) + require.Len(codec.hashes, 2) + + // Assert all of c1 is gone + require.NoError(db.View(func(tx *bolt.Tx) error { + return nil + })) +} diff --git a/client/state/state_database.go b/client/state/state_database.go index 8793d651e..4e0e2d946 100644 --- a/client/state/state_database.go +++ b/client/state/state_database.go @@ -3,7 +3,6 @@ package state import ( "fmt" "path/filepath" - "strings" "github.com/boltdb/bolt" trstate "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" @@ -39,56 +38,6 @@ var ( taskStateKey = []byte("task_state") ) -//TODO delete from kvcodec -// DeleteAllocationBucket is used to delete an allocation bucket if it exists. -func DeleteAllocationBucket(tx *bolt.Tx, allocID string) error { - if !tx.Writable() { - return fmt.Errorf("transaction must be writable") - } - - // Retrieve the root allocations bucket - allocations := tx.Bucket(allocationsBucket) - if allocations == nil { - return nil - } - - // Check if the bucket exists - key := []byte(allocID) - if allocBkt := allocations.Bucket(key); allocBkt == nil { - return nil - } - - return allocations.DeleteBucket(key) -} - -//TODO delete from kvcodec -// DeleteTaskBucket is used to delete a task bucket if it exists. -func DeleteTaskBucket(tx *bolt.Tx, allocID, taskName string) error { - if !tx.Writable() { - return fmt.Errorf("transaction must be writable") - } - - // Retrieve the root allocations bucket - allocations := tx.Bucket(allocationsBucket) - if allocations == nil { - return nil - } - - // Retrieve the specific allocations bucket - alloc := allocations.Bucket([]byte(allocID)) - if alloc == nil { - return nil - } - - // Check if the bucket exists - key := []byte(taskName) - if taskBkt := alloc.Bucket(key); taskBkt == nil { - return nil - } - - return alloc.DeleteBucket(key) -} - // NewStateDBFunc creates a StateDB given a state directory. type NewStateDBFunc func(stateDir string) (StateDB, error) @@ -153,7 +102,7 @@ type allocEntry struct { } func (s *BoltStateDB) getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map[string]error) { - allocationsBkt := tx.Bucket(allocationsBucket) + allocationsBkt := newNamedBucket(tx, allocationsBucket) if allocationsBkt == nil { // No allocs return nil, nil @@ -163,7 +112,7 @@ func (s *BoltStateDB) getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map errs := map[string]error{} // Create a cursor for iteration. - c := allocationsBkt.Cursor() + c := allocationsBkt.bkt.Cursor() // Iterate over all the allocation buckets for k, _ := c.First(); k != nil; k, _ = c.Next() { @@ -190,7 +139,7 @@ func (s *BoltStateDB) getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map func (s *BoltStateDB) PutAllocation(alloc *structs.Allocation) error { return s.db.Update(func(tx *bolt.Tx) error { // Retrieve the root allocations bucket - allocsBkt, err := tx.CreateBucketIfNotExists(allocationsBucket) + allocsBkt, err := createNamedBucketIfNotExists(tx, allocationsBucket) if err != nil { return err } @@ -205,7 +154,7 @@ func (s *BoltStateDB) PutAllocation(alloc *structs.Allocation) error { allocState := allocEntry{ Alloc: alloc, } - return s.codec.Put(allocBkt, alloc.ID, allocKey, &allocState) + return s.codec.Put(allocBkt, allocKey, &allocState) }) } @@ -256,8 +205,7 @@ func (s *BoltStateDB) PutTaskRunnerLocalState(allocID, taskName string, val inte return fmt.Errorf("failed to retrieve allocation bucket: %v", err) } - path := strings.Join([]string{allocID, taskName, string(taskLocalStateKey)}, "-") - if err := s.codec.Put(taskBkt, path, taskLocalStateKey, val); err != nil { + if err := s.codec.Put(taskBkt, taskLocalStateKey, val); err != nil { return fmt.Errorf("failed to write task_runner state: %v", err) } @@ -273,8 +221,42 @@ func (s *BoltStateDB) PutTaskState(allocID, taskName string, state *structs.Task return fmt.Errorf("failed to retrieve allocation bucket: %v", err) } - path := strings.Join([]string{allocID, taskName, string(taskStateKey)}, "-") - return s.codec.Put(taskBkt, path, taskStateKey, state) + return s.codec.Put(taskBkt, taskStateKey, state) + }) +} + +// DeleteTaskBucket is used to delete a task bucket if it exists. +func (s *BoltStateDB) DeleteTaskBucket(allocID, taskName string) error { + return s.db.Update(func(tx *bolt.Tx) error { + // Retrieve the root allocations bucket + allocations := newNamedBucket(tx, allocationsBucket) + if allocations == nil { + return nil + } + + // Retrieve the specific allocations bucket + alloc := allocations.Bucket([]byte(allocID)) + if alloc == nil { + return nil + } + + // Check if the bucket exists + key := []byte(taskName) + return s.codec.DeleteBucket(alloc, key) + }) +} + +// DeleteAllocationBucket is used to delete an allocation bucket if it exists. +func (s *BoltStateDB) DeleteAllocationBucket(allocID string) error { + return s.db.Update(func(tx *bolt.Tx) error { + // Retrieve the root allocations bucket + allocations := newNamedBucket(tx, allocationsBucket) + if allocations == nil { + return nil + } + + key := []byte(allocID) + return s.codec.DeleteBucket(allocations, key) }) } @@ -288,18 +270,18 @@ func (s *BoltStateDB) Close() error { // particular allocation. If the root allocation bucket or the specific // allocation bucket doesn't exist, it will be created as long as the // transaction is writable. -func getAllocationBucket(tx *bolt.Tx, allocID string) (*bolt.Bucket, error) { +func getAllocationBucket(tx *bolt.Tx, allocID string) (kvStore, error) { var err error w := tx.Writable() // Retrieve the root allocations bucket - allocations := tx.Bucket(allocationsBucket) + allocations := newNamedBucket(tx, allocationsBucket) if allocations == nil { if !w { return nil, fmt.Errorf("Allocations bucket doesn't exist and transaction is not writable") } - allocations, err = tx.CreateBucket(allocationsBucket) + allocations, err = createNamedBucketIfNotExists(tx, allocationsBucket) if err != nil { return nil, err } @@ -326,7 +308,7 @@ func getAllocationBucket(tx *bolt.Tx, allocID string) (*bolt.Bucket, error) { // particular task. If the root allocation bucket, the specific // allocation or task bucket doesn't exist, they will be created as long as the // transaction is writable. -func getTaskBucket(tx *bolt.Tx, allocID, taskName string) (*bolt.Bucket, error) { +func getTaskBucket(tx *bolt.Tx, allocID, taskName string) (kvStore, error) { alloc, err := getAllocationBucket(tx, allocID) if err != nil { return nil, err From 820af27171c58076fdbbe46926f120004a1e053c Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 16 Aug 2018 11:14:52 -0700 Subject: [PATCH 047/152] wrap boltdb in a write deduplicator Saves a tiny bit of cpu and some IO. Sadly doesn't prevent all IO on duplicate writes as the transactions are still created and committed. $ go test -bench=. -benchmem goos: linux goarch: amd64 pkg: github.com/hashicorp/nomad/helper/boltdd BenchmarkWriteDeduplication_On-4 500 4059591 ns/op 23736 B/op 56 allocs/op BenchmarkWriteDeduplication_Off-4 300 4115319 ns/op 25942 B/op 55 allocs/op --- client/allocrunner/alloc_runner.go | 18 +- client/allocrunner/taskrunner/task_runner.go | 19 +- client/state/kvcodec.go | 132 ------- client/state/kvcodec_test.go | 80 ----- client/state/named_bucket.go | 123 ------- client/state/named_bucket_test.go | 151 -------- client/state/state_database.go | 60 ++-- helper/boltdd/boltdd.go | 360 +++++++++++++++++++ helper/boltdd/boltdd_test.go | 344 ++++++++++++++++++ 9 files changed, 752 insertions(+), 535 deletions(-) delete mode 100644 client/state/kvcodec.go delete mode 100644 client/state/kvcodec_test.go delete mode 100644 client/state/named_bucket.go delete mode 100644 client/state/named_bucket_test.go create mode 100644 helper/boltdd/boltdd.go create mode 100644 helper/boltdd/boltdd_test.go diff --git a/client/allocrunner/alloc_runner.go b/client/allocrunner/alloc_runner.go index 9bf22149e..b60432332 100644 --- a/client/allocrunner/alloc_runner.go +++ b/client/allocrunner/alloc_runner.go @@ -15,7 +15,6 @@ import ( "github.com/hashicorp/nomad/client/allocrunner/taskrunner" "github.com/hashicorp/nomad/client/config" consulApi "github.com/hashicorp/nomad/client/consul" - "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/helper" "github.com/hashicorp/nomad/nomad/structs" @@ -432,15 +431,16 @@ func (r *AllocRunner) saveAllocRunnerState() error { // DestroyState is used to cleanup after ourselves func (r *AllocRunner) DestroyState() error { - r.allocStateLock.Lock() - defer r.allocStateLock.Unlock() + //r.allocStateLock.Lock() + //defer r.allocStateLock.Unlock() - return r.stateDB.Update(func(tx *bolt.Tx) error { - if err := state.DeleteAllocationBucket(tx, r.allocID); err != nil { - return fmt.Errorf("failed to delete allocation bucket: %v", err) - } - return nil - }) + //return r.stateDB.Update(func(tx *bolt.Tx) error { + // if err := state.DeleteAllocationBucket(tx, r.allocID); err != nil { + // return fmt.Errorf("failed to delete allocation bucket: %v", err) + // } + // return nil + //}) + panic("deprecated: use allocrunnerv2") } // DestroyContext is used to destroy the context diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index 60032e60d..386fb672a 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -26,7 +26,6 @@ import ( "github.com/hashicorp/nomad/client/config" consulApi "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/driver" - "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" "github.com/ugorji/go/codec" @@ -536,15 +535,17 @@ func (r *TaskRunner) SaveState() error { // DestroyState is used to cleanup after ourselves func (r *TaskRunner) DestroyState() error { - r.persistLock.Lock() - defer r.persistLock.Unlock() + //r.persistLock.Lock() + //defer r.persistLock.Unlock() - return r.stateDB.Update(func(tx *bolt.Tx) error { - if err := state.DeleteTaskBucket(tx, r.alloc.ID, r.task.Name); err != nil { - return fmt.Errorf("failed to delete task bucket: %v", err) - } - return nil - }) + //return r.stateDB.Update(func(tx *bolt.Tx) error { + // if err := state.DeleteTaskBucket(tx, r.alloc.ID, r.task.Name); err != nil { + // return fmt.Errorf("failed to delete task bucket: %v", err) + // } + // return nil + //}) + //XXX Deprecated: see allocrunnerv2 + panic("deprecated") } // setState is used to update the state of the task runner diff --git a/client/state/kvcodec.go b/client/state/kvcodec.go deleted file mode 100644 index 61bca1a4f..000000000 --- a/client/state/kvcodec.go +++ /dev/null @@ -1,132 +0,0 @@ -package state - -import ( - "bytes" - "fmt" - "io" - "sync" - - "github.com/hashicorp/nomad/nomad/structs" - "github.com/ugorji/go/codec" - "golang.org/x/crypto/blake2b" -) - -type kvStore interface { - Path() []byte - Bucket(name []byte) kvStore - CreateBucket(key []byte) (kvStore, error) - CreateBucketIfNotExists(key []byte) (kvStore, error) - DeleteBucket(key []byte) error - Get(key []byte) (val []byte) - Put(key, val []byte) error -} - -// keyValueCodec handles encoding and decoding values from a key/value store -// such as boltdb. -type keyValueCodec struct { - // hashes maps buckets to keys to the hash of the last content written: - // bucket -> key -> hash for example: - // allocations/1234 -> alloc -> abcd - // allocations/1234/redis -> task_state -> efff - hashes map[string]map[string][]byte - hashesLock sync.Mutex -} - -func newKeyValueCodec() *keyValueCodec { - return &keyValueCodec{ - hashes: make(map[string]map[string][]byte), - } -} - -// Put into kv store iff it has changed since the last write. A globally -// unique key is constructed for each value by concatinating the path and key -// passed in. -func (c *keyValueCodec) Put(bkt kvStore, key []byte, val interface{}) error { - // buffer for writing serialized state to - var buf bytes.Buffer - - // Hash for skipping unnecessary writes - h, err := blake2b.New256(nil) - if err != nil { - // Programming error that should never happen! - return err - } - - // Multiplex writes to both hasher and buffer - w := io.MultiWriter(h, &buf) - - // Serialize the object - if err := codec.NewEncoder(w, structs.MsgpackHandle).Encode(val); err != nil { - return fmt.Errorf("failed to encode passed object: %v", err) - } - - // If the hashes are equal, skip the write - hashPath := string(bkt.Path()) - hashKey := string(key) - hashVal := h.Sum(nil) - - // lastHash value or nil if it hasn't been hashed yet - var lastHash []byte - - c.hashesLock.Lock() - if hashBkt, ok := c.hashes[hashPath]; ok { - lastHash = hashBkt[hashKey] - } else { - // Create hash bucket - c.hashes[hashPath] = make(map[string][]byte, 2) - } - c.hashesLock.Unlock() - - if bytes.Equal(hashVal, lastHash) { - return nil - } - - // New value: write it to the underlying store - if err := bkt.Put(key, buf.Bytes()); err != nil { - return fmt.Errorf("failed to write data at key %s: %v", key, err) - } - - // New value written, store hash (bucket path map was created above) - c.hashesLock.Lock() - c.hashes[hashPath][hashKey] = hashVal - c.hashesLock.Unlock() - - return nil - -} - -// Get value by key from boltdb. -func (c *keyValueCodec) Get(bkt kvStore, key []byte, obj interface{}) error { - // Get the data - data := bkt.Get(key) - if data == nil { - return fmt.Errorf("no data at key %v", string(key)) - } - - // Deserialize the object - if err := codec.NewDecoderBytes(data, structs.MsgpackHandle).Decode(obj); err != nil { - return fmt.Errorf("failed to decode data into passed object: %v", err) - } - - return nil -} - -// DeleteBucket or do nothing if bucket doesn't exist. -func (c *keyValueCodec) DeleteBucket(parent kvStore, bktName []byte) error { - // Get the path of the bucket being deleted - bkt := parent.Bucket(bktName) - if bkt == nil { - // Doesn't exist! Nothing to delete - return nil - } - - // Delete the bucket - err := parent.DeleteBucket(bktName) - - // Always purge all corresponding hashes to prevent memory leaks - c.hashesLock.Lock() - delete(c.hashes, string(bkt.Path())) - c.hashesLock.Unlock() - - return err -} diff --git a/client/state/kvcodec_test.go b/client/state/kvcodec_test.go deleted file mode 100644 index ee625d016..000000000 --- a/client/state/kvcodec_test.go +++ /dev/null @@ -1,80 +0,0 @@ -package state - -import ( - "testing" - - "github.com/stretchr/testify/require" -) - -// mockKVStore tracks puts and is useful for testing KVCodec's write-on-change -// code. -type mockKVStore struct { - puts int -} - -func (mockKVStore) Path() []byte { - return []byte{} -} - -func (m *mockKVStore) Bucket(name []byte) kvStore { - return m -} - -func (m *mockKVStore) CreateBucket(key []byte) (kvStore, error) { - return m, nil -} - -func (m *mockKVStore) CreateBucketIfNotExists(key []byte) (kvStore, error) { - return m, nil -} - -func (m *mockKVStore) DeleteBucket(key []byte) error { - return nil -} - -func (mockKVStore) Get(key []byte) (val []byte) { - return nil -} - -func (m *mockKVStore) Put(key, val []byte) error { - m.puts++ - return nil -} - -// TestKVCodec_PutHash asserts that Puts on the underlying kvstore only occur -// when the data actually changes. -func TestKVCodec_PutHash(t *testing.T) { - require := require.New(t) - codec := newKeyValueCodec() - - // Create arguments for Put - kv := new(mockKVStore) - key := []byte("key1") - val := &struct { - Val int - }{ - Val: 1, - } - - // Initial Put should be written - require.NoError(codec.Put(kv, key, val)) - require.Equal(1, kv.puts) - - // Writing the same values again should be a noop - require.NoError(codec.Put(kv, key, val)) - require.Equal(1, kv.puts) - - // Changing the value should write again - val.Val++ - require.NoError(codec.Put(kv, key, val)) - require.Equal(2, kv.puts) - - // Changing the key should write again - key = []byte("key2") - require.NoError(codec.Put(kv, key, val)) - require.Equal(3, kv.puts) - - // Writing the same values again should be a noop - require.NoError(codec.Put(kv, key, val)) - require.Equal(3, kv.puts) -} diff --git a/client/state/named_bucket.go b/client/state/named_bucket.go deleted file mode 100644 index b5383e838..000000000 --- a/client/state/named_bucket.go +++ /dev/null @@ -1,123 +0,0 @@ -package state - -import "github.com/boltdb/bolt" - -// namedBucket is a wrapper around bolt.Bucket's to preserve their path -// information and expose it via the Path() method. -// -// Knowing the full bucket path to a key is necessary for tracking accesses in -// another datastructure such as the hashing writer keyValueCodec. -type namedBucket struct { - path []byte - name []byte - bkt *bolt.Bucket -} - -// newNamedBucket from a bolt transaction. -func newNamedBucket(tx *bolt.Tx, root []byte) *namedBucket { - b := tx.Bucket(root) - if b == nil { - return nil - } - - return &namedBucket{ - path: root, - name: root, - bkt: b, - } -} - -// createNamedBucketIfNotExists from a bolt transaction. -func createNamedBucketIfNotExists(tx *bolt.Tx, root []byte) (*namedBucket, error) { - b, err := tx.CreateBucketIfNotExists(root) - if err != nil { - return nil, err - } - - return &namedBucket{ - path: root, - name: root, - bkt: b, - }, nil -} - -// Path to this bucket (including this bucket). -func (n *namedBucket) Path() []byte { - return n.path -} - -// Name of this bucket. -func (n *namedBucket) Name() []byte { - return n.name -} - -// Bucket returns a bucket inside the current one or nil if the bucket does not -// exist. -func (n *namedBucket) Bucket(name []byte) kvStore { - b := n.bkt.Bucket(name) - if b == nil { - return nil - } - - return &namedBucket{ - path: n.chBkt(name), - name: name, - bkt: b, - } -} - -// CreateBucketIfNotExists creates a bucket if it doesn't exist and returns it -// or an error. -func (n *namedBucket) CreateBucketIfNotExists(name []byte) (kvStore, error) { - b, err := n.bkt.CreateBucketIfNotExists(name) - if err != nil { - return nil, err - } - - return &namedBucket{ - path: n.chBkt(name), - name: name, - bkt: b, - }, nil -} - -// CreateBucket creates a bucket and returns it. -func (n *namedBucket) CreateBucket(name []byte) (kvStore, error) { - b, err := n.bkt.CreateBucket(name) - if err != nil { - return nil, err - } - - return &namedBucket{ - path: n.chBkt(name), - name: name, - bkt: b, - }, nil -} - -// DeleteBucket calls DeleteBucket on the underlying bolt.Bucket. -func (n *namedBucket) DeleteBucket(name []byte) error { - return n.bkt.DeleteBucket(name) -} - -// Get calls Get on the underlying bolt.Bucket. -func (n *namedBucket) Get(key []byte) []byte { - return n.bkt.Get(key) -} - -// Put calls Put on the underlying bolt.Bucket. -func (n *namedBucket) Put(key, value []byte) error { - return n.bkt.Put(key, value) -} - -// chBkt is like chdir but for buckets: it appends the new name to the end of -// a copy of the path and returns it. -func (n *namedBucket) chBkt(name []byte) []byte { - // existing path + new path element + path separator - path := make([]byte, len(n.path)+len(name)+1) - copy(path[0:len(n.path)], n.path) - path[len(n.path)] = '/' - copy(path[len(n.path)+1:], name) - - return path -} diff --git a/client/state/named_bucket_test.go b/client/state/named_bucket_test.go deleted file mode 100644 index 705483a16..000000000 --- a/client/state/named_bucket_test.go +++ /dev/null @@ -1,151 +0,0 @@ -package state - -import ( - "io/ioutil" - "os" - "path/filepath" - "testing" - - "github.com/boltdb/bolt" - "github.com/stretchr/testify/require" -) - -func setupBoltDB(t *testing.T) (*bolt.DB, func()) { - dir, err := ioutil.TempDir("", "nomadtest_") - require.NoError(t, err) - cleanup := func() { - if err := os.RemoveAll(dir); err != nil { - t.Logf("error removing test dir: %v", err) - } - } - - dbFilename := filepath.Join(dir, "nomadtest.db") - db, err := bolt.Open(dbFilename, 0600, nil) - if err != nil { - cleanup() - t.Fatalf("error creating boltdb: %v", err) - } - - return db, func() { - db.Close() - cleanup() - } -} - -// TestNamedBucket_Path asserts that creating and changing buckets are tracked -// properly by the namedBucket wrapper. -func TestNamedBucket_Path(t *testing.T) { - t.Parallel() - require := require.New(t) - - db, cleanup := setupBoltDB(t) - defer cleanup() - - parentBktName, childBktName := []byte("root"), []byte("child") - parentKey, parentVal := []byte("pkey"), []byte("pval") - childKey, childVal := []byte("ckey"), []byte("cval") - - require.NoError(db.Update(func(tx *bolt.Tx) error { - // Trying to open a named bucket from a nonexistent bucket - // should return nil. - require.Nil(newNamedBucket(tx, []byte("nonexistent"))) - - // Creating a named bucket from a bolt tx should work and set - // the path and name properly. - b, err := createNamedBucketIfNotExists(tx, parentBktName) - require.NoError(err) - require.Equal(parentBktName, b.Name()) - require.Equal(parentBktName, b.Path()) - - // Trying to descend into a nonexistent bucket should return - // nil. - require.Nil(b.Bucket([]byte("nonexistent"))) - - // Descending into a new bucket should update the path. - childBkt, err := b.CreateBucket(childBktName) - require.NoError(err) - require.Equal(childBktName, childBkt.(*namedBucket).Name()) - require.Equal([]byte("root/child"), childBkt.Path()) - - // Assert the parent bucket did not get changed. - require.Equal(parentBktName, b.Name()) - require.Equal(parentBktName, b.Path()) - - // Add entries to both buckets - require.NoError(b.Put(parentKey, parentVal)) - require.NoError(childBkt.Put(childKey, childVal)) - return nil - })) - - // Read buckets and values back out - require.NoError(db.View(func(tx *bolt.Tx) error { - b := newNamedBucket(tx, parentBktName) - require.NotNil(b) - require.Equal(parentVal, b.Get(parentKey)) - require.Nil(b.Get(childKey)) - - childBkt := b.Bucket(childBktName) - require.NotNil(childBkt) - require.Nil(childBkt.Get(parentKey)) - require.Equal(childVal, childBkt.Get(childKey)) - return nil - })) -} - -// TestNamedBucket_DeleteBucket asserts that deleting a bucket properly purges -// all related keys from the internal hashes map. -func TestNamedBucket_DeleteBucket(t *testing.T) { - t.Parallel() - require := require.New(t) - - db, cleanup := setupBoltDB(t) - defer cleanup() - - // Create some nested buckets and keys (key values will just be their names) - b1Name, c1Name, c2Name, c1c1Name := []byte("b1"), []byte("c1"), []byte("c2"), []byte("c1c1") - b1k1, c1k1, c2k1, c1c1k1 := []byte("b1k1"), []byte("c1k1"), []byte("c2k1"), []byte("c1c1k1") - - codec := newKeyValueCodec() - - // Create initial db state - require.NoError(db.Update(func(tx *bolt.Tx) error { - // Create bucket 1 and key - b1, err := createNamedBucketIfNotExists(tx, b1Name) - require.NoError(err) - require.NoError(codec.Put(b1, b1k1, b1k1)) - - // Create child bucket 1 and key - c1, err := b1.CreateBucketIfNotExists(c1Name) - require.NoError(err) - require.NoError(codec.Put(c1, c1k1, c1k1)) - - // Create child-child bucket 1 and key - c1c1, err := c1.CreateBucketIfNotExists(c1c1Name) - require.NoError(err) - require.NoError(codec.Put(c1c1, c1c1k1, c1c1k1)) - - // Create child bucket 2 and key - c2, err := b1.CreateBucketIfNotExists(c2Name) - require.NoError(err) - require.NoError(codec.Put(c2, c2k1, c2k1)) - return nil - })) - - // codec should be tracking 4 hash buckets (b1, c1, c2, c1c1) - require.Len(codec.hashes, 4) - - // Delete c1 - require.NoError(db.Update(func(tx *bolt.Tx) error { - b1 := newNamedBucket(tx, b1Name) - return codec.DeleteBucket(b1, c1Name) - })) - - START HERE // We don't appear to be properly deleting the sub-bucket - // codec should be tracking 2 hash buckets (b1, c2) - require.Len(codec.hashes, 2) - - // Assert all of c1 is gone - require.NoError(db.View(func(tx *bolt.Tx) error { - return nil - })) -} diff --git a/client/state/state_database.go b/client/state/state_database.go index 4e0e2d946..ff89f6dff 100644 --- a/client/state/state_database.go +++ b/client/state/state_database.go @@ -4,8 +4,8 @@ import ( "fmt" "path/filepath" - "github.com/boltdb/bolt" trstate "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + "github.com/hashicorp/nomad/helper/boltdd" "github.com/hashicorp/nomad/nomad/structs" ) @@ -57,20 +57,18 @@ func GetStateDBFactory(devMode bool) NewStateDBFunc { // methods are safe for concurrent access. Create via NewStateDB by setting // devMode=false. type BoltStateDB struct { - db *bolt.DB - codec *keyValueCodec + db *boltdd.DB } func NewBoltStateDB(stateDir string) (StateDB, error) { // Create or open the boltdb state database - db, err := bolt.Open(filepath.Join(stateDir, "state.db"), 0600, nil) + db, err := boltdd.Open(filepath.Join(stateDir, "state.db"), 0600, nil) if err != nil { return nil, fmt.Errorf("failed to create state database: %v", err) } sdb := &BoltStateDB{ - db: db, - codec: newKeyValueCodec(), + db: db, } return sdb, nil } @@ -83,7 +81,7 @@ func NewBoltStateDB(stateDir string) (StateDB, error) { func (s *BoltStateDB) GetAllAllocations() ([]*structs.Allocation, map[string]error, error) { var allocs []*structs.Allocation var errs map[string]error - err := s.db.View(func(tx *bolt.Tx) error { + err := s.db.View(func(tx *boltdd.Tx) error { allocs, errs = s.getAllAllocations(tx) return nil }) @@ -101,8 +99,8 @@ type allocEntry struct { Alloc *structs.Allocation } -func (s *BoltStateDB) getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map[string]error) { - allocationsBkt := newNamedBucket(tx, allocationsBucket) +func (s *BoltStateDB) getAllAllocations(tx *boltdd.Tx) ([]*structs.Allocation, map[string]error) { + allocationsBkt := tx.Bucket(allocationsBucket) if allocationsBkt == nil { // No allocs return nil, nil @@ -112,7 +110,7 @@ func (s *BoltStateDB) getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map errs := map[string]error{} // Create a cursor for iteration. - c := allocationsBkt.bkt.Cursor() + c := allocationsBkt.BoltBucket().Cursor() // Iterate over all the allocation buckets for k, _ := c.First(); k != nil; k, _ = c.Next() { @@ -124,7 +122,7 @@ func (s *BoltStateDB) getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map } var allocState allocEntry - if err := s.codec.Get(allocBkt, allocKey, &allocState); err != nil { + if err := allocBkt.Get(allocKey, &allocState); err != nil { errs[allocID] = fmt.Errorf("failed to decode alloc %v", err) continue } @@ -137,9 +135,9 @@ func (s *BoltStateDB) getAllAllocations(tx *bolt.Tx) ([]*structs.Allocation, map // PutAllocation stores an allocation or returns an error. func (s *BoltStateDB) PutAllocation(alloc *structs.Allocation) error { - return s.db.Update(func(tx *bolt.Tx) error { + return s.db.Update(func(tx *boltdd.Tx) error { // Retrieve the root allocations bucket - allocsBkt, err := createNamedBucketIfNotExists(tx, allocationsBucket) + allocsBkt, err := tx.CreateBucketIfNotExists(allocationsBucket) if err != nil { return err } @@ -154,7 +152,7 @@ func (s *BoltStateDB) PutAllocation(alloc *structs.Allocation) error { allocState := allocEntry{ Alloc: alloc, } - return s.codec.Put(allocBkt, allocKey, &allocState) + return allocBkt.Put(allocKey, &allocState) }) } @@ -164,7 +162,7 @@ func (s *BoltStateDB) GetTaskRunnerState(allocID, taskName string) (*trstate.Loc var ls trstate.LocalState var ts structs.TaskState - err := s.db.View(func(tx *bolt.Tx) error { + err := s.db.View(func(tx *boltdd.Tx) error { bkt, err := getTaskBucket(tx, allocID, taskName) if err != nil { return fmt.Errorf("failed to get task %q bucket: %v", taskName, err) @@ -172,12 +170,12 @@ func (s *BoltStateDB) GetTaskRunnerState(allocID, taskName string) (*trstate.Loc // Restore Local State //XXX set persisted hash to avoid immediate write on first use? - if err := s.codec.Get(bkt, taskLocalStateKey, &ls); err != nil { + if err := bkt.Get(taskLocalStateKey, &ls); err != nil { return fmt.Errorf("failed to read local task runner state: %v", err) } // Restore Task State - if err := s.codec.Get(bkt, taskStateKey, &ts); err != nil { + if err := bkt.Get(taskStateKey, &ts); err != nil { return fmt.Errorf("failed to read task state: %v", err) } @@ -199,13 +197,13 @@ func (s *BoltStateDB) GetTaskRunnerState(allocID, taskName string) (*trstate.Loc // PutTaskRunnerLocalState stores TaskRunner's LocalState or returns an error. // It is up to the caller to serialize the state to bytes. func (s *BoltStateDB) PutTaskRunnerLocalState(allocID, taskName string, val interface{}) error { - return s.db.Update(func(tx *bolt.Tx) error { + return s.db.Update(func(tx *boltdd.Tx) error { taskBkt, err := getTaskBucket(tx, allocID, taskName) if err != nil { return fmt.Errorf("failed to retrieve allocation bucket: %v", err) } - if err := s.codec.Put(taskBkt, taskLocalStateKey, val); err != nil { + if err := taskBkt.Put(taskLocalStateKey, val); err != nil { return fmt.Errorf("failed to write task_runner state: %v", err) } @@ -215,21 +213,21 @@ func (s *BoltStateDB) PutTaskRunnerLocalState(allocID, taskName string, val inte // PutTaskState stores a task's state or returns an error. func (s *BoltStateDB) PutTaskState(allocID, taskName string, state *structs.TaskState) error { - return s.db.Update(func(tx *bolt.Tx) error { + return s.db.Update(func(tx *boltdd.Tx) error { taskBkt, err := getTaskBucket(tx, allocID, taskName) if err != nil { return fmt.Errorf("failed to retrieve allocation bucket: %v", err) } - return s.codec.Put(taskBkt, taskStateKey, state) + return taskBkt.Put(taskStateKey, state) }) } // DeleteTaskBucket is used to delete a task bucket if it exists. func (s *BoltStateDB) DeleteTaskBucket(allocID, taskName string) error { - return s.db.Update(func(tx *bolt.Tx) error { + return s.db.Update(func(tx *boltdd.Tx) error { // Retrieve the root allocations bucket - allocations := newNamedBucket(tx, allocationsBucket) + allocations := tx.Bucket(allocationsBucket) if allocations == nil { return nil } @@ -242,21 +240,21 @@ func (s *BoltStateDB) DeleteTaskBucket(allocID, taskName string) error { // Check if the bucket exists key := []byte(taskName) - return s.codec.DeleteBucket(alloc, key) + return alloc.DeleteBucket(key) }) } // DeleteAllocationBucket is used to delete an allocation bucket if it exists. func (s *BoltStateDB) DeleteAllocationBucket(allocID string) error { - return s.db.Update(func(tx *bolt.Tx) error { + return s.db.Update(func(tx *boltdd.Tx) error { // Retrieve the root allocations bucket - allocations := newNamedBucket(tx, allocationsBucket) + allocations := tx.Bucket(allocationsBucket) if allocations == nil { return nil } key := []byte(allocID) - return s.codec.DeleteBucket(allocations, key) + return allocations.DeleteBucket(key) }) } @@ -270,18 +268,18 @@ func (s *BoltStateDB) Close() error { // particular allocation. If the root allocation bucket or the specific // allocation bucket doesn't exist, it will be created as long as the // transaction is writable. -func getAllocationBucket(tx *bolt.Tx, allocID string) (kvStore, error) { +func getAllocationBucket(tx *boltdd.Tx, allocID string) (*boltdd.Bucket, error) { var err error w := tx.Writable() // Retrieve the root allocations bucket - allocations := newNamedBucket(tx, allocationsBucket) + allocations := tx.Bucket(allocationsBucket) if allocations == nil { if !w { return nil, fmt.Errorf("Allocations bucket doesn't exist and transaction is not writable") } - allocations, err = createNamedBucketIfNotExists(tx, allocationsBucket) + allocations, err = tx.CreateBucketIfNotExists(allocationsBucket) if err != nil { return nil, err } @@ -308,7 +306,7 @@ func getAllocationBucket(tx *bolt.Tx, allocID string) (kvStore, error) { // particular task. If the root allocation bucket, the specific // allocation or task bucket doesn't exist, they will be created as long as the // transaction is writable. -func getTaskBucket(tx *bolt.Tx, allocID, taskName string) (kvStore, error) { +func getTaskBucket(tx *boltdd.Tx, allocID, taskName string) (*boltdd.Bucket, error) { alloc, err := getAllocationBucket(tx, allocID) if err != nil { return nil, err diff --git a/helper/boltdd/boltdd.go b/helper/boltdd/boltdd.go new file mode 100644 index 000000000..0a9ec9897 --- /dev/null +++ b/helper/boltdd/boltdd.go @@ -0,0 +1,360 @@ +// boltdd contains a wrapper around BoltDB to deduplicate writes and encode +// values using mgspack. (dd stands for DeDuplicate) +package boltdd + +import ( + "bytes" + "fmt" + "os" + "sync" + + "github.com/boltdb/bolt" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/ugorji/go/codec" + "golang.org/x/crypto/blake2b" +) + +// DB wraps an underlying bolt.DB to create write deduplicating buckets and +// msgpack encoded values. +type DB struct { + rootBuckets map[string]*bucketMeta + rootBucketsLock sync.Mutex + + bdb *bolt.DB +} + +// Open a bolt.DB and wrap it in a write-deduplicating msgpack-encoding +// implementation. +func Open(path string, mode os.FileMode, options *bolt.Options) (*DB, error) { + bdb, err := bolt.Open(path, mode, options) + if err != nil { + return nil, err + } + + return &DB{ + rootBuckets: make(map[string]*bucketMeta), + bdb: bdb, + }, nil +} + +func (db *DB) bucket(btx *bolt.Tx, name []byte) *Bucket { + bb := btx.Bucket(name) + if bb == nil { + return nil + } + + db.rootBucketsLock.Lock() + defer db.rootBucketsLock.Unlock() + + b, ok := db.rootBuckets[string(name)] + if !ok { + b = newBucketMeta() + db.rootBuckets[string(name)] = b + } + + return newBucket(b, bb) +} + +func (db *DB) createBucket(btx *bolt.Tx, name []byte) (*Bucket, error) { + bb, err := btx.CreateBucket(name) + if err != nil { + return nil, err + } + + db.rootBucketsLock.Lock() + defer db.rootBucketsLock.Unlock() + + // Always create a new Bucket since CreateBucket above fails if the + // bucket already exists. + b := newBucketMeta() + db.rootBuckets[string(name)] = b + + return newBucket(b, bb), nil +} + +func (db *DB) createBucketIfNotExists(btx *bolt.Tx, name []byte) (*Bucket, error) { + bb, err := btx.CreateBucketIfNotExists(name) + if err != nil { + return nil, err + } + + db.rootBucketsLock.Lock() + defer db.rootBucketsLock.Unlock() + + b, ok := db.rootBuckets[string(name)] + if !ok { + b = newBucketMeta() + db.rootBuckets[string(name)] = b + } + + return newBucket(b, bb), nil +} + +func (db *DB) Update(fn func(*Tx) error) error { + return db.bdb.Update(func(btx *bolt.Tx) error { + tx := newTx(db, btx) + return fn(tx) + }) +} + +func (db *DB) View(fn func(*Tx) error) error { + return db.bdb.View(func(btx *bolt.Tx) error { + tx := newTx(db, btx) + return fn(tx) + }) +} + +// Close closes the underlying bolt.DB and clears all bucket hashes. DB is +// unusable after closing. +func (db *DB) Close() error { + db.rootBuckets = nil + return db.bdb.Close() +} + +// BoltDB returns the underlying bolt.DB. +func (db *DB) BoltDB() *bolt.DB { + return db.bdb +} + +type Tx struct { + db *DB + btx *bolt.Tx +} + +func newTx(db *DB, btx *bolt.Tx) *Tx { + return &Tx{ + db: db, + btx: btx, + } +} + +// Bucket returns a root bucket or nil if it doesn't exist. +func (tx *Tx) Bucket(name []byte) *Bucket { + return tx.db.bucket(tx.btx, name) +} + +func (tx *Tx) CreateBucket(name []byte) (*Bucket, error) { + return tx.db.createBucket(tx.btx, name) +} + +// CreateBucketIfNotExists returns a root bucket or creates a new one if it +// doesn't already exist. +func (tx *Tx) CreateBucketIfNotExists(name []byte) (*Bucket, error) { + return tx.db.createBucketIfNotExists(tx.btx, name) +} + +// Writable wraps boltdb Tx.Writable. +func (tx *Tx) Writable() bool { + return tx.btx.Writable() +} + +// BoltTx returns the unerlying bolt.Tx +func (tx *Tx) BoltTx() *bolt.Tx { + return tx.btx +} + +// bucketMeta persists metadata -- such as key hashes and child buckets -- +// about boltdb Buckets across transactions. +type bucketMeta struct { + // hashes holds all of the value hashes for keys in this bucket + hashes map[string][]byte + hashesLock sync.Mutex + + // buckets holds all of the child buckets + buckets map[string]*bucketMeta + bucketsLock sync.Mutex +} + +func newBucketMeta() *bucketMeta { + return &bucketMeta{ + hashes: make(map[string][]byte), + buckets: make(map[string]*bucketMeta), + } +} + +// getHash of last value written to a key or nil if no hash exists. +func (bm *bucketMeta) getHash(hashKey string) []byte { + bm.hashesLock.Lock() + lastHash := bm.hashes[hashKey] + bm.hashesLock.Unlock() + return lastHash +} + +// setHash of last value written to key. +func (bm *bucketMeta) setHash(hashKey string, hashVal []byte) { + bm.hashesLock.Lock() + bm.hashes[hashKey] = hashVal + bm.hashesLock.Unlock() +} + +// delHash deletes a hash value or does nothing if the hash key does not exist. +func (bm *bucketMeta) delHash(hashKey string) { + bm.hashesLock.Lock() + delete(bm.hashes, hashKey) + bm.hashesLock.Unlock() +} + +// createBucket metadata entry for the given nested bucket. Overwrites any +// existing entry so caller should ensure bucket does not already exist. +func (bm *bucketMeta) createBucket(name []byte) *bucketMeta { + bm.bucketsLock.Lock() + defer bm.bucketsLock.Unlock() + + // Always create a new Bucket since CreateBucket above fails if the + // bucket already exists. + b := newBucketMeta() + bm.buckets[string(name)] = b + return b +} + +// deleteBucket metadata entry for the given nested bucket. Does nothing if +// nested bucket metadata does not exist. +func (bm *bucketMeta) deleteBucket(name []byte) { + bm.bucketsLock.Lock() + delete(bm.buckets, string(name)) + bm.bucketsLock.Unlock() + +} + +// getOrCreateBucket metadata entry for the given nested bucket. +func (bm *bucketMeta) getOrCreateBucket(name []byte) *bucketMeta { + bm.bucketsLock.Lock() + defer bm.bucketsLock.Unlock() + + b, ok := bm.buckets[string(name)] + if !ok { + b = newBucketMeta() + bm.buckets[string(name)] = b + } + return b +} + +type Bucket struct { + bm *bucketMeta + boltBucket *bolt.Bucket +} + +// newBucket creates a new view into a bucket backed by a boltdb +// transaction. +func newBucket(b *bucketMeta, bb *bolt.Bucket) *Bucket { + return &Bucket{ + bm: b, + boltBucket: bb, + } +} + +// Put into boltdb iff it has changed since the last write. +func (b *Bucket) Put(key []byte, val interface{}) error { + // buffer for writing serialized state to + var buf bytes.Buffer + + // Serialize the object + if err := codec.NewEncoder(&buf, structs.MsgpackHandle).Encode(val); err != nil { + return fmt.Errorf("failed to encode passed object: %v", err) + } + + // Hash for skipping unnecessary writes + hashKey := string(key) + hashVal := blake2b.Sum256(buf.Bytes()) + + // lastHash value or nil if it hasn't been hashed yet + lastHash := b.bm.getHash(hashKey) + + // If the hashes are equal, skip the write + if bytes.Equal(hashVal[:], lastHash) { + return nil + } + + // New value: write it to the underlying boltdb + if err := b.boltBucket.Put(key, buf.Bytes()); err != nil { + return fmt.Errorf("failed to write data at key %s: %v", key, err) + } + + // New value written, store hash (bucket path map was created above) + b.bm.setHash(hashKey, hashVal[:]) + + return nil + +} + +// Get value by key from boltdb or return an error if key not found. +func (b *Bucket) Get(key []byte, obj interface{}) error { + // Get the raw data from the underlying boltdb + data := b.boltBucket.Get(key) + if data == nil { + return fmt.Errorf("no data at key %v", string(key)) + } + + // Deserialize the object + if err := codec.NewDecoderBytes(data, structs.MsgpackHandle).Decode(obj); err != nil { + return fmt.Errorf("failed to decode data into passed object: %v", err) + } + + return nil +} + +// Delete removes a key from the bucket. If the key does not exist then nothing +// is done and a nil error is returned. Returns an error if the bucket was +// created from a read-only transaction. +func (b *Bucket) Delete(key []byte) error { + err := b.boltBucket.Delete(key) + b.bm.delHash(string(key)) + return err +} + +// Bucket represents a boltdb Bucket and its associated metadata necessary for +// write deduplication. Like bolt.Buckets it is only valid for the duration of +// the transaction that created it. +func (b *Bucket) Bucket(name []byte) *Bucket { + bb := b.boltBucket.Bucket(name) + if bb == nil { + return nil + } + + bmeta := b.bm.getOrCreateBucket(name) + return newBucket(bmeta, bb) +} + +// CreateBucket creates a new bucket at the given key and returns the new +// bucket. Returns an error if the key already exists, if the bucket name is +// blank, or if the bucket name is too long. The bucket instance is only valid +// for the lifetime of the transaction. +func (b *Bucket) CreateBucket(name []byte) (*Bucket, error) { + bb, err := b.boltBucket.CreateBucket(name) + if err != nil { + return nil, err + } + + bmeta := b.bm.createBucket(name) + return newBucket(bmeta, bb), nil +} + +// CreateBucketIfNotExists creates a new bucket if it doesn't already exist and +// returns a reference to it. The bucket instance is only valid for the +// lifetime of the transaction. +func (b *Bucket) CreateBucketIfNotExists(name []byte) (*Bucket, error) { + bb, err := b.boltBucket.CreateBucketIfNotExists(name) + if err != nil { + return nil, err + } + + bmeta := b.bm.getOrCreateBucket(name) + return newBucket(bmeta, bb), nil +} + +// DeleteBucket deletes a child bucket. Returns an error if the bucket does not +// exist or corresponds to a non-bucket key. +func (b *Bucket) DeleteBucket(name []byte) error { + // Delete the bucket from the underlying boltdb + err := b.boltBucket.DeleteBucket(name) + + // Remove reference to child bucket + b.bm.deleteBucket(name) + return err +} + +// BoltBucket returns the internal bolt.Bucket for this Bucket. Only valid +// for the duration of the current transaction. +func (b *Bucket) BoltBucket() *bolt.Bucket { + return b.boltBucket +} diff --git a/helper/boltdd/boltdd_test.go b/helper/boltdd/boltdd_test.go new file mode 100644 index 000000000..da3681ef6 --- /dev/null +++ b/helper/boltdd/boltdd_test.go @@ -0,0 +1,344 @@ +package boltdd + +import ( + "bytes" + "fmt" + "io/ioutil" + "os" + "path/filepath" + "testing" + + "github.com/hashicorp/nomad/nomad/mock" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/stretchr/testify/require" + "github.com/ugorji/go/codec" +) + +type testingT interface { + Fatalf(format string, args ...interface{}) + Logf(format string, args ...interface{}) +} + +func setupBoltDB(t testingT) (*DB, func()) { + dir, err := ioutil.TempDir("", "nomadtest_") + if err != nil { + t.Fatalf("error creating tempdir: %v", err) + } + + cleanup := func() { + if err := os.RemoveAll(dir); err != nil { + t.Logf("error removing test dir: %v", err) + } + } + + dbFilename := filepath.Join(dir, "nomadtest.db") + db, err := Open(dbFilename, 0600, nil) + if err != nil { + cleanup() + t.Fatalf("error creating boltdb: %v", err) + } + + return db, func() { + db.Close() + cleanup() + } +} + +func TestDB_Open(t *testing.T) { + t.Parallel() + require := require.New(t) + + db, cleanup := setupBoltDB(t) + defer cleanup() + + require.Equal(0, db.BoltDB().Stats().TxStats.Write) +} + +func TestBucket_Create(t *testing.T) { + t.Parallel() + require := require.New(t) + + db, cleanup := setupBoltDB(t) + defer cleanup() + + name := []byte("create_test") + + require.NoError(db.Update(func(tx *Tx) error { + // Trying to get a nonexistent bucket should return nil + require.Nil(tx.Bucket(name)) + + // Creating a nonexistent bucket should work + b, err := tx.CreateBucket(name) + require.NoError(err) + require.NotNil(b) + + // Recreating a bucket that exists should fail + b, err = tx.CreateBucket(name) + require.Error(err) + require.Nil(b) + + // get or create should work + b, err = tx.CreateBucketIfNotExists(name) + require.NoError(err) + require.NotNil(b) + return nil + })) + + // Bucket should be visible + require.NoError(db.View(func(tx *Tx) error { + require.NotNil(tx.Bucket(name)) + return nil + })) +} + +func TestBucket_DedupeWrites(t *testing.T) { + t.Parallel() + require := require.New(t) + + db, cleanup := setupBoltDB(t) + defer cleanup() + + bname := []byte("dedupewrites_test") + k1name := []byte("k1") + k2name := []byte("k2") + + // Put 2 keys + require.NoError(db.Update(func(tx *Tx) error { + b, err := tx.CreateBucket(bname) + require.NoError(err) + + require.NoError(b.Put(k1name, k1name)) + require.NoError(b.Put(k2name, k2name)) + return nil + })) + + // Assert there was at least 1 write + origWrites := db.BoltDB().Stats().TxStats.Write + require.NotZero(origWrites) + + // Write the same values again and expect no new writes + require.NoError(db.Update(func(tx *Tx) error { + b := tx.Bucket(bname) + require.NoError(b.Put(k1name, k1name)) + require.NoError(b.Put(k2name, k2name)) + return nil + })) + + putWrites := db.BoltDB().Stats().TxStats.Write + + // Unforunately every committed transaction causes two writes, so this + // only saves 1 write operation + require.Equal(origWrites+2, putWrites) + + // Write new values and assert more writes took place + require.NoError(db.Update(func(tx *Tx) error { + b := tx.Bucket(bname) + require.NoError(b.Put(k1name, []byte("newval1"))) + require.NoError(b.Put(k2name, []byte("newval2"))) + return nil + })) + + putWrites2 := db.BoltDB().Stats().TxStats.Write + + // Expect 3 additional writes: 2 for the transaction and one for the + // dirty page + require.Equal(putWrites+3, putWrites2) +} + +func TestBucket_Delete(t *testing.T) { + t.Parallel() + require := require.New(t) + + db, cleanup := setupBoltDB(t) + defer cleanup() + + parentName := []byte("delete_test") + parentKey := []byte("parent_key") + childName := []byte("child") + childKey := []byte("child_key") + grandchildName1 := []byte("grandchild1") + grandchildKey1 := []byte("grandchild_key1") + grandchildName2 := []byte("grandchild2") + grandchildKey2 := []byte("grandchild_key2") + + // Create a parent bucket with 1 child and 2 grandchildren + require.NoError(db.Update(func(tx *Tx) error { + pb, err := tx.CreateBucket(parentName) + require.NoError(err) + + require.NoError(pb.Put(parentKey, parentKey)) + + child, err := pb.CreateBucket(childName) + require.NoError(err) + + require.NoError(child.Put(childKey, childKey)) + + grandchild1, err := child.CreateBucket(grandchildName1) + require.NoError(err) + + require.NoError(grandchild1.Put(grandchildKey1, grandchildKey1)) + + grandchild2, err := child.CreateBucket(grandchildName2) + require.NoError(err) + + require.NoError(grandchild2.Put(grandchildKey2, grandchildKey2)) + return nil + })) + + // Verify grandchild keys wrote + require.NoError(db.View(func(tx *Tx) error { + grandchild1 := tx.Bucket(parentName).Bucket(childName).Bucket(grandchildName1) + var v1 []byte + grandchild1.Get(grandchildKey1, &v1) + require.Equal(grandchildKey1, v1) + + grandchild2 := tx.Bucket(parentName).Bucket(childName).Bucket(grandchildName2) + var v2 []byte + grandchild2.Get(grandchildKey2, &v2) + require.Equal(grandchildKey2, v2) + return nil + })) + + // Delete grandchildKey1 and grandchild2 + require.NoError(db.Update(func(tx *Tx) error { + child := tx.Bucket(parentName).Bucket(childName) + + require.NoError(child.DeleteBucket(grandchildName2)) + + grandchild1 := child.Bucket(grandchildName1) + require.NoError(grandchild1.Delete(grandchildKey1)) + return nil + })) + + // Ensure grandchild2 alone was deleted + require.NoError(db.View(func(tx *Tx) error { + grandchild1 := tx.Bucket(parentName).Bucket(childName).Bucket(grandchildName1) + var v1 []byte + grandchild1.Get(grandchildKey1, &v1) + require.Equal(([]byte)(nil), v1) + + grandchild2 := tx.Bucket(parentName).Bucket(childName).Bucket(grandchildName2) + require.Nil(grandchild2) + return nil + })) + + // Deleting child bucket should delete grandchild1 as well + require.NoError(db.Update(func(tx *Tx) error { + parent := tx.Bucket(parentName) + require.NoError(parent.DeleteBucket(childName)) + + // Recreate child bucket and ensure childKey and grandchild are gone + child, err := parent.CreateBucket(childName) + require.NoError(err) + + var v []byte + require.Error(child.Get(childKey, &v)) + require.Equal(([]byte)(nil), v) + + require.Nil(child.Bucket(grandchildName1)) + + // Rewrite childKey1 to make sure it doesn't get dedupe incorrectly + require.NoError(child.Put(childKey, childKey)) + return nil + })) + + // Ensure childKey1 was rewritten and not deduped incorrectly + require.NoError(db.View(func(tx *Tx) error { + var v []byte + require.NoError(tx.Bucket(parentName).Bucket(childName).Get(childKey, &v)) + require.Equal(childKey, v) + return nil + })) +} + +func BenchmarkWriteDeduplication_On(b *testing.B) { + db, cleanup := setupBoltDB(b) + defer cleanup() + + bucketName := []byte("allocations") + alloc := mock.Alloc() + allocID := []byte(alloc.ID) + + err := db.Update(func(tx *Tx) error { + allocs, err := tx.CreateBucket(bucketName) + if err != nil { + return err + } + + return allocs.Put(allocID, alloc) + }) + + if err != nil { + b.Fatalf("error setting up: %v", err) + } + + b.ResetTimer() + for i := 0; i < b.N; i++ { + err := db.Update(func(tx *Tx) error { + return tx.Bucket(bucketName).Put(allocID, alloc) + }) + + if err != nil { + b.Fatalf("error at runtime: %v", err) + } + } +} + +func BenchmarkWriteDeduplication_Off(b *testing.B) { + dir, err := ioutil.TempDir("", "nomadtest_") + if err != nil { + b.Fatalf("error creating tempdir: %v", err) + } + + defer func() { + if err := os.RemoveAll(dir); err != nil { + b.Logf("error removing test dir: %v", err) + } + }() + + dbFilename := filepath.Join(dir, "nomadtest.db") + db, err := Open(dbFilename, 0600, nil) + if err != nil { + b.Fatalf("error creating boltdb: %v", err) + } + + defer db.Close() + + bucketName := []byte("allocations") + alloc := mock.Alloc() + allocID := []byte(alloc.ID) + + err = db.Update(func(tx *Tx) error { + allocs, err := tx.CreateBucket(bucketName) + if err != nil { + return err + } + + var buf bytes.Buffer + if err := codec.NewEncoder(&buf, structs.MsgpackHandle).Encode(alloc); err != nil { + return fmt.Errorf("failed to encode passed object: %v", err) + } + + return allocs.Put(allocID, buf) + }) + + if err != nil { + b.Fatalf("error setting up: %v", err) + } + + b.ResetTimer() + for i := 0; i < b.N; i++ { + err := db.Update(func(tx *Tx) error { + var buf bytes.Buffer + if err := codec.NewEncoder(&buf, structs.MsgpackHandle).Encode(alloc); err != nil { + return fmt.Errorf("failed to encode passed object: %v", err) + } + + return tx.Bucket(bucketName).Put(allocID, buf) + }) + + if err != nil { + b.Fatalf("error at runtime: %v", err) + } + } +} From 357641c3642b2b4f52d567b56ddf4835d567ecbe Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Fri, 17 Aug 2018 10:34:44 -0700 Subject: [PATCH 048/152] persist alloc state on changes, not periodically Allow alloc and task runners to persist their own state when something changes instead of periodically syncing all state. --- client/allocrunnerv2/alloc_runner.go | 42 ++++++---------------------- client/client.go | 42 +++++++++++----------------- client/interfaces/client.go | 2 +- 3 files changed, 27 insertions(+), 59 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 68e1f90ee..b9cce6f12 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -177,16 +177,6 @@ MAIN: case <-taskWaitCh: // TaskRunners have all exited break MAIN - case updated := <-ar.updateCh: - // Update ar.alloc - ar.setAlloc(updated) - - //TODO Run AR Update hooks - - // Update task runners - for _, tr := range ar.tasks { - tr.Update(updated) - } } } @@ -230,18 +220,6 @@ func (ar *allocRunner) setAlloc(updated *structs.Allocation) { ar.allocLock.Unlock() } -// SaveState does all the state related stuff. Who knows. FIXME -//XXX do we need to do periodic syncing? if Saving is only called *before* Run -// *and* within Run -- *and* Updates are applid within Run -- we may be able to -// skip quite a bit of locking? maybe? -func (ar *allocRunner) SaveState() error { - //XXX Do we move this to the client - //XXX Track AllocModifyIndex to only write alloc on change? - - // Write the allocation - return ar.stateDB.PutAllocation(ar.Alloc()) -} - // Restore state from database. Must be called after NewAllocRunner but before // Run. func (ar *allocRunner) Restore() error { @@ -308,9 +286,7 @@ func (ar *allocRunner) TaskStateUpdated(taskName string, state *structs.TaskStat calloc := ar.clientAlloc(states) // Update the server - if err := ar.stateUpdater.AllocStateUpdated(calloc); err != nil { - ar.logger.Error("failed to update remote allocation state", "error", err) - } + ar.stateUpdater.AllocStateUpdated(calloc) } // clientAlloc takes in the task states and returns an Allocation populated @@ -408,14 +384,14 @@ func getClientStatus(taskStates map[string]*structs.TaskState) (status, descript // If there is already a pending update it will be discarded and replaced by // the latest update. func (ar *allocRunner) Update(update *structs.Allocation) { - select { - case ar.updateCh <- update: - // Updated alloc sent - case <-ar.updateCh: - // There was a pending update; replace it with the new update. - // This also prevents Update() from blocking if its called - // concurrently with Run() exiting. - ar.updateCh <- update + // Update ar.alloc + ar.setAlloc(update) + + //TODO Run AR Update hooks + + // Update task runners + for _, tr := range ar.tasks { + tr.Update(update) } } diff --git a/client/client.go b/client/client.go index 13b3ec415..7605f22e3 100644 --- a/client/client.go +++ b/client/client.go @@ -107,7 +107,6 @@ type AllocRunner interface { IsWaiting() bool IsMigrating() bool WaitCh() <-chan struct{} - SaveState() error Update(*structs.Allocation) Alloc() *structs.Allocation Restore() error @@ -547,7 +546,7 @@ func (c *Client) Shutdown() error { c.shutdown = true close(c.shutdownCh) c.connPool.Shutdown() - return c.saveState() + return nil } // Stats is used to return statistics for debugging and insight @@ -811,10 +810,6 @@ func (c *Client) restoreState() error { // saveState is used to snapshot our state into the data dir. func (c *Client) saveState() error { - if c.config.DevMode { - return nil - } - var wg sync.WaitGroup var l sync.Mutex var mErr multierror.Error @@ -823,7 +818,7 @@ func (c *Client) saveState() error { for id, ar := range runners { go func(id string, ar AllocRunner) { - err := ar.SaveState() + err := c.stateDB.PutAllocation(ar.Alloc()) if err != nil { c.logger.Printf("[ERR] client: failed to save state for alloc %q: %v", id, err) l.Lock() @@ -1506,18 +1501,9 @@ func (c *Client) updateNodeStatus() error { return nil } -// XXX combine with below -func (c *Client) AllocStateUpdated(alloc *structs.Allocation) error { - if alloc == nil { - return fmt.Errorf("nil allocation provided to AllocStateUpdated") - } - - c.updateAllocStatus(alloc) - return nil -} - -// updateAllocStatus is used to update the status of an allocation -func (c *Client) updateAllocStatus(alloc *structs.Allocation) { +// AllocStateUpdated asynchronously updates the server with the current state +// of an allocations and its tasks. +func (c *Client) AllocStateUpdated(alloc *structs.Allocation) { if alloc.Terminated() { // Terminated, mark for GC if we're still tracking this alloc // runner. If it's not being tracked that means the server has @@ -1901,6 +1887,12 @@ func (c *Client) updateAlloc(update *structs.Allocation) { return } + // Update local copy of alloc + if err := c.stateDB.PutAllocation(update); err != nil { + c.logger.Printf("[ERR] client: dropping alloc update. Unable to persist locally: %q", update.ID) + } + + // Update alloc runner ar.Update(update) } @@ -1915,6 +1907,12 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error return nil } + // Initialize local copy of alloc before creating the alloc runner so + // we can't end up with an alloc runner that does not have an alloc. + if err := c.stateDB.PutAllocation(alloc); err != nil { + return err + } + //FIXME disabled previous alloc waiting/migrating // get the previous alloc runner - if one exists - for the // blocking/migrating watcher @@ -1931,7 +1929,6 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error // Copy the config since the node can be swapped out as it is being updated. // The long term fix is to pass in the config and node separately and then // we don't have to do a copy. - //ar := allocrunner.NewAllocRunner(c.logger, c.configCopy.Copy(), c.stateDB, c.updateAllocStatus, alloc, c.vaultClient, c.consulService, prevAlloc) //XXX FIXME create a root logger logger := hclog.New(&hclog.LoggerOptions{ Name: "nomad", @@ -1959,11 +1956,6 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error // Store the alloc runner. c.allocs[alloc.ID] = ar - // Initialize local state - if err := ar.SaveState(); err != nil { - c.logger.Printf("[WARN] client: initial save state for alloc %q failed: %v", alloc.ID, err) - } - go ar.Run() return nil } diff --git a/client/interfaces/client.go b/client/interfaces/client.go index 6525df693..f79db6620 100644 --- a/client/interfaces/client.go +++ b/client/interfaces/client.go @@ -10,5 +10,5 @@ type Client interface { type AllocStateHandler interface { // AllocStateUpdated is used to emit an updated allocation. This allocation // is stripped to only include client settable fields. - AllocStateUpdated(alloc *structs.Allocation) error + AllocStateUpdated(alloc *structs.Allocation) } From 5749ede04e6d82859be9409557877b87d93065fa Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Fri, 17 Aug 2018 14:06:43 -0700 Subject: [PATCH 049/152] keep forgetting lxc --- client/driver/lxc.go | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/client/driver/lxc.go b/client/driver/lxc.go index 5f724f98b..ba46c829e 100644 --- a/client/driver/lxc.go +++ b/client/driver/lxc.go @@ -470,6 +470,10 @@ func (h *lxcDriverHandle) Signal(s os.Signal) error { return fmt.Errorf("LXC does not support signals") } +func (d *lxcDriverHandle) Network() *cstructs.DriverNetwork { + return nil +} + func (h *lxcDriverHandle) Stats() (*cstructs.TaskResourceUsage, error) { cpuStats, err := h.container.CPUStats() if err != nil { From 4236255686e4c6b05b04702018c8226cba120d0e Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 22 Aug 2018 10:18:49 -0700 Subject: [PATCH 050/152] lots of comment/log fixes --- client/client.go | 2 +- client/state/state_database.go | 20 +++++++------------- helper/boltdd/boltdd.go | 2 +- 3 files changed, 9 insertions(+), 15 deletions(-) diff --git a/client/client.go b/client/client.go index 7605f22e3..9bbb8fb65 100644 --- a/client/client.go +++ b/client/client.go @@ -1889,7 +1889,7 @@ func (c *Client) updateAlloc(update *structs.Allocation) { // Update local copy of alloc if err := c.stateDB.PutAllocation(update); err != nil { - c.logger.Printf("[ERR] client: dropping alloc update. Unable to persist locally: %q", update.ID) + c.logger.Printf("[ERR] client: error persisting updated alloc locally: %q", update.ID) } // Update alloc runner diff --git a/client/state/state_database.go b/client/state/state_database.go index ff89f6dff..50f811209 100644 --- a/client/state/state_database.go +++ b/client/state/state_database.go @@ -54,12 +54,13 @@ func GetStateDBFactory(devMode bool) NewStateDBFunc { } // BoltStateDB persists and restores Nomad client state in a boltdb. All -// methods are safe for concurrent access. Create via NewStateDB by setting -// devMode=false. +// methods are safe for concurrent access. type BoltStateDB struct { db *boltdd.DB } +// NewBoltStateDB creates or opens an existing boltdb state file or returns an +// error. func NewBoltStateDB(stateDir string) (StateDB, error) { // Create or open the boltdb state database db, err := boltdd.Open(filepath.Join(stateDir, "state.db"), 0600, nil) @@ -121,13 +122,13 @@ func (s *BoltStateDB) getAllAllocations(tx *boltdd.Tx) ([]*structs.Allocation, m continue } - var allocState allocEntry - if err := allocBkt.Get(allocKey, &allocState); err != nil { - errs[allocID] = fmt.Errorf("failed to decode alloc %v", err) + var ae allocEntry + if err := allocBkt.Get(allocKey, &ae); err != nil { + errs[allocID] = fmt.Errorf("failed to decode alloc: %v", err) continue } - allocs = append(allocs, allocState.Alloc) + allocs = append(allocs, ae.Alloc) } return allocs, errs @@ -157,7 +158,6 @@ func (s *BoltStateDB) PutAllocation(alloc *structs.Allocation) error { } // GetTaskRunnerState restores TaskRunner specific state. -//TODO wrap in a single struct? func (s *BoltStateDB) GetTaskRunnerState(allocID, taskName string) (*trstate.LocalState, *structs.TaskState, error) { var ls trstate.LocalState var ts structs.TaskState @@ -169,7 +169,6 @@ func (s *BoltStateDB) GetTaskRunnerState(allocID, taskName string) (*trstate.Loc } // Restore Local State - //XXX set persisted hash to avoid immediate write on first use? if err := bkt.Get(taskLocalStateKey, &ls); err != nil { return fmt.Errorf("failed to read local task runner state: %v", err) } @@ -179,10 +178,6 @@ func (s *BoltStateDB) GetTaskRunnerState(allocID, taskName string) (*trstate.Loc return fmt.Errorf("failed to read task state: %v", err) } - // XXX if driver has task { - // tr.restoreDriver() - // } - return nil }) @@ -195,7 +190,6 @@ func (s *BoltStateDB) GetTaskRunnerState(allocID, taskName string) (*trstate.Loc } // PutTaskRunnerLocalState stores TaskRunner's LocalState or returns an error. -// It is up to the caller to serialize the state to bytes. func (s *BoltStateDB) PutTaskRunnerLocalState(allocID, taskName string, val interface{}) error { return s.db.Update(func(tx *boltdd.Tx) error { taskBkt, err := getTaskBucket(tx, allocID, taskName) diff --git a/helper/boltdd/boltdd.go b/helper/boltdd/boltdd.go index 0a9ec9897..1fa31870b 100644 --- a/helper/boltdd/boltdd.go +++ b/helper/boltdd/boltdd.go @@ -148,7 +148,7 @@ func (tx *Tx) Writable() bool { return tx.btx.Writable() } -// BoltTx returns the unerlying bolt.Tx +// BoltTx returns the underlying bolt.Tx. func (tx *Tx) BoltTx() *bolt.Tx { return tx.btx } From e42154fc46cacbe51bd1cf7f7e2b9a2cdaa720be Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 23 Aug 2018 12:03:17 -0700 Subject: [PATCH 051/152] implement stopping, destroying, and disk migration * Stopping an alloc is implemented via Updates but update hooks are *not* run. * Destroying an alloc is a best effort cleanup. * AllocRunner destroy hooks implemented. * Disk migration and blocking on a previous allocation exiting moved to its own package to avoid cycles. Now only depends on alloc broadcaster instead of also using a waitch. * AllocBroadcaster now only drops stale allocations and always keeps the latest version. * Made AllocDir safe for concurrent use Lots of internal contexts that are currently unused. Unsure if they should be used or removed. --- client/allocdir/alloc_dir.go | 30 ++- client/allocrunner/alloc_runner.go | 45 ++--- client/allocrunner/testing.go | 3 +- client/allocrunnerv2/alloc_runner.go | 109 ++++++++--- client/allocrunnerv2/alloc_runner_hooks.go | 116 +++++++++-- client/allocrunnerv2/config.go | 5 + .../interfaces/runner_lifecycle.go | 4 +- client/allocrunnerv2/taskrunner/lifecycle.go | 2 + .../allocrunnerv2/taskrunner/task_runner.go | 14 +- .../alloc_watcher.go | 32 ++- .../alloc_watcher_test.go | 182 ++++++++++++++---- client/allocwatcher/doc.go | 4 + client/client.go | 34 ++-- client/state/interface.go | 2 + client/state/noopdb.go | 8 + client/structs/funcs.go | 71 +++++-- 16 files changed, 492 insertions(+), 169 deletions(-) rename client/{allocrunner => allocwatcher}/alloc_watcher.go (95%) rename client/{allocrunner => allocwatcher}/alloc_watcher_test.go (58%) create mode 100644 client/allocwatcher/doc.go diff --git a/client/allocdir/alloc_dir.go b/client/allocdir/alloc_dir.go index 1c31a68b4..65831af66 100644 --- a/client/allocdir/alloc_dir.go +++ b/client/allocdir/alloc_dir.go @@ -9,6 +9,7 @@ import ( "log" "os" "path/filepath" + "sync" "time" "github.com/hashicorp/go-multierror" @@ -58,6 +59,8 @@ var ( TaskDirs = map[string]os.FileMode{TmpDirName: os.ModeSticky | 0777} ) +// AllocDir allows creating, destroying, and accessing an allocation's +// directory. All methods are safe for concurrent use. type AllocDir struct { // AllocDir is the directory used for storing any state // of this allocation. It will be purged on alloc destroy. @@ -73,6 +76,8 @@ type AllocDir struct { // built is true if Build has successfully run built bool + mu sync.RWMutex + logger *log.Logger } @@ -100,6 +105,9 @@ func NewAllocDir(logger *log.Logger, allocDir string) *AllocDir { // Copy an AllocDir and all of its TaskDirs. Returns nil if AllocDir is // nil. func (d *AllocDir) Copy() *AllocDir { + d.mu.RLock() + defer d.mu.RUnlock() + if d == nil { return nil } @@ -117,6 +125,9 @@ func (d *AllocDir) Copy() *AllocDir { // NewTaskDir creates a new TaskDir and adds it to the AllocDirs TaskDirs map. func (d *AllocDir) NewTaskDir(name string) *TaskDir { + d.mu.Lock() + defer d.mu.Unlock() + td := newTaskDir(d.logger, d.AllocDir, name) d.TaskDirs[name] = td return td @@ -129,6 +140,9 @@ func (d *AllocDir) NewTaskDir(name string) *TaskDir { // file "NOMAD-${ALLOC_ID}-ERROR.log" will be appended to the tar with the // error message as the contents. func (d *AllocDir) Snapshot(w io.Writer) error { + d.mu.RLock() + defer d.mu.RUnlock() + allocDataDir := filepath.Join(d.SharedDir, SharedDataDir) rootPaths := []string{allocDataDir} for _, taskdir := range d.TaskDirs { @@ -206,11 +220,16 @@ func (d *AllocDir) Snapshot(w io.Writer) error { // Move other alloc directory's shared path and local dir to this alloc dir. func (d *AllocDir) Move(other *AllocDir, tasks []*structs.Task) error { + d.mu.RLock() if !d.built { // Enforce the invariant that Build is called before Move + d.mu.RUnlock() return fmt.Errorf("unable to move to %q - alloc dir is not built", d.AllocDir) } + // Moving is slow and only reads immutable fields, so unlock during heavy IO + d.mu.RUnlock() + // Move the data directory otherDataDir := filepath.Join(other.SharedDir, SharedDataDir) dataDir := filepath.Join(d.SharedDir, SharedDataDir) @@ -246,7 +265,6 @@ func (d *AllocDir) Move(other *AllocDir, tasks []*structs.Task) error { // Tears down previously build directory structure. func (d *AllocDir) Destroy() error { - // Unmount all mounted shared alloc dirs. var mErr multierror.Error if err := d.UnmountAll(); err != nil { @@ -258,12 +276,17 @@ func (d *AllocDir) Destroy() error { } // Unset built since the alloc dir has been destroyed. + d.mu.Lock() d.built = false + d.mu.Unlock() return mErr.ErrorOrNil() } // UnmountAll linked/mounted directories in task dirs. func (d *AllocDir) UnmountAll() error { + d.mu.RLock() + defer d.mu.RUnlock() + var mErr multierror.Error for _, dir := range d.TaskDirs { // Check if the directory has the shared alloc mounted. @@ -322,7 +345,9 @@ func (d *AllocDir) Build() error { } // Mark as built + d.mu.Lock() d.built = true + d.mu.Unlock() return nil } @@ -386,11 +411,14 @@ func (d *AllocDir) ReadAt(path string, offset int64) (io.ReadCloser, error) { p := filepath.Join(d.AllocDir, path) // Check if it is trying to read into a secret directory + d.mu.RLock() for _, dir := range d.TaskDirs { if filepath.HasPrefix(p, dir.SecretsDir) { + d.mu.RUnlock() return nil, fmt.Errorf("Reading secret file prohibited: %s", path) } } + d.mu.RUnlock() f, err := os.Open(p) if err != nil { diff --git a/client/allocrunner/alloc_runner.go b/client/allocrunner/alloc_runner.go index b60432332..cee65299d 100644 --- a/client/allocrunner/alloc_runner.go +++ b/client/allocrunner/alloc_runner.go @@ -13,6 +13,7 @@ import ( "github.com/hashicorp/go-multierror" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner/taskrunner" + "github.com/hashicorp/nomad/client/allocwatcher" "github.com/hashicorp/nomad/client/config" consulApi "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/vaultclient" @@ -77,7 +78,7 @@ type AllocRunner struct { // the migrates it data. If sticky volumes aren't used and there's no // previous allocation a noop implementation is used so it always safe // to call. - prevAlloc prevAllocWatcher + prevAlloc allocwatcher.PrevAllocWatcher // ctx is cancelled with exitFn to cause the alloc to be destroyed // (stopped and GC'd). @@ -133,26 +134,26 @@ type allocRunnerMutableState struct { // NewAllocRunner is used to create a new allocation context func NewAllocRunner(logger *log.Logger, config *config.Config, stateDB *bolt.DB, updater AllocStateUpdater, alloc *structs.Allocation, vaultClient vaultclient.VaultClient, consulClient consulApi.ConsulServiceAPI, - prevAlloc prevAllocWatcher) *AllocRunner { + prevAlloc allocwatcher.PrevAllocWatcher) *AllocRunner { ar := &AllocRunner{ - config: config, - stateDB: stateDB, - updater: updater, - logger: logger, - alloc: alloc, - allocID: alloc.ID, - allocBroadcast: cstructs.NewAllocBroadcaster(8), - prevAlloc: prevAlloc, - dirtyCh: make(chan struct{}, 1), - allocDir: allocdir.NewAllocDir(logger, filepath.Join(config.AllocDir, alloc.ID)), - tasks: make(map[string]*taskrunner.TaskRunner), - taskStates: copyTaskStates(alloc.TaskStates), - restored: make(map[string]struct{}), - updateCh: make(chan *structs.Allocation, 64), - waitCh: make(chan struct{}), - vaultClient: vaultClient, - consulClient: consulClient, + config: config, + stateDB: stateDB, + updater: updater, + logger: logger, + alloc: alloc, + allocID: alloc.ID, + //allocBroadcast: cstructs.NewAllocBroadcaster(8), + prevAlloc: prevAlloc, + dirtyCh: make(chan struct{}, 1), + allocDir: allocdir.NewAllocDir(logger, filepath.Join(config.AllocDir, alloc.ID)), + tasks: make(map[string]*taskrunner.TaskRunner), + taskStates: copyTaskStates(alloc.TaskStates), + restored: make(map[string]struct{}), + updateCh: make(chan *structs.Allocation, 64), + waitCh: make(chan struct{}), + vaultClient: vaultClient, + consulClient: consulClient, } // TODO Should be passed a context @@ -612,9 +613,9 @@ func (r *AllocRunner) sendBroadcast(alloc *structs.Allocation) { // Try to send the alloc up to three times with a delay to allow recovery. sent := false for i := 0; i < 3; i++ { - if sent = r.allocBroadcast.Send(alloc); sent { - break - } + //if sent = r.allocBroadcast.Send(alloc); sent { + // break + //} time.Sleep(500 * time.Millisecond) } if !sent { diff --git a/client/allocrunner/testing.go b/client/allocrunner/testing.go index f9e8152bc..4c4c1a3ff 100644 --- a/client/allocrunner/testing.go +++ b/client/allocrunner/testing.go @@ -7,6 +7,7 @@ import ( "testing" "github.com/boltdb/bolt" + "github.com/hashicorp/nomad/client/allocwatcher" "github.com/hashicorp/nomad/client/config" consulApi "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/vaultclient" @@ -51,7 +52,7 @@ func TestAllocRunnerFromAlloc(t *testing.T, alloc *structs.Allocation, restarts alloc.Job.Type = structs.JobTypeBatch } vclient := vaultclient.NewMockVaultClient() - ar := NewAllocRunner(testlog.Logger(t), conf, db, upd.Update, alloc, vclient, consulApi.NewMockConsulServiceClient(t, testlog.HCLogger(t)), NoopPrevAlloc{}) + ar := NewAllocRunner(testlog.Logger(t), conf, db, upd.Update, alloc, vclient, consulApi.NewMockConsulServiceClient(t), allocwatcher.NoopPrevAlloc{}) return upd, ar } diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index b9cce6f12..4172fcd52 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -13,6 +13,7 @@ import ( "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" "github.com/hashicorp/nomad/client/allocrunnerv2/state" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" + "github.com/hashicorp/nomad/client/allocwatcher" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/consul" cinterfaces "github.com/hashicorp/nomad/client/interfaces" @@ -51,10 +52,14 @@ type allocRunner struct { // vaultClient is the used to manage Vault tokens vaultClient vaultclient.VaultClient - // waitCh is closed when the alloc runner has transitioned to a terminal - // state + // waitCh is closed when the Run() loop has exited waitCh chan struct{} + // destroyed is true when the Run() loop has exited, postrun hooks have + // run, and alloc runner has been destroyed + destroyed bool + destroyedLock sync.Mutex + // Alloc captures the allocation being run. alloc *structs.Allocation allocLock sync.RWMutex @@ -81,6 +86,13 @@ type allocRunner struct { // have buffer size 1 in order to support dropping pending updates when // a newer allocation is received. updateCh chan *structs.Allocation + + // allocBroadcaster sends client allocation updates to all listeners + allocBroadcaster *cstructs.AllocBroadcaster + + // prevAllocWatcher allows waiting for a previous allocation to exit + // and if necessary migrate its alloc dir. + prevAllocWatcher allocwatcher.PrevAllocWatcher } // NewAllocRunner returns a new allocation runner. @@ -92,17 +104,19 @@ func NewAllocRunner(config *Config) (*allocRunner, error) { } ar := &allocRunner{ - id: alloc.ID, - alloc: alloc, - clientConfig: config.ClientConfig, - consulClient: config.Consul, - vaultClient: config.Vault, - tasks: make(map[string]*taskrunner.TaskRunner, len(tg.Tasks)), - waitCh: make(chan struct{}), - updateCh: make(chan *structs.Allocation, updateChCap), - state: &state.State{}, - stateDB: config.StateDB, - stateUpdater: config.StateUpdater, + id: alloc.ID, + alloc: alloc, + clientConfig: config.ClientConfig, + consulClient: config.Consul, + vaultClient: config.Vault, + tasks: make(map[string]*taskrunner.TaskRunner, len(tg.Tasks)), + waitCh: make(chan struct{}), + updateCh: make(chan *structs.Allocation, updateChCap), + state: &state.State{}, + stateDB: config.StateDB, + stateUpdater: config.StateUpdater, + allocBroadcaster: cstructs.NewAllocBroadcaster(alloc), + prevAllocWatcher: config.PrevAllocWatcher, } // Create alloc dir @@ -154,7 +168,7 @@ func (ar *allocRunner) WaitCh() <-chan struct{} { } // XXX How does alloc Restart work -// Run is the main go-routine that executes all the tasks. +// Run is the main goroutine that executes all the tasks. func (ar *allocRunner) Run() { // Close the wait channel defer close(ar.waitCh) @@ -220,6 +234,11 @@ func (ar *allocRunner) setAlloc(updated *structs.Allocation) { ar.allocLock.Unlock() } +// GetAllocDir returns the alloc dir which is safe for concurrent use. +func (ar *allocRunner) GetAllocDir() *allocdir.AllocDir { + return ar.allocDir +} + // Restore state from database. Must be called after NewAllocRunner but before // Run. func (ar *allocRunner) Restore() error { @@ -287,6 +306,9 @@ func (ar *allocRunner) TaskStateUpdated(taskName string, state *structs.TaskStat // Update the server ar.stateUpdater.AllocStateUpdated(calloc) + + // Broadcast client alloc to listeners + ar.allocBroadcaster.Send(calloc) } // clientAlloc takes in the task states and returns an Allocation populated @@ -395,19 +417,45 @@ func (ar *allocRunner) Update(update *structs.Allocation) { } } -// Destroy the alloc runner by stopping it if it is still running and cleaning -// up all of its resources. -// -// This method is safe for calling concurrently with Run(). Callers must -// receive on WaitCh() to block until alloc runner has stopped and been -// destroyed. -//XXX TODO -func (ar *allocRunner) Destroy() { - //TODO +func (ar *allocRunner) Listener() *cstructs.AllocListener { + return ar.allocBroadcaster.Listen() +} - for _, tr := range ar.tasks { - tr.Kill(context.Background(), structs.NewTaskEvent(structs.TaskKilled)) +// Destroy the alloc runner by synchronously stopping it if it is still running +// and cleaning up all of its resources. +// +// This method is safe for calling concurrently with Run() and will cause it to +// exit (thus closing WaitCh). +func (ar *allocRunner) Destroy() { + // Stop tasks + for name, tr := range ar.tasks { + err := tr.Kill(context.TODO(), structs.NewTaskEvent(structs.TaskKilled)) + if err != nil { + if err == taskrunner.ErrTaskNotRunning { + ar.logger.Trace("task not running", "task_name", name) + } else { + ar.logger.Warn("failed to kill task", "error", err, "task_name", name) + } + } } + + // Wait for tasks to exit and postrun hooks to finish + <-ar.waitCh + + // Run destroy hooks + if err := ar.destroy(); err != nil { + ar.logger.Warn("error running destroy hooks", "error", err) + } + + // Cleanup state db + if err := ar.stateDB.DeleteAllocationBucket(ar.id); err != nil { + ar.logger.Warn("failed to delete allocation state", "error", err) + } + + // Mark alloc as destroyed + ar.destroyedLock.Lock() + ar.destroyed = true + ar.destroyedLock.Unlock() } // IsDestroyed returns true if the alloc runner has been destroyed (stopped and @@ -416,27 +464,26 @@ func (ar *allocRunner) Destroy() { // This method is safe for calling concurrently with Run(). Callers must // receive on WaitCh() to block until alloc runner has stopped and been // destroyed. -//XXX TODO func (ar *allocRunner) IsDestroyed() bool { - return false + ar.destroyedLock.Lock() + defer ar.destroyedLock.Unlock() + return ar.destroyed } // IsWaiting returns true if the alloc runner is waiting for its previous // allocation to terminate. // // This method is safe for calling concurrently with Run(). -//XXX TODO func (ar *allocRunner) IsWaiting() bool { - return false + return ar.prevAllocWatcher.IsWaiting() } // IsMigrating returns true if the alloc runner is migrating data from its // previous allocation. // // This method is safe for calling concurrently with Run(). -//XXX TODO func (ar *allocRunner) IsMigrating() bool { - return false + return ar.prevAllocWatcher.IsMigrating() } // StatsReporter needs implementing diff --git a/client/allocrunnerv2/alloc_runner_hooks.go b/client/allocrunnerv2/alloc_runner_hooks.go index a3cac16f0..4ac7b9ab8 100644 --- a/client/allocrunnerv2/alloc_runner_hooks.go +++ b/client/allocrunnerv2/alloc_runner_hooks.go @@ -6,17 +6,22 @@ import ( "time" log "github.com/hashicorp/go-hclog" + multierror "github.com/hashicorp/go-multierror" + "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocwatcher" ) // initRunnerHooks intializes the runners hooks. func (ar *allocRunner) initRunnerHooks() { hookLogger := ar.logger.Named("runner_hook") - ar.runnerHooks = make([]interfaces.RunnerHook, 0, 3) // Create the alloc directory hook. This is run first to ensure the // directoy path exists for other hooks. - ar.runnerHooks = append(ar.runnerHooks, newAllocDirHook(ar, hookLogger)) + ar.runnerHooks = []interfaces.RunnerHook{ + newAllocDirHook(hookLogger, ar), + newDiskMigrationHook(hookLogger, ar.prevAllocWatcher, ar.allocDir), + } } // prerun is used to run the runners prerun hooks. @@ -36,17 +41,21 @@ func (ar *allocRunner) prerun() error { continue } + //TODO Check hook state + name := pre.Name() var start time.Time if ar.logger.IsTrace() { start = time.Now() - ar.logger.Trace("running prestart hook", "name", name, "start", start) + ar.logger.Trace("running pre-run hook", "name", name, "start", start) } - if err := pre.Prerun(); err != nil { + if err := pre.Prerun(context.TODO()); err != nil { return fmt.Errorf("hook %q failed: %v", name, err) } + //TODO Persist hook state locally + if ar.logger.IsTrace() { end := time.Now() ar.logger.Trace("finished pre-run hooks", "name", name, "end", end, "duration", end.Sub(start)) @@ -93,28 +102,53 @@ func (ar *allocRunner) postrun() error { return nil } -/* -What state is needed to transfer: -*/ +// destroy is used to run the runners destroy hooks. All hooks are run and +// errors are returned as a multierror. +func (ar *allocRunner) destroy() error { + if ar.logger.IsTrace() { + start := time.Now() + ar.logger.Trace("running destroy hooks", "start", start) + defer func() { + end := time.Now() + ar.logger.Trace("finished destroy hooks", "end", end, "duration", end.Sub(start)) + }() + } -/* -AR Hooks: -Alloc Dir Build: -Needs to know the folder to create + var merr multierror.Error + for _, hook := range ar.runnerHooks { + h, ok := hook.(interfaces.RunnerDestroyHook) + if !ok { + continue + } -Alloc Migrate -Needs access to RPC + name := h.Name() + var start time.Time + if ar.logger.IsTrace() { + start = time.Now() + ar.logger.Trace("running destroy hook", "name", name, "start", start) + } -Alloc Health Watcher: -Requires: Access to consul to watch health, access to every task event, task status change -*/ + if err := h.Destroy(); err != nil { + merr.Errors = append(merr.Errors, fmt.Errorf("destroy hook %q failed: %v", name, err)) + } + if ar.logger.IsTrace() { + end := time.Now() + ar.logger.Trace("finished destroy hooks", "name", name, "end", end, "duration", end.Sub(start)) + } + } + + return nil +} + +// allocDirHook creates and destroys the root directory and shared directories +// for an allocation. type allocDirHook struct { runner *allocRunner logger log.Logger } -func newAllocDirHook(runner *allocRunner, logger log.Logger) *allocDirHook { +func newAllocDirHook(logger log.Logger, runner *allocRunner) *allocDirHook { ad := &allocDirHook{ runner: runner, } @@ -126,7 +160,7 @@ func (h *allocDirHook) Name() string { return "alloc_dir" } -func (h *allocDirHook) Prerun() error { +func (h *allocDirHook) Prerun(context.Context) error { return h.runner.allocDir.Build() } @@ -134,6 +168,52 @@ func (h *allocDirHook) Destroy() error { return h.runner.allocDir.Destroy() } +// diskMigrationHook migrates ephemeral disk volumes. Depends on alloc dir +// being built but must be run before anything else manipulates the alloc dir. +type diskMigrationHook struct { + allocDir *allocdir.AllocDir + allocWatcher allocwatcher.PrevAllocWatcher + logger log.Logger +} + +func newDiskMigrationHook(logger log.Logger, allocWatcher allocwatcher.PrevAllocWatcher, allocDir *allocdir.AllocDir) *diskMigrationHook { + h := &diskMigrationHook{ + allocDir: allocDir, + allocWatcher: allocWatcher, + } + h.logger = logger.Named(h.Name()) + return h +} + +func (h *diskMigrationHook) Name() string { + return "migrate_disk" +} + +func (h *diskMigrationHook) Prerun(ctx context.Context) error { + // Wait for a previous alloc - if any - to terminate + if err := h.allocWatcher.Wait(ctx); err != nil { + return err + } + + // Wait for data to be migrated from a previous alloc if applicable + if err := h.allocWatcher.Migrate(ctx, h.allocDir); err != nil { + if err == context.Canceled { + return err + } + + // Soft-fail on migration errors + h.logger.Warn("error migrating data from previous alloc", "error", err) + + // Recreate alloc dir to ensure a clean slate + h.allocDir.Destroy() + if err := h.allocDir.Build(); err != nil { + return fmt.Errorf("failed to clean task directories after failed migration: %v", err) + } + } + + return nil +} + // TODO type allocHealthWatcherHook struct { runner *allocRunner diff --git a/client/allocrunnerv2/config.go b/client/allocrunnerv2/config.go index 6d5865299..50b8a672a 100644 --- a/client/allocrunnerv2/config.go +++ b/client/allocrunnerv2/config.go @@ -2,6 +2,7 @@ package allocrunnerv2 import ( log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocwatcher" clientconfig "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/interfaces" @@ -32,4 +33,8 @@ type Config struct { // StateUpdater is used to emit updated task state StateUpdater interfaces.AllocStateHandler + + // PrevAllocWatcher handles waiting on previous allocations and + // migrating their ephemeral disk when necessary. + PrevAllocWatcher allocwatcher.PrevAllocWatcher } diff --git a/client/allocrunnerv2/interfaces/runner_lifecycle.go b/client/allocrunnerv2/interfaces/runner_lifecycle.go index 5ccece2e8..63164dce9 100644 --- a/client/allocrunnerv2/interfaces/runner_lifecycle.go +++ b/client/allocrunnerv2/interfaces/runner_lifecycle.go @@ -1,6 +1,8 @@ package interfaces import ( + "context" + "github.com/hashicorp/nomad/client/allocrunnerv2/state" ) @@ -11,7 +13,7 @@ type RunnerHook interface { type RunnerPrerunHook interface { RunnerHook - Prerun() error + Prerun(context.Context) error } type RunnerPostrunHook interface { diff --git a/client/allocrunnerv2/taskrunner/lifecycle.go b/client/allocrunnerv2/taskrunner/lifecycle.go index 9de87f528..521c0157d 100644 --- a/client/allocrunnerv2/taskrunner/lifecycle.go +++ b/client/allocrunnerv2/taskrunner/lifecycle.go @@ -55,6 +55,8 @@ func (tr *TaskRunner) Signal(event *structs.TaskEvent, s os.Signal) error { return handle.Signal(s) } +// Kill a task. Blocks until task exits or context is canceled. State is set to +// dead. func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error { // Grab the handle handle := tr.getDriverHandle() diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 9844bb28e..634a4f804 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -333,11 +333,21 @@ func (tr *TaskRunner) handleUpdates() { for { select { case <-tr.triggerUpdateCh: - // Update triggered; run hooks - tr.updateHooks() case <-tr.waitCh: return } + + if tr.Alloc().TerminalStatus() { + // Terminal update: kill TaskRunner and let Run execute postrun hooks + err := tr.Kill(context.TODO(), structs.NewTaskEvent(structs.TaskKilled)) + if err != nil { + tr.logger.Warn("error stopping task", "error", err) + } + continue + } + + // Non-terminal update; run hooks + tr.updateHooks() } } diff --git a/client/allocrunner/alloc_watcher.go b/client/allocwatcher/alloc_watcher.go similarity index 95% rename from client/allocrunner/alloc_watcher.go rename to client/allocwatcher/alloc_watcher.go index 1cf47bd67..34774e85b 100644 --- a/client/allocrunner/alloc_watcher.go +++ b/client/allocwatcher/alloc_watcher.go @@ -1,4 +1,4 @@ -package allocrunner +package allocwatcher import ( "archive/tar" @@ -38,10 +38,18 @@ type terminated interface { Terminated() bool } -// prevAllocWatcher allows AllocRunners to wait for a previous allocation to +// AllocRunnerMeta provides metadata about an AllocRunner such as its alloc and +// alloc dir. +type AllocRunnerMeta interface { + GetAllocDir() *allocdir.AllocDir + Listener() *cstructs.AllocListener + Alloc() *structs.Allocation +} + +// PrevAllocWatcher allows AllocRunners to wait for a previous allocation to // terminate and migrate its data whether or not the previous allocation is // local or remote. -type prevAllocWatcher interface { +type PrevAllocWatcher interface { // Wait for previous alloc to terminate Wait(context.Context) error @@ -55,10 +63,10 @@ type prevAllocWatcher interface { IsMigrating() bool } -// NewAllocWatcher creates a prevAllocWatcher appropriate for whether this +// NewAllocWatcher creates a PrevAllocWatcher appropriate for whether this // alloc's previous allocation was local or remote. If this alloc has no // previous alloc then a noop implementation is returned. -func NewAllocWatcher(alloc *structs.Allocation, prevAR *AllocRunner, rpc rpcer, config *config.Config, l *log.Logger, migrateToken string) prevAllocWatcher { +func NewAllocWatcher(alloc *structs.Allocation, prevAR AllocRunnerMeta, rpc rpcer, config *config.Config, l *log.Logger, migrateToken string) PrevAllocWatcher { if alloc.PreviousAllocation == "" { // No previous allocation, use noop transitioner return NoopPrevAlloc{} @@ -74,8 +82,7 @@ func NewAllocWatcher(alloc *structs.Allocation, prevAR *AllocRunner, rpc rpcer, tasks: tg.Tasks, sticky: tg.EphemeralDisk != nil && tg.EphemeralDisk.Sticky, prevAllocDir: prevAR.GetAllocDir(), - prevListener: prevAR.GetListener(), - prevWaitCh: prevAR.WaitCh(), + prevListener: prevAR.Listener(), prevStatus: prevAR.Alloc(), logger: l, } @@ -118,10 +125,6 @@ type localPrevAlloc struct { // terminated (and therefore won't send updates to the listener) prevStatus terminated - // prevWaitCh is closed when the previous alloc is garbage collected - // which is a failsafe against blocking the new alloc forever - prevWaitCh <-chan struct{} - // waiting and migrating are true when alloc runner is waiting on the // prevAllocWatcher. Writers must acquire the waitingLock and readers // should use the helper methods IsWaiting and IsMigrating. @@ -161,11 +164,6 @@ func (p *localPrevAlloc) Wait(ctx context.Context) error { defer p.prevListener.Close() - if p.prevStatus.Terminated() { - // Fast path - previous alloc already terminated! - return nil - } - // Block until previous alloc exits p.logger.Printf("[DEBUG] client: alloc %q waiting for previous alloc %q to terminate", p.allocID, p.prevAllocID) for { @@ -174,8 +172,6 @@ func (p *localPrevAlloc) Wait(ctx context.Context) error { if !ok || prevAlloc.Terminated() { return nil } - case <-p.prevWaitCh: - return nil case <-ctx.Done(): return ctx.Err() } diff --git a/client/allocrunner/alloc_watcher_test.go b/client/allocwatcher/alloc_watcher_test.go similarity index 58% rename from client/allocrunner/alloc_watcher_test.go rename to client/allocwatcher/alloc_watcher_test.go index 907f4a521..64509b03f 100644 --- a/client/allocrunner/alloc_watcher_test.go +++ b/client/allocwatcher/alloc_watcher_test.go @@ -1,4 +1,4 @@ -package allocrunner +package allocwatcher import ( "archive/tar" @@ -14,26 +14,118 @@ import ( "testing" "time" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/testutil" + cstructs "github.com/hashicorp/nomad/client/structs" + ctestutil "github.com/hashicorp/nomad/client/testutil" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/nomad/mock" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/testutil" + "github.com/stretchr/testify/require" ) +// fakeAllocRunner implements AllocRunnerMeta +type fakeAllocRunner struct { + alloc *structs.Allocation + AllocDir *allocdir.AllocDir + Broadcaster *cstructs.AllocBroadcaster +} + +// newFakeAllocRunner creates a new AllocRunnerMeta. Callers must call +// AllocDir.Destroy() when finished. +func newFakeAllocRunner(t *testing.T, logger hclog.Logger) *fakeAllocRunner { + alloc := mock.Alloc() + alloc.Job.TaskGroups[0].EphemeralDisk.Sticky = true + alloc.Job.TaskGroups[0].EphemeralDisk.Migrate = true + + path, err := ioutil.TempDir("", "nomad_test_wathcer") + require.NoError(t, err) + + return &fakeAllocRunner{ + alloc: alloc, + AllocDir: allocdir.NewAllocDir(logger, path), + Broadcaster: cstructs.NewAllocBroadcaster(alloc), + } +} + +func (f *fakeAllocRunner) GetAllocDir() *allocdir.AllocDir { + return f.AllocDir +} + +func (f *fakeAllocRunner) Listener() *cstructs.AllocListener { + return f.Broadcaster.Listen() +} + +func (f *fakeAllocRunner) Alloc() *structs.Allocation { + return f.alloc +} + +// newConfig returns a new Config and cleanup func +func newConfig(t *testing.T) (Config, func()) { + logger := testlog.HCLogger(t) + + prevAR := newFakeAllocRunner(t, logger) + + alloc := mock.Alloc() + alloc.PreviousAllocation = prevAR.Alloc().ID + alloc.Job.TaskGroups[0].EphemeralDisk.Sticky = true + alloc.Job.TaskGroups[0].EphemeralDisk.Migrate = true + + config := Config{ + Alloc: alloc, + PreviousRunner: prevAR, + RPC: nil, + Config: nil, + MigrateToken: "fake_token", + Logger: logger, + } + + cleanup := func() { + prevAR.AllocDir.Destroy() + } + + return config, cleanup +} + +// TestPrevAlloc_Noop asserts that when no previous allocation is set the noop +// implementation is returned that does not block or perform migrations. +func TestPrevAlloc_Noop(t *testing.T) { + conf, cleanup := newConfig(t) + defer cleanup() + + conf.Alloc.PreviousAllocation = "" + + watcher := NewAllocWatcher(conf) + require.NotNil(t, watcher) + _, ok := watcher.(NoopPrevAlloc) + require.True(t, ok, "expected watcher to be NoopPrevAlloc") + + done := make(chan int, 2) + go func() { + watcher.Wait(context.Background()) + done <- 1 + watcher.Migrate(context.Background(), nil) + done <- 1 + }() + require.False(t, watcher.IsWaiting()) + require.False(t, watcher.IsMigrating()) + <-done + <-done +} + // TestPrevAlloc_LocalPrevAlloc asserts that when a previous alloc runner is // set a localPrevAlloc will block on it. func TestPrevAlloc_LocalPrevAlloc(t *testing.T) { - _, prevAR := TestAllocRunner(t, false) - prevAR.alloc.Job.TaskGroups[0].Tasks[0].Config["run_for"] = "10s" + t.Parallel() + conf, cleanup := newConfig(t) - newAlloc := mock.Alloc() - newAlloc.PreviousAllocation = prevAR.Alloc().ID - newAlloc.Job.TaskGroups[0].EphemeralDisk.Sticky = false - task := newAlloc.Job.TaskGroups[0].Tasks[0] - task.Driver = "mock_driver" - task.Config["run_for"] = "500ms" + defer cleanup() - waiter := NewAllocWatcher(newAlloc, prevAR, nil, nil, testlog.Logger(t), "") + conf.Alloc.Job.TaskGroups[0].Tasks[0].Driver = "mock_driver" + conf.Alloc.Job.TaskGroups[0].Tasks[0].Config["run_for"] = "500ms" + + waiter := NewAllocWatcher(conf) // Wait in a goroutine with a context to make sure it exits at the right time ctx, cancel := context.WithCancel(context.Background()) @@ -43,39 +135,55 @@ func TestPrevAlloc_LocalPrevAlloc(t *testing.T) { waiter.Wait(ctx) }() - select { - case <-ctx.Done(): - t.Fatalf("Wait exited too early") - case <-time.After(33 * time.Millisecond): - // Good! It's blocking - } + // Assert watcher is waiting + testutil.WaitForResult(func() (bool, error) { + return waiter.IsWaiting(), fmt.Errorf("expected watcher to be waiting") + }, func(err error) { + t.Fatalf("error: %v", err) + }) - // Start the previous allocs to cause it to update but not terminate - go prevAR.Run() - defer prevAR.Destroy() + // Broadcast a non-terminal alloc update to assert only terminal + // updates break out of waiting. + update := conf.PreviousRunner.Alloc().Copy() + update.DesiredStatus = structs.AllocDesiredStatusStop + update.ModifyIndex++ + update.AllocModifyIndex++ - select { - case <-ctx.Done(): - t.Fatalf("Wait exited too early") - case <-time.After(33 * time.Millisecond): - // Good! It's still blocking - } + broadcaster := conf.PreviousRunner.(*fakeAllocRunner).Broadcaster + err := broadcaster.Send(update) + require.NoError(t, err) - // Stop the previous alloc - prevAR.Destroy() + // Assert watcher is still waiting because alloc isn't terminal + testutil.WaitForResult(func() (bool, error) { + return waiter.IsWaiting(), fmt.Errorf("expected watcher to be waiting") + }, func(err error) { + t.Fatalf("error: %v", err) + }) - select { - case <-ctx.Done(): - // Good! We unblocked when the previous alloc stopped - case <-time.After(time.Second): - t.Fatalf("Wait exited too early") - } + // Stop the previous alloc and assert watcher stops blocking + update = update.Copy() + update.DesiredStatus = structs.AllocDesiredStatusStop + update.ClientStatus = structs.AllocClientStatusComplete + update.ModifyIndex++ + update.AllocModifyIndex++ + + err = broadcaster.Send(update) + require.NoError(t, err) + + testutil.WaitForResult(func() (bool, error) { + if waiter.IsWaiting() { + return false, fmt.Errorf("did not expect watcher to be waiting") + } + return !waiter.IsMigrating(), fmt.Errorf("did not expect watcher to be migrating") + }, func(err error) { + t.Fatalf("error: %v", err) + }) } // TestPrevAlloc_StreamAllocDir_Ok asserts that streaming a tar to an alloc dir // works. func TestPrevAlloc_StreamAllocDir_Ok(t *testing.T) { - testutil.RequireRoot(t) + ctestutil.RequireRoot(t) t.Parallel() dir, err := ioutil.TempDir("", "") if err != nil { @@ -178,7 +286,7 @@ func TestPrevAlloc_StreamAllocDir_Ok(t *testing.T) { defer os.RemoveAll(dir1) rc := ioutil.NopCloser(buf) - prevAlloc := &remotePrevAlloc{logger: testlog.Logger(t)} + prevAlloc := &remotePrevAlloc{logger: testlog.HCLogger(t)} if err := prevAlloc.streamAllocDir(context.Background(), rc, dir1); err != nil { t.Fatalf("err: %v", err) } @@ -228,7 +336,7 @@ func TestPrevAlloc_StreamAllocDir_Error(t *testing.T) { // This test only unit tests streamAllocDir so we only need a partially // complete remotePrevAlloc prevAlloc := &remotePrevAlloc{ - logger: testlog.Logger(t), + logger: testlog.HCLogger(t), allocID: "123", prevAllocID: "abc", migrate: true, diff --git a/client/allocwatcher/doc.go b/client/allocwatcher/doc.go new file mode 100644 index 000000000..0580784c9 --- /dev/null +++ b/client/allocwatcher/doc.go @@ -0,0 +1,4 @@ +// Package allocwatcher allows blocking until another allocation - whether +// running locally or remotely - completes and migrates the allocation +// directory if necessary. +package allocwatcher diff --git a/client/client.go b/client/client.go index 9bbb8fb65..0b31c9e2e 100644 --- a/client/client.go +++ b/client/client.go @@ -30,6 +30,7 @@ import ( "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner" "github.com/hashicorp/nomad/client/allocrunnerv2" + "github.com/hashicorp/nomad/client/allocwatcher" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/servers" "github.com/hashicorp/nomad/client/state" @@ -103,9 +104,11 @@ type ClientStatsReporter interface { type AllocRunner interface { StatsReporter() allocrunner.AllocStatsReporter Destroy() + GetAllocDir() *allocdir.AllocDir IsDestroyed() bool IsWaiting() bool IsMigrating() bool + Listener() *cstructs.AllocListener WaitCh() <-chan struct{} Update(*structs.Allocation) Alloc() *structs.Allocation @@ -1913,18 +1916,10 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error return err } - //FIXME disabled previous alloc waiting/migrating - // get the previous alloc runner - if one exists - for the - // blocking/migrating watcher - /* - var prevAR *allocrunner.AllocRunner - if alloc.PreviousAllocation != "" { - prevAR = c.allocs[alloc.PreviousAllocation] - } - - c.configLock.RLock() - prevAlloc := allocrunner.NewAllocWatcher(alloc, prevAR, c, c.configCopy, c.logger, migrateToken) - */ + // Since only the Client has access to other AllocRunners and the RPC + // client, create the previous allocation watcher here. + prevAlloc := c.allocs[alloc.PreviousAllocation] + prevAllocWatcher := allocwatcher.NewAllocWatcher(alloc, prevAlloc, c, c.configCopy, c.logger, migrateToken) // Copy the config since the node can be swapped out as it is being updated. // The long term fix is to pass in the config and node separately and then @@ -1938,13 +1933,14 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error c.configLock.RLock() arConf := &allocrunnerv2.Config{ - Alloc: alloc, - Logger: logger, - ClientConfig: c.config, - StateDB: c.stateDB, - Consul: c.consulService, - Vault: c.vaultClient, - StateUpdater: c, + Alloc: alloc, + Logger: logger, + ClientConfig: c.config, + StateDB: c.stateDB, + Consul: c.consulService, + Vault: c.vaultClient, + StateUpdater: c, + PrevAllocWatcher: prevAllocWatcher, } c.configLock.RUnlock() diff --git a/client/state/interface.go b/client/state/interface.go index 59d6bd08b..952620422 100644 --- a/client/state/interface.go +++ b/client/state/interface.go @@ -12,5 +12,7 @@ type StateDB interface { GetTaskRunnerState(allocID, taskName string) (*state.LocalState, *structs.TaskState, error) PutTaskRunnerLocalState(allocID, taskName string, val interface{}) error PutTaskState(allocID, taskName string, state *structs.TaskState) error + DeleteTaskBucket(allocID, taskName string) error + DeleteAllocationBucket(allocID string) error Close() error } diff --git a/client/state/noopdb.go b/client/state/noopdb.go index 7a31c461f..65d105fd3 100644 --- a/client/state/noopdb.go +++ b/client/state/noopdb.go @@ -27,6 +27,14 @@ func (n noopDB) PutTaskState(allocID string, taskName string, state *structs.Tas return nil } +func (n noopDB) DeleteTaskBucket(allocID, taskName string) error { + return nil +} + +func (n noopDB) DeleteAllocationBucket(allocID string) error { + return nil +} + func (n noopDB) Close() error { return nil } diff --git a/client/structs/funcs.go b/client/structs/funcs.go index ec6d5421c..021e95d2d 100644 --- a/client/structs/funcs.go +++ b/client/structs/funcs.go @@ -1,27 +1,44 @@ package structs import ( + "errors" "sync" "github.com/hashicorp/nomad/nomad/structs" ) -// AllocBroadcaster implements an allocation broadcast channel. -// The zero value is a usable unbuffered channel. +const ( + // listenerCap is the capacity of the listener chans. Must be exactly 1 + // to prevent Sends from blocking and allows them to pop old pending + // updates from the chan before enqueueing the latest update. + listenerCap = 1 +) + +var ErrAllocBroadcasterClosed = errors.New("alloc broadcaster closed") + +// AllocBroadcaster implements an allocation broadcast channel where each +// listener receives allocation updates. Pending updates are dropped and +// replaced by newer allocation updates, so listeners may not receive every +// allocation update. However this ensures Sends never block and listeners only +// receive the latest allocation update -- never a stale version. type AllocBroadcaster struct { m sync.Mutex - listeners map[int]chan<- *structs.Allocation // lazy init + alloc *structs.Allocation + listeners map[int]chan *structs.Allocation // lazy init nextId int - capacity int closed bool } -// NewAllocBroadcaster returns a new AllocBroadcaster with the given capacity (0 means unbuffered). -func NewAllocBroadcaster(n int) *AllocBroadcaster { - return &AllocBroadcaster{capacity: n} +// NewAllocBroadcaster returns a new AllocBroadcaster with the given initial +// allocation. +func NewAllocBroadcaster(initial *structs.Allocation) *AllocBroadcaster { + return &AllocBroadcaster{ + alloc: initial, + } } -// AllocListener implements a listening endpoint for an allocation broadcast channel. +// AllocListener implements a listening endpoint for an allocation broadcast +// channel. type AllocListener struct { // Ch receives the broadcast messages. Ch <-chan *structs.Allocation @@ -29,27 +46,33 @@ type AllocListener struct { id int } -// Send broadcasts a message to the channel. Send returns whether the message -// was sent to all channels. -func (b *AllocBroadcaster) Send(v *structs.Allocation) bool { +// Send broadcasts an allocation update. Any pending updates are replaced with +// this version of the allocation to prevent blocking on slow receivers. +func (b *AllocBroadcaster) Send(v *structs.Allocation) error { b.m.Lock() defer b.m.Unlock() if b.closed { - return false + return ErrAllocBroadcasterClosed } - sent := true + + // Update alloc on broadcaster to send to newly created listeners + b.alloc = v + + // Send alloc to already created listeners for _, l := range b.listeners { select { case l <- v: - default: - sent = false + case <-l: + // Pop pending update and replace with new update + l <- v } } - return sent + return nil } -// Close closes the channel, disabling the sending of further messages. +// Close closes the channel, disabling the sending of further allocation +// updates. func (b *AllocBroadcaster) Close() { b.m.Lock() defer b.m.Unlock() @@ -57,6 +80,7 @@ func (b *AllocBroadcaster) Close() { return } + b.alloc = nil b.closed = true for _, l := range b.listeners { close(l) @@ -68,16 +92,25 @@ func (b *AllocBroadcaster) Listen() *AllocListener { b.m.Lock() defer b.m.Unlock() if b.listeners == nil { - b.listeners = make(map[int]chan<- *structs.Allocation) + b.listeners = make(map[int]chan *structs.Allocation) } + for b.listeners[b.nextId] != nil { b.nextId++ } - ch := make(chan *structs.Allocation, b.capacity) + + ch := make(chan *structs.Allocation, listenerCap) + if b.closed { + // Broadcaster is already closed, close this listener close(ch) + } else { + // Send the current allocation to the listener + ch <- b.alloc } + b.listeners[b.nextId] = ch + return &AllocListener{ch, b, b.nextId} } From 9d1ea3b2287a600834029982c436134be2ae78ca Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 29 Aug 2018 15:05:03 -0700 Subject: [PATCH 052/152] client: hclog-ify most of the client Leaving fingerprinters in case that interface changes with plugins. --- client/acl.go | 4 +- client/allocdir/alloc_dir.go | 9 +- client/allocdir/alloc_dir_test.go | 16 +-- client/allocdir/task_dir.go | 9 +- client/allocdir/task_dir_linux_test.go | 2 +- client/allocdir/task_dir_test.go | 8 +- client/allocrunner/alloc_runner.go | 6 +- client/allocrunner/testing.go | 2 +- client/allocrunnerv2/alloc_runner.go | 7 +- client/allocwatcher/alloc_watcher.go | 98 +++++++++------ client/client.go | 159 +++++++++++++------------ client/fingerprint_manager.go | 47 +++++--- client/gc.go | 34 +++--- client/rpc.go | 16 +-- client/servers/manager.go | 17 +-- client/stats/host.go | 12 +- client/vaultclient/vaultclient.go | 29 +++-- 17 files changed, 260 insertions(+), 215 deletions(-) diff --git a/client/acl.go b/client/acl.go index eb45115cf..dc9c21f61 100644 --- a/client/acl.go +++ b/client/acl.go @@ -134,7 +134,7 @@ func (c *Client) resolveTokenValue(secretID string) (*structs.ACLToken, error) { if err := c.RPC("ACL.ResolveToken", &req, &resp); err != nil { // If we encounter an error but have a cached value, mask the error and extend the cache if ok { - c.logger.Printf("[WARN] client: failed to resolve token, using expired cached value: %v", err) + c.logger.Warn("failed to resolve token, using expired cached value", "error", err) cached := raw.(*cachedACLValue) return cached.Token, nil } @@ -198,7 +198,7 @@ func (c *Client) resolvePolicies(secretID string, policies []string) ([]*structs if err := c.RPC("ACL.GetPolicies", &req, &resp); err != nil { // If we encounter an error but have cached policies, mask the error and extend the cache if len(missing) == 0 { - c.logger.Printf("[WARN] client: failed to resolve policies, using expired cached value: %v", err) + c.logger.Warn("failed to resolve policies, using expired cached value", "error", err) out = append(out, expired...) return out, nil } diff --git a/client/allocdir/alloc_dir.go b/client/allocdir/alloc_dir.go index 65831af66..793c5c81d 100644 --- a/client/allocdir/alloc_dir.go +++ b/client/allocdir/alloc_dir.go @@ -6,12 +6,12 @@ import ( "fmt" "io" "io/ioutil" - "log" "os" "path/filepath" "sync" "time" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/go-multierror" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" @@ -78,7 +78,7 @@ type AllocDir struct { mu sync.RWMutex - logger *log.Logger + logger hclog.Logger } // AllocDirFS exposes file operations on the alloc dir @@ -93,7 +93,8 @@ type AllocDirFS interface { // NewAllocDir initializes the AllocDir struct with allocDir as base path for // the allocation directory. -func NewAllocDir(logger *log.Logger, allocDir string) *AllocDir { +func NewAllocDir(logger hclog.Logger, allocDir string) *AllocDir { + logger = logger.Named("alloc_dir") return &AllocDir{ AllocDir: allocDir, SharedDir: filepath.Join(allocDir, SharedAllocName), @@ -209,7 +210,7 @@ func (d *AllocDir) Snapshot(w io.Writer) error { // the snapshotting side closed the connect // prematurely and won't try to use the tar // anyway. - d.logger.Printf("[WARN] client: snapshotting failed and unable to write error marker: %v", writeErr) + d.logger.Warn("snapshotting failed and unable to write error marker", "error", writeErr) } return fmt.Errorf("failed to snapshot %s: %v", path, err) } diff --git a/client/allocdir/alloc_dir_test.go b/client/allocdir/alloc_dir_test.go index d3a71e3b7..811b29f08 100644 --- a/client/allocdir/alloc_dir_test.go +++ b/client/allocdir/alloc_dir_test.go @@ -54,7 +54,7 @@ func TestAllocDir_BuildAlloc(t *testing.T) { } defer os.RemoveAll(tmp) - d := NewAllocDir(testlog.Logger(t), tmp) + d := NewAllocDir(testlog.HCLogger(t), tmp) defer d.Destroy() d.NewTaskDir(t1.Name) d.NewTaskDir(t2.Name) @@ -91,7 +91,7 @@ func TestAllocDir_MountSharedAlloc(t *testing.T) { } defer os.RemoveAll(tmp) - d := NewAllocDir(testlog.Logger(t), tmp) + d := NewAllocDir(testlog.HCLogger(t), tmp) defer d.Destroy() if err := d.Build(); err != nil { t.Fatalf("Build() failed: %v", err) @@ -136,7 +136,7 @@ func TestAllocDir_Snapshot(t *testing.T) { } defer os.RemoveAll(tmp) - d := NewAllocDir(testlog.Logger(t), tmp) + d := NewAllocDir(testlog.HCLogger(t), tmp) defer d.Destroy() if err := d.Build(); err != nil { t.Fatalf("Build() failed: %v", err) @@ -223,13 +223,13 @@ func TestAllocDir_Move(t *testing.T) { defer os.RemoveAll(tmp2) // Create two alloc dirs - d1 := NewAllocDir(testlog.Logger(t), tmp1) + d1 := NewAllocDir(testlog.HCLogger(t), tmp1) if err := d1.Build(); err != nil { t.Fatalf("Build() failed: %v", err) } defer d1.Destroy() - d2 := NewAllocDir(testlog.Logger(t), tmp2) + d2 := NewAllocDir(testlog.HCLogger(t), tmp2) if err := d2.Build(); err != nil { t.Fatalf("Build() failed: %v", err) } @@ -284,7 +284,7 @@ func TestAllocDir_EscapeChecking(t *testing.T) { } defer os.RemoveAll(tmp) - d := NewAllocDir(testlog.Logger(t), tmp) + d := NewAllocDir(testlog.HCLogger(t), tmp) if err := d.Build(); err != nil { t.Fatalf("Build() failed: %v", err) } @@ -325,7 +325,7 @@ func TestAllocDir_ReadAt_SecretDir(t *testing.T) { } defer os.RemoveAll(tmp) - d := NewAllocDir(testlog.Logger(t), tmp) + d := NewAllocDir(testlog.HCLogger(t), tmp) if err := d.Build(); err != nil { t.Fatalf("Build() failed: %v", err) } @@ -410,7 +410,7 @@ func TestAllocDir_CreateDir(t *testing.T) { // TestAllocDir_Copy asserts that AllocDir.Copy does a deep copy of itself and // all TaskDirs. func TestAllocDir_Copy(t *testing.T) { - a := NewAllocDir(testlog.Logger(t), "foo") + a := NewAllocDir(testlog.HCLogger(t), "foo") a.NewTaskDir("bar") a.NewTaskDir("baz") diff --git a/client/allocdir/task_dir.go b/client/allocdir/task_dir.go index bcf6e6da8..9486258d9 100644 --- a/client/allocdir/task_dir.go +++ b/client/allocdir/task_dir.go @@ -3,10 +3,10 @@ package allocdir import ( "fmt" "io/ioutil" - "log" "os" "path/filepath" + hclog "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" ) @@ -37,15 +37,18 @@ type TaskDir struct { // /secrets/ SecretsDir string - logger *log.Logger + logger hclog.Logger } // newTaskDir creates a TaskDir struct with paths set. Call Build() to // create paths on disk. // // Call AllocDir.NewTaskDir to create new TaskDirs -func newTaskDir(logger *log.Logger, allocDir, taskName string) *TaskDir { +func newTaskDir(logger hclog.Logger, allocDir, taskName string) *TaskDir { taskDir := filepath.Join(allocDir, taskName) + + logger = logger.Named("task_dir").With("task_name", taskName) + return &TaskDir{ Dir: taskDir, SharedAllocDir: filepath.Join(allocDir, SharedAllocName), diff --git a/client/allocdir/task_dir_linux_test.go b/client/allocdir/task_dir_linux_test.go index 0f4da6bce..f3f9a7423 100644 --- a/client/allocdir/task_dir_linux_test.go +++ b/client/allocdir/task_dir_linux_test.go @@ -22,7 +22,7 @@ func TestLinuxSpecialDirs(t *testing.T) { } defer os.RemoveAll(allocDir) - td := newTaskDir(testlog.Logger(t), allocDir, "test") + td := newTaskDir(testlog.HCLogger(t), allocDir, "test") // Despite the task dir not existing, unmountSpecialDirs should *not* // return an error diff --git a/client/allocdir/task_dir_test.go b/client/allocdir/task_dir_test.go index 318e48413..a4c1b7dc9 100644 --- a/client/allocdir/task_dir_test.go +++ b/client/allocdir/task_dir_test.go @@ -18,7 +18,7 @@ func TestTaskDir_EmbedNonexistent(t *testing.T) { } defer os.RemoveAll(tmp) - d := NewAllocDir(testlog.Logger(t), tmp) + d := NewAllocDir(testlog.HCLogger(t), tmp) defer d.Destroy() td := d.NewTaskDir(t1.Name) if err := d.Build(); err != nil { @@ -40,7 +40,7 @@ func TestTaskDir_EmbedDirs(t *testing.T) { } defer os.RemoveAll(tmp) - d := NewAllocDir(testlog.Logger(t), tmp) + d := NewAllocDir(testlog.HCLogger(t), tmp) defer d.Destroy() td := d.NewTaskDir(t1.Name) if err := d.Build(); err != nil { @@ -97,7 +97,7 @@ func TestTaskDir_NonRoot_Image(t *testing.T) { } defer os.RemoveAll(tmp) - d := NewAllocDir(testlog.Logger(t), tmp) + d := NewAllocDir(testlog.HCLogger(t), tmp) defer d.Destroy() td := d.NewTaskDir(t1.Name) if err := d.Build(); err != nil { @@ -120,7 +120,7 @@ func TestTaskDir_NonRoot(t *testing.T) { } defer os.RemoveAll(tmp) - d := NewAllocDir(testlog.Logger(t), tmp) + d := NewAllocDir(testlog.HCLogger(t), tmp) defer d.Destroy() td := d.NewTaskDir(t1.Name) if err := d.Build(); err != nil { diff --git a/client/allocrunner/alloc_runner.go b/client/allocrunner/alloc_runner.go index cee65299d..f436340f0 100644 --- a/client/allocrunner/alloc_runner.go +++ b/client/allocrunner/alloc_runner.go @@ -144,9 +144,9 @@ func NewAllocRunner(logger *log.Logger, config *config.Config, stateDB *bolt.DB, alloc: alloc, allocID: alloc.ID, //allocBroadcast: cstructs.NewAllocBroadcaster(8), - prevAlloc: prevAlloc, - dirtyCh: make(chan struct{}, 1), - allocDir: allocdir.NewAllocDir(logger, filepath.Join(config.AllocDir, alloc.ID)), + prevAlloc: prevAlloc, + dirtyCh: make(chan struct{}, 1), + //allocDir: allocdir.NewAllocDir(logger, filepath.Join(config.AllocDir, alloc.ID)), tasks: make(map[string]*taskrunner.TaskRunner), taskStates: copyTaskStates(alloc.TaskStates), restored: make(map[string]struct{}), diff --git a/client/allocrunner/testing.go b/client/allocrunner/testing.go index 4c4c1a3ff..7c4853623 100644 --- a/client/allocrunner/testing.go +++ b/client/allocrunner/testing.go @@ -52,7 +52,7 @@ func TestAllocRunnerFromAlloc(t *testing.T, alloc *structs.Allocation, restarts alloc.Job.Type = structs.JobTypeBatch } vclient := vaultclient.NewMockVaultClient() - ar := NewAllocRunner(testlog.Logger(t), conf, db, upd.Update, alloc, vclient, consulApi.NewMockConsulServiceClient(t), allocwatcher.NoopPrevAlloc{}) + ar := NewAllocRunner(testlog.Logger(t), conf, db, upd.Update, alloc, vclient, consulApi.NewMockConsulServiceClient(t, testlog.HCLogger(t)), allocwatcher.NoopPrevAlloc{}) return upd, ar } diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 4172fcd52..ec8fd3c46 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -119,13 +119,12 @@ func NewAllocRunner(config *Config) (*allocRunner, error) { prevAllocWatcher: config.PrevAllocWatcher, } - // Create alloc dir - //XXX update AllocDir to hc log - ar.allocDir = allocdir.NewAllocDir(nil, filepath.Join(config.ClientConfig.AllocDir, alloc.ID)) - // Create the logger based on the allocation ID ar.logger = config.Logger.With("alloc_id", alloc.ID) + // Create alloc dir + ar.allocDir = allocdir.NewAllocDir(ar.logger, filepath.Join(config.ClientConfig.AllocDir, alloc.ID)) + // Initialize the runners hooks. ar.initRunnerHooks() diff --git a/client/allocwatcher/alloc_watcher.go b/client/allocwatcher/alloc_watcher.go index 34774e85b..f6e9d4510 100644 --- a/client/allocwatcher/alloc_watcher.go +++ b/client/allocwatcher/alloc_watcher.go @@ -5,7 +5,6 @@ import ( "context" "fmt" "io" - "log" "os" "path/filepath" "sync" @@ -13,6 +12,7 @@ import ( "time" "github.com/hashicorp/consul/lib" + "github.com/hashicorp/go-hclog" nomadapi "github.com/hashicorp/nomad/api" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/config" @@ -26,8 +26,8 @@ const ( getRemoteRetryIntv = 30 * time.Second ) -// rpcer is the interface needed by a prevAllocWatcher to make RPC calls. -type rpcer interface { +// RPCer is the interface needed by a prevAllocWatcher to make RPC calls. +type RPCer interface { // RPC allows retrieving remote allocs. RPC(method string, args interface{}, reply interface{}) error } @@ -63,40 +63,66 @@ type PrevAllocWatcher interface { IsMigrating() bool } +type Config struct { + // Alloc is the current allocation which may need to block on its + // previous allocation stopping. + Alloc *structs.Allocation + + // PreviousRunner is non-nil iff All has a PreviousAllocation and it is + // running locally. + PreviousRunner AllocRunnerMeta + + // RPC allows the alloc watcher to monitor remote allocations. + RPC RPCer + + // Config is necessary for using the RPC. + Config *config.Config + + // MigrateToken is used to migrate remote alloc dirs when ACLs are + // enabled. + MigrateToken string + + Logger hclog.Logger +} + // NewAllocWatcher creates a PrevAllocWatcher appropriate for whether this // alloc's previous allocation was local or remote. If this alloc has no // previous alloc then a noop implementation is returned. -func NewAllocWatcher(alloc *structs.Allocation, prevAR AllocRunnerMeta, rpc rpcer, config *config.Config, l *log.Logger, migrateToken string) PrevAllocWatcher { - if alloc.PreviousAllocation == "" { +func NewAllocWatcher(c Config) PrevAllocWatcher { + if c.Alloc.PreviousAllocation == "" { // No previous allocation, use noop transitioner return NoopPrevAlloc{} } - tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup) + logger := c.Logger.Named("alloc_watcher") + logger = logger.With("alloc_id", c.Alloc.ID) + logger = logger.With("previous_alloc", c.Alloc.PreviousAllocation) - if prevAR != nil { + tg := c.Alloc.Job.LookupTaskGroup(c.Alloc.TaskGroup) + + if c.PreviousRunner != nil { // Previous allocation is local, use local transitioner return &localPrevAlloc{ - allocID: alloc.ID, - prevAllocID: alloc.PreviousAllocation, + allocID: c.Alloc.ID, + prevAllocID: c.Alloc.PreviousAllocation, tasks: tg.Tasks, sticky: tg.EphemeralDisk != nil && tg.EphemeralDisk.Sticky, - prevAllocDir: prevAR.GetAllocDir(), - prevListener: prevAR.Listener(), - prevStatus: prevAR.Alloc(), - logger: l, + prevAllocDir: c.PreviousRunner.GetAllocDir(), + prevListener: c.PreviousRunner.Listener(), + prevStatus: c.PreviousRunner.Alloc(), + logger: logger, } } return &remotePrevAlloc{ - allocID: alloc.ID, - prevAllocID: alloc.PreviousAllocation, + allocID: c.Alloc.ID, + prevAllocID: c.Alloc.PreviousAllocation, tasks: tg.Tasks, - config: config, + config: c.Config, migrate: tg.EphemeralDisk != nil && tg.EphemeralDisk.Migrate, - rpc: rpc, - logger: l, - migrateToken: migrateToken, + rpc: c.RPC, + migrateToken: c.MigrateToken, + logger: logger, } } @@ -132,7 +158,7 @@ type localPrevAlloc struct { migrating bool waitingLock sync.RWMutex - logger *log.Logger + logger hclog.Logger } // IsWaiting returns true if there's a concurrent call inside Wait @@ -165,7 +191,7 @@ func (p *localPrevAlloc) Wait(ctx context.Context) error { defer p.prevListener.Close() // Block until previous alloc exits - p.logger.Printf("[DEBUG] client: alloc %q waiting for previous alloc %q to terminate", p.allocID, p.prevAllocID) + p.logger.Debug("waiting for previous alloc to terminate") for { select { case prevAlloc, ok := <-p.prevListener.Ch: @@ -194,13 +220,14 @@ func (p *localPrevAlloc) Migrate(ctx context.Context, dest *allocdir.AllocDir) e p.waitingLock.Unlock() }() - p.logger.Printf("[DEBUG] client: alloc %q copying previous alloc %q", p.allocID, p.prevAllocID) + p.logger.Debug("copying previous alloc") moveErr := dest.Move(p.prevAllocDir, p.tasks) // Always cleanup previous alloc if err := p.prevAllocDir.Destroy(); err != nil { - p.logger.Printf("[ERR] client: error destroying allocdir %v: %v", p.prevAllocDir.AllocDir, err) + p.logger.Error("error destroying alloc dir", + "error", err, "previous_alloc_dir", p.prevAllocDir.AllocDir) } return moveErr @@ -226,7 +253,7 @@ type remotePrevAlloc struct { // rpc provides an RPC method for watching for updates to the previous // alloc and determining what node it was on. - rpc rpcer + rpc RPCer // nodeID is the node the previous alloc. Set by Wait() for use in // Migrate() iff the previous alloc has not already been GC'd. @@ -239,7 +266,7 @@ type remotePrevAlloc struct { migrating bool waitingLock sync.RWMutex - logger *log.Logger + logger hclog.Logger // migrateToken allows a client to migrate data in an ACL-protected remote // volume @@ -273,7 +300,7 @@ func (p *remotePrevAlloc) Wait(ctx context.Context) error { p.waitingLock.Unlock() }() - p.logger.Printf("[DEBUG] client: alloc %q waiting for remote previous alloc %q to terminate", p.allocID, p.prevAllocID) + p.logger.Debug("waiting for remote previous alloc to terminate") req := structs.AllocSpecificRequest{ AllocID: p.prevAllocID, QueryOptions: structs.QueryOptions{ @@ -296,7 +323,7 @@ func (p *remotePrevAlloc) Wait(ctx context.Context) error { resp := structs.SingleAllocResponse{} err := p.rpc.RPC("Alloc.GetAlloc", &req, &resp) if err != nil { - p.logger.Printf("[ERR] client: failed to query previous alloc %q: %v", p.prevAllocID, err) + p.logger.Error("error querying previous alloc", "error", err) retry := getRemoteRetryIntv + lib.RandomStagger(getRemoteRetryIntv) select { case <-time.After(retry): @@ -306,7 +333,7 @@ func (p *remotePrevAlloc) Wait(ctx context.Context) error { } } if resp.Alloc == nil { - p.logger.Printf("[DEBUG] client: blocking alloc %q has been GC'd", p.prevAllocID) + p.logger.Debug("blocking alloc was GC'd") return nil } if resp.Alloc.Terminated() { @@ -341,12 +368,11 @@ func (p *remotePrevAlloc) Migrate(ctx context.Context, dest *allocdir.AllocDir) p.waitingLock.Unlock() }() - p.logger.Printf("[DEBUG] client: alloc %q copying from remote previous alloc %q", p.allocID, p.prevAllocID) + p.logger.Debug("copying from remote previous alloc") if p.nodeID == "" { // NodeID couldn't be found; likely alloc was GC'd - p.logger.Printf("[WARN] client: alloc %q couldn't migrate data from previous alloc %q; previous alloc may have been GC'd", - p.allocID, p.prevAllocID) + p.logger.Warn("unable to migrate data from previous alloc; previous alloc may have been GC'd") return nil } @@ -367,7 +393,8 @@ func (p *remotePrevAlloc) Migrate(ctx context.Context, dest *allocdir.AllocDir) } if err := prevAllocDir.Destroy(); err != nil { - p.logger.Printf("[ERR] client: error destroying allocdir %q: %v", prevAllocDir.AllocDir, err) + p.logger.Error("error destroying alloc dir", + "error", err, "previous_alloc_dir", prevAllocDir.AllocDir) } return nil } @@ -387,7 +414,7 @@ func (p *remotePrevAlloc) getNodeAddr(ctx context.Context, nodeID string) (strin for { err := p.rpc.RPC("Node.GetNode", &req, &resp) if err != nil { - p.logger.Printf("[ERR] client: failed to query node info %q: %v", nodeID, err) + p.logger.Error("failed to query node", "error", err, "node", nodeID) retry := getRemoteRetryIntv + lib.RandomStagger(getRemoteRetryIntv) select { case <-time.After(retry): @@ -452,7 +479,7 @@ func (p *remotePrevAlloc) migrateAllocDir(ctx context.Context, nodeAddr string) // stream remote alloc to dir to a local path. Caller should cleanup dest on // error. func (p *remotePrevAlloc) streamAllocDir(ctx context.Context, resp io.ReadCloser, dest string) error { - p.logger.Printf("[DEBUG] client: alloc %q streaming snapshot of previous alloc %q to %q", p.allocID, p.prevAllocID, dest) + p.logger.Debug("streaming snapshot of previous alloc", "destination", dest) tr := tar.NewReader(resp) defer resp.Close() @@ -462,8 +489,7 @@ func (p *remotePrevAlloc) streamAllocDir(ctx context.Context, resp io.ReadCloser canceled := func() bool { select { case <-ctx.Done(): - p.logger.Printf("[INFO] client: stopping migration of previous alloc %q for new alloc: %v", - p.prevAllocID, p.allocID) + p.logger.Info("migration of previous alloc canceled") return true default: return false diff --git a/client/client.go b/client/client.go index 0b31c9e2e..afca7d1a8 100644 --- a/client/client.go +++ b/client/client.go @@ -4,7 +4,6 @@ import ( "errors" "fmt" "io/ioutil" - "log" "net" "net/rpc" "os" @@ -25,7 +24,6 @@ import ( nconfig "github.com/hashicorp/nomad/nomad/structs/config" vaultapi "github.com/hashicorp/vault/api" - "github.com/boltdb/bolt" "github.com/hashicorp/consul/lib" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner" @@ -130,7 +128,8 @@ type Client struct { configCopy *config.Config configLock sync.RWMutex - logger *log.Logger + logger hclog.Logger + rpcLogger hclog.Logger connPool *pool.ConnPool @@ -230,7 +229,7 @@ func NewClient(cfg *config.Config, consulCatalog consul.CatalogAPI, consulServic } // Create the logger - logger := cfg.Logger.ResetNamed("") + logger := cfg.Logger.Named("client") // Create the client c := &Client{ @@ -241,7 +240,8 @@ func NewClient(cfg *config.Config, consulCatalog consul.CatalogAPI, consulServic connPool: pool.NewPool(logger, clientRPCCache, clientMaxStreams, tlsWrap), tlsWrap: tlsWrap, streamingRpcs: structs.NewStreamingRpcRegistry(), - logger: logger.StandardLogger(&hclog.StandardLoggerOptions{InferLevels: true}), + rpcLogger: logger.Named("rpc"), + allocs: make(map[string]AllocRunner), allocUpdates: make(chan *structs.Allocation, 64), shutdownCh: make(chan struct{}), triggerDiscoveryCh: make(chan struct{}), @@ -305,7 +305,7 @@ func NewClient(cfg *config.Config, consulCatalog consul.CatalogAPI, consulServic c.configLock.RLock() if len(c.configCopy.Servers) > 0 { if _, err := c.setServersImpl(c.configCopy.Servers, true); err != nil { - c.logger.Printf("[WARN] client: None of the configured servers are valid: %v", err) + logger.Warn("none of the configured servers are valid", "error", err) } } c.configLock.RUnlock() @@ -326,13 +326,13 @@ func NewClient(cfg *config.Config, consulCatalog consul.CatalogAPI, consulServic // Restore the state if err := c.restoreState(); err != nil { - c.logger.Printf("[ERR] client: failed to restore state: %v", err) - c.logger.Printf("[ERR] client: Nomad is unable to start due to corrupt state. "+ + logger.Error("failed to restore state", "error", err) + logger.Error("Nomad is unable to start due to corrupt state. "+ "The safest way to proceed is to manually stop running task processes "+ - "and remove Nomad's state (%q) and alloc (%q) directories before "+ + "and remove Nomad's state and alloc directories before "+ "restarting. Lost allocations will be rescheduled.", - c.config.StateDir, c.config.AllocDir) - c.logger.Printf("[ERR] client: Corrupt state is often caused by a bug. Please " + + "state_dir", c.config.StateDir, "alloc_dir", c.config.AllocDir) + logger.Error("Corrupt state is often caused by a bug. Please " + "report as much information as possible to " + "https://github.com/hashicorp/nomad/issues") return nil, fmt.Errorf("failed to restore state") @@ -353,7 +353,7 @@ func NewClient(cfg *config.Config, consulCatalog consul.CatalogAPI, consulServic // Start collecting stats go c.emitStats() - c.logger.Printf("[INFO] client: Node ID %q", c.NodeID()) + c.logger.Info("started client", "node_id", c.NodeID()) return c, nil } @@ -380,7 +380,7 @@ func (c *Client) init() error { c.config.StateDir = p } - c.logger.Printf("[INFO] client: using state directory %v", c.config.StateDir) + c.logger.Info("using state directory", "state_dir", c.config.StateDir) // Open the state database db, err := state.GetStateDBFactory(c.config.DevMode)(c.config.StateDir) @@ -414,7 +414,7 @@ func (c *Client) init() error { c.config.AllocDir = p } - c.logger.Printf("[INFO] client: using alloc directory %v", c.config.AllocDir) + c.logger.Info("using alloc directory", "alloc_dir", c.config.AllocDir) return nil } @@ -455,7 +455,7 @@ func (c *Client) reloadTLSConnections(newConfig *nconfig.TLSConfig) error { func (c *Client) Reload(newConfig *config.Config) error { shouldReloadTLS, err := tlsutil.ShouldReloadRPCConnections(c.config.TLSConfig, newConfig.TLSConfig) if err != nil { - c.logger.Printf("[ERR] nomad: error parsing server TLS configuration: %s", err) + c.logger.Error("error parsing TLS configuration", "error", err) return err } @@ -513,7 +513,7 @@ func (c *Client) RPCMinorVersion() int { // Shutdown is used to tear down the client func (c *Client) Shutdown() error { - c.logger.Printf("[INFO] client: shutting down") + c.logger.Info("shutting down") c.shutdownLock.Lock() defer c.shutdownLock.Unlock() @@ -524,7 +524,7 @@ func (c *Client) Shutdown() error { // Defer closing the database defer func() { if err := c.stateDB.Close(); err != nil { - c.logger.Printf("[ERR] client: failed to close state database on shutdown: %v", err) + c.logger.Error("error closing state database on shutdown", "error", err) } }() @@ -626,15 +626,12 @@ func (c *Client) GetAllocFS(allocID string) (allocdir.AllocDirFS, error) { c.allocLock.RLock() defer c.allocLock.RUnlock() - _, ok := c.allocs[allocID] + ar, ok := c.allocs[allocID] if !ok { return nil, structs.NewErrUnknownAllocation(allocID) } - //XXX Experiment in splitting AllocDir and TaskDir structs as the FS - //API only needs the immutable AllocDir struct - //FIXME needs a better logger though -- good reason to still get it from alloc runner? - return allocdir.NewAllocDir(c.logger, filepath.Join(c.config.AllocDir, allocID)), nil + return ar.GetAllocDir(), nil } // GetClientAlloc returns the allocation from the client @@ -683,8 +680,8 @@ func (c *Client) setServersImpl(in []string, force bool) (int, error) { defer wg.Done() addr, err := resolveServer(srv) if err != nil { - c.logger.Printf("[DEBUG] client: ignoring server %s due to resolution error: %v", srv, err) mu.Lock() + c.logger.Debug("ignoring server due to resolution error", "error", err, "server", srv) merr.Errors = append(merr.Errors, err) mu.Unlock() return @@ -744,8 +741,8 @@ func (c *Client) restoreState() error { } for allocID, err := range allocErrs { - c.logger.Printf("[ERR] client: failed to restore alloc %q: %v", allocID, err) - //XXX Cleanup? + c.logger.Error("error restoring alloc", "error", err, "alloc_id", allocID) + //TODO Cleanup // Try to clean up alloc dir // Remove boltdb entries? // Send to server with clientstatus=failed @@ -754,17 +751,11 @@ func (c *Client) restoreState() error { // Load each alloc back var mErr multierror.Error for _, alloc := range allocs { - //XXX FIXME create a root logger - logger := hclog.New(&hclog.LoggerOptions{ - Name: "nomad", - Level: hclog.LevelFromString(c.configCopy.LogLevel), - TimeFormat: time.RFC3339, - }) c.configLock.RLock() arConf := &allocrunnerv2.Config{ Alloc: alloc, - Logger: logger, + Logger: c.logger, ClientConfig: c.config, StateDB: c.stateDB, StateUpdater: c, @@ -775,14 +766,14 @@ func (c *Client) restoreState() error { ar, err := allocrunnerv2.NewAllocRunner(arConf) if err != nil { - c.logger.Printf("[ERR] client: failed to create alloc %q: %v", alloc.ID, err) + c.logger.Error("error running alloc", "error", err, "alloc_id", alloc.ID) mErr.Errors = append(mErr.Errors, err) continue } // Restore state if err := ar.Restore(); err != nil { - c.logger.Printf("[ERR] client: failed to restore alloc %q: %v", alloc.ID, err) + c.logger.Error("error restoring alloc", "error", err, "alloc_id", alloc.ID) mErr.Errors = append(mErr.Errors, err) //TODO Cleanup allocrunner continue @@ -823,7 +814,7 @@ func (c *Client) saveState() error { go func(id string, ar AllocRunner) { err := c.stateDB.PutAllocation(ar.Alloc()) if err != nil { - c.logger.Printf("[ERR] client: failed to save state for alloc %q: %v", id, err) + c.logger.Error("error saving alloc state", "error", err, "alloc_id", id) l.Lock() multierror.Append(&mErr, err) l.Unlock() @@ -1202,12 +1193,12 @@ func (c *Client) registerAndHeartbeat() { // registered before if strings.Contains(err.Error(), "node not found") { // Re-register the node - c.logger.Printf("[INFO] client: re-registering node") + c.logger.Info("re-registering node") c.retryRegisterNode() heartbeat = time.After(lib.RandomStagger(initialHeartbeatStagger)) } else { intv := c.getHeartbeatRetryIntv(err) - c.logger.Printf("[ERR] client: heartbeating failed. Retrying in %v: %v", intv, err) + c.logger.Error("error heartbeating. retrying", "error", err, "period", intv) heartbeat = time.After(intv) // If heartbeating fails, trigger Consul discovery @@ -1282,7 +1273,7 @@ func (c *Client) periodicSnapshot() { case <-snapshot: snapshot = time.After(stateSnapshotIntv) if err := c.saveState(); err != nil { - c.logger.Printf("[ERR] client: failed to save state: %v", err) + c.logger.Error("error saving state", "error", err) } case <-c.shutdownCh: @@ -1349,13 +1340,13 @@ func (c *Client) watchNodeEvents() { batchEvents = append(batchEvents, event) } else { // Drop the oldest event - c.logger.Printf("[WARN] client: dropping node event: %v", batchEvents[0]) + c.logger.Warn("dropping node event", "node_event", batchEvents[0]) batchEvents = append(batchEvents[1:], event) } timer.Reset(c.retryIntv(nodeUpdateRetryIntv)) case <-timer.C: if err := c.submitNodeEvents(batchEvents); err != nil { - c.logger.Printf("[ERR] client: submitting node events failed: %v", err) + c.logger.Error("error submitting node events", "error", err) timer.Reset(c.retryIntv(nodeUpdateRetryIntv)) } else { // Reset the events since we successfully sent them. @@ -1388,10 +1379,10 @@ func (c *Client) retryRegisterNode() { } if err == noServersErr { - c.logger.Print("[DEBUG] client: registration waiting on servers") + c.logger.Debug("registration waiting on servers") c.triggerDiscovery() } else { - c.logger.Printf("[ERR] client: registration failure: %v", err) + c.logger.Error("error registering", "error", err) } select { case <-c.rpcRetryWatcher(): @@ -1420,9 +1411,9 @@ func (c *Client) registerNode() error { c.config.Node.Status = structs.NodeStatusReady c.configLock.Unlock() - c.logger.Printf("[INFO] client: node registration complete") + c.logger.Info("node registration complete") if len(resp.EvalIDs) != 0 { - c.logger.Printf("[DEBUG] client: %d evaluations triggered by node registration", len(resp.EvalIDs)) + c.logger.Debug("evaluations triggered by node registration", "num_evals", len(resp.EvalIDs)) } c.heartbeatLock.Lock() @@ -1448,7 +1439,7 @@ func (c *Client) updateNodeStatus() error { end := time.Now() if len(resp.EvalIDs) != 0 { - c.logger.Printf("[DEBUG] client: %d evaluations triggered by node update", len(resp.EvalIDs)) + c.logger.Debug("evaluations triggered by node update", "num_evals", len(resp.EvalIDs)) } // Update the last heartbeat and the new TTL, capturing the old values @@ -1460,15 +1451,15 @@ func (c *Client) updateNodeStatus() error { c.heartbeatTTL = resp.HeartbeatTTL c.haveHeartbeated = true c.heartbeatLock.Unlock() - c.logger.Printf("[TRACE] client: next heartbeat in %v", resp.HeartbeatTTL) + c.logger.Trace("next heartbeat", "period", resp.HeartbeatTTL) if resp.Index != 0 { - c.logger.Printf("[DEBUG] client: state updated to %s", req.Status) + c.logger.Debug("state updated", "node_status", req.Status) // We have potentially missed our TTL log how delayed we were if haveHeartbeated { - c.logger.Printf("[WARN] client: heartbeat missed (request took %v). Heartbeat TTL was %v and heartbeated after %v", - end.Sub(start), oldTTL, time.Since(last)) + c.logger.Warn("missed heartbeat", + "req_latency", end.Sub(start), "heartbeat_ttl", oldTTL, "since_last_heartbeat", time.Since(last)) } } @@ -1481,7 +1472,7 @@ func (c *Client) updateNodeStatus() error { for _, s := range resp.Servers { addr, err := resolveServer(s.RPCAdvertiseAddr) if err != nil { - c.logger.Printf("[WARN] client: ignoring invalid server %q: %v", s.RPCAdvertiseAddr, err) + c.logger.Warn("ignoring invalid server", "error", err, "server", s.RPCAdvertiseAddr) continue } e := &servers.Server{DC: s.Datacenter, Addr: addr} @@ -1573,7 +1564,7 @@ func (c *Client) allocSync() { var resp structs.GenericResponse if err := c.RPC("Node.UpdateAlloc", &args, &resp); err != nil { - c.logger.Printf("[ERR] client: failed to update allocations: %v", err) + c.logger.Error("error updating allocations", "error", err) syncTicker.Stop() syncTicker = time.NewTicker(c.retryIntv(allocSyncRetryIntv)) staggered = true @@ -1648,10 +1639,10 @@ OUTER: // servers are not fully upgraded before the clients register. This // can cause the SecretID to be lost if strings.Contains(err.Error(), "node secret ID does not match") { - c.logger.Printf("[DEBUG] client: re-registering node as there was a secret ID mismatch: %v", err) + c.logger.Debug("secret mismatch; re-registering node", "error", err) c.retryRegisterNode() } else if err != noServersErr { - c.logger.Printf("[ERR] client: failed to query for node allocations: %v", err) + c.logger.Error("error querying node allocations", "error", err) } retry := c.retryIntv(getAllocRetryIntv) select { @@ -1709,7 +1700,7 @@ OUTER: allocsReq.MinQueryIndex = pullIndex - 1 allocsResp = structs.AllocsGetResponse{} if err := c.RPC("Alloc.GetAllocs", &allocsReq, &allocsResp); err != nil { - c.logger.Printf("[ERR] client: failed to query updated allocations: %v", err) + c.logger.Error("error querying updated allocations", "error", err) retry := c.retryIntv(getAllocRetryIntv) select { case <-c.rpcRetryWatcher(): @@ -1751,8 +1742,8 @@ OUTER: } } - c.logger.Printf("[DEBUG] client: updated allocations at index %d (total %d) (pulled %d) (filtered %d)", - resp.Index, len(resp.Allocs), len(allocsResp.Allocs), len(filtered)) + c.logger.Debug("updated allocations", "index", resp.Index, + "total", len(resp.Allocs), "pulled", len(allocsResp.Allocs), "filtered", len(filtered)) // Update the query index. if resp.Index > req.MinQueryIndex { @@ -1799,7 +1790,7 @@ func (c *Client) watchNodeUpdates() { for { select { case <-timer.C: - c.logger.Printf("[DEBUG] client: state changed, updating node and re-registering.") + c.logger.Debug("state changed, updating node and re-registering") c.retryRegisterNode() hasChanged = false case <-c.triggerNodeUpdate: @@ -1826,7 +1817,8 @@ func (c *Client) runAllocs(update *allocUpdates) { // Diff the existing and updated allocations diff := diffAllocs(existing, update) - c.logger.Printf("[DEBUG] client: %#v", diff) + c.logger.Debug("allocation updates", "added", len(diff.added), "removed", len(diff.removed), + "updated", len(diff.updated), "ignored", len(diff.ignore)) // Remove the old allocations for _, remove := range diff.removed { @@ -1835,21 +1827,20 @@ func (c *Client) runAllocs(update *allocUpdates) { // Update the existing allocations for _, update := range diff.updated { - c.logger.Printf("[TRACE] client: updating alloc %q to index %d", update.ID, update.AllocModifyIndex) + c.logger.Trace("updating alloc", "alloc_id", update.ID, "index", update.AllocModifyIndex) c.updateAlloc(update) } // Make room for new allocations before running if err := c.garbageCollector.MakeRoomFor(diff.added); err != nil { - c.logger.Printf("[ERR] client: error making room for new allocations: %v", err) + c.logger.Error("error making room for new allocations", "error", err) } // Start the new allocations for _, add := range diff.added { migrateToken := update.migrateTokens[add.ID] if err := c.addAlloc(add, migrateToken); err != nil { - c.logger.Printf("[ERR] client: failed to add alloc '%s': %v", - add.ID, err) + c.logger.Error("error adding alloc", "error", err, "alloc_id", add.ID) } } @@ -1865,7 +1856,7 @@ func (c *Client) removeAlloc(allocID string) { defer c.allocLock.Unlock() ar, ok := c.allocs[allocID] if !ok { - c.logger.Printf("[WARN] client: missing context for alloc '%s'", allocID) + c.logger.Warn("cannot remove nonexistent alloc", "alloc_id", allocID) return } @@ -1886,13 +1877,13 @@ func (c *Client) updateAlloc(update *structs.Allocation) { defer c.allocLock.Unlock() ar, ok := c.allocs[update.ID] if !ok { - c.logger.Printf("[WARN] client: missing context for alloc %q", update.ID) + c.logger.Warn("cannot update nonexistent alloc", "alloc_id", update.ID) return } // Update local copy of alloc if err := c.stateDB.PutAllocation(update); err != nil { - c.logger.Printf("[ERR] client: error persisting updated alloc locally: %q", update.ID) + c.logger.Error("error persisting updated alloc locally", "error", err, "alloc_id", update.ID) } // Update alloc runner @@ -1906,7 +1897,7 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error // Check if we already have an alloc runner if _, ok := c.allocs[alloc.ID]; ok { - c.logger.Printf("[DEBUG]: client: dropping duplicate add allocation request: %q", alloc.ID) + c.logger.Debug("dropping duplicate add allocation request", "alloc_id", alloc.ID) return nil } @@ -1918,8 +1909,15 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error // Since only the Client has access to other AllocRunners and the RPC // client, create the previous allocation watcher here. - prevAlloc := c.allocs[alloc.PreviousAllocation] - prevAllocWatcher := allocwatcher.NewAllocWatcher(alloc, prevAlloc, c, c.configCopy, c.logger, migrateToken) + watcherConfig := allocwatcher.Config{ + Alloc: alloc, + PreviousRunner: c.allocs[alloc.PreviousAllocation], + RPC: c, + Config: c.configCopy, + MigrateToken: migrateToken, + Logger: c.logger, + } + prevAllocWatcher := allocwatcher.NewAllocWatcher(watcherConfig) // Copy the config since the node can be swapped out as it is being updated. // The long term fix is to pass in the config and node separately and then @@ -1966,7 +1964,7 @@ func (c *Client) setupVaultClient() error { } if c.vaultClient == nil { - c.logger.Printf("[ERR] client: failed to create vault client") + c.logger.Error("failed to create vault client") return fmt.Errorf("failed to create vault client") } @@ -1980,6 +1978,7 @@ func (c *Client) setupVaultClient() error { // tokens for each of the tasks, unwraps all of them using the supplied vault // client and returns a map of unwrapped tokens, indexed by the task name. func (c *Client) deriveToken(alloc *structs.Allocation, taskNames []string, vclient *vaultapi.Client) (map[string]string, error) { + vlogger := c.logger.Named("vault") if alloc == nil { return nil, fmt.Errorf("nil allocation") } @@ -2003,7 +2002,7 @@ func (c *Client) deriveToken(alloc *structs.Allocation, taskNames []string, vcli } } if !found { - c.logger.Printf("[ERR] task %q not found in the allocation", taskName) + vlogger.Error("task not found in the allocation", "task_name", taskName) return nil, fmt.Errorf("task %q not found in the allocation", taskName) } verifiedTasks = append(verifiedTasks, taskName) @@ -2025,15 +2024,15 @@ func (c *Client) deriveToken(alloc *structs.Allocation, taskNames []string, vcli // Derive the tokens var resp structs.DeriveVaultTokenResponse if err := c.RPC("Node.DeriveVaultToken", &req, &resp); err != nil { - c.logger.Printf("[ERR] client.vault: DeriveVaultToken RPC failed: %v", err) + vlogger.Error("error making derive token RPC", "error", err) return nil, fmt.Errorf("DeriveVaultToken RPC failed: %v", err) } if resp.Error != nil { - c.logger.Printf("[ERR] client.vault: failed to derive vault tokens: %v", resp.Error) + vlogger.Error("error deriving vault tokens", "error", resp.Error) return nil, structs.NewWrappedServerError(resp.Error) } if resp.Tasks == nil { - c.logger.Printf("[ERR] client.vault: failed to derive vault token: invalid response") + vlogger.Error("error derivng vault token", "error", "invalid response") return nil, fmt.Errorf("failed to derive vault tokens: invalid response") } @@ -2044,7 +2043,7 @@ func (c *Client) deriveToken(alloc *structs.Allocation, taskNames []string, vcli // Get the wrapped token wrappedToken, ok := resp.Tasks[taskName] if !ok { - c.logger.Printf("[ERR] client.vault: wrapped token missing for task %q", taskName) + vlogger.Error("wrapped token missing for task", "task_name", taskName) return nil, fmt.Errorf("wrapped token missing for task %q", taskName) } @@ -2070,7 +2069,7 @@ func (c *Client) deriveToken(alloc *structs.Allocation, taskNames []string, vcli validationErr = fmt.Errorf("Vault returned unwrap secret with empty Auth.ClientToken. Secret warnings: %v", unwrapResp.Warnings) } if validationErr != nil { - c.logger.Printf("[WARN] client.vault: failed to unwrap token: %v", err) + vlogger.Warn("error unwrapping token", "error", err) return nil, structs.NewRecoverableError(validationErr, true) } @@ -2099,7 +2098,7 @@ func (c *Client) consulDiscovery() { select { case <-c.triggerDiscoveryCh: if err := c.consulDiscoveryImpl(); err != nil { - c.logger.Printf("[ERR] client.consul: error discovering nomad servers: %v", err) + c.logger.Error("error discovering nomad servers", "error", err) } case <-c.shutdownCh: return @@ -2108,6 +2107,8 @@ func (c *Client) consulDiscovery() { } func (c *Client) consulDiscoveryImpl() error { + consulLogger := c.logger.Named("consul") + // Acquire heartbeat lock to prevent heartbeat from running // concurrently with discovery. Concurrent execution is safe, however // discovery is usually triggered when heartbeating has failed so @@ -2142,7 +2143,7 @@ func (c *Client) consulDiscoveryImpl() error { serviceName := c.configCopy.ConsulConfig.ServerServiceName var mErr multierror.Error var nomadServers servers.Servers - c.logger.Printf("[DEBUG] client.consul: bootstrap contacting following Consul DCs: %+q", dcs) + consulLogger.Debug("bootstrap contacting Consul DCs", "consul_dcs", dcs) DISCOLOOP: for _, dc := range dcs { consulOpts := &consulapi.QueryOptions{ @@ -2196,7 +2197,7 @@ DISCOLOOP: return fmt.Errorf("no Nomad Servers advertising service %q in Consul datacenters: %+q", serviceName, dcs) } - c.logger.Printf("[INFO] client.consul: discovered following Servers: %s", nomadServers) + consulLogger.Info("discovered following servers", "servers", nomadServers) // Fire the retry trigger if we have updated the set of servers. if c.servers.SetServers(nomadServers) { @@ -2238,7 +2239,7 @@ func (c *Client) emitStats() { err := c.hostStatsCollector.Collect() next.Reset(c.config.StatsCollectionInterval) if err != nil { - c.logger.Printf("[WARN] client: error fetching host resource usage stats: %v", err) + c.logger.Warn("error fetching host resource usage stats", "error", err) continue } diff --git a/client/fingerprint_manager.go b/client/fingerprint_manager.go index f4c6d44c3..340bac876 100644 --- a/client/fingerprint_manager.go +++ b/client/fingerprint_manager.go @@ -3,9 +3,11 @@ package client import ( "fmt" "log" + "os" "sync" "time" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/fingerprint" @@ -28,7 +30,7 @@ type FingerprintManager struct { // updateNodeFromDriver is a callback to the client to update the state of a // specific driver for the node updateNodeFromDriver func(string, *structs.DriverInfo, *structs.DriverInfo) *structs.Node - logger *log.Logger + logger hclog.Logger } // NewFingerprintManager is a constructor that creates and returns an instance @@ -38,7 +40,9 @@ func NewFingerprintManager(getConfig func() *config.Config, shutdownCh chan struct{}, updateNodeAttributes func(*cstructs.FingerprintResponse) *structs.Node, updateNodeFromDriver func(string, *structs.DriverInfo, *structs.DriverInfo) *structs.Node, - logger *log.Logger) *FingerprintManager { + logger hclog.Logger) *FingerprintManager { + + logger = logger.Named("fingerprint_mgr") return &FingerprintManager{ getConfig: getConfig, updateNodeAttributes: updateNodeAttributes, @@ -74,7 +78,7 @@ func (fp *FingerprintManager) Run() error { whitelistFingerprintsEnabled := len(whitelistFingerprints) > 0 blacklistFingerprints := cfg.ReadStringListToMap("fingerprint.blacklist") - fp.logger.Printf("[DEBUG] client.fingerprint_manager: built-in fingerprints: %v", fingerprint.BuiltinFingerprints()) + fp.logger.Debug("built-in fingerprints", "fingerprinters", fingerprint.BuiltinFingerprints()) var availableFingerprints []string var skippedFingerprints []string @@ -98,7 +102,8 @@ func (fp *FingerprintManager) Run() error { } if len(skippedFingerprints) != 0 { - fp.logger.Printf("[DEBUG] client.fingerprint_manager: fingerprint modules skipped due to white/blacklist: %v", skippedFingerprints) + fp.logger.Debug("fingerprint modules skipped due to white/blacklist", + "skipped_fingerprinters", skippedFingerprints) } // Next, set up drivers @@ -131,7 +136,7 @@ func (fp *FingerprintManager) Run() error { } if len(skippedDrivers) > 0 { - fp.logger.Printf("[DEBUG] client.fingerprint_manager: drivers skipped due to white/blacklist: %v", skippedDrivers) + fp.logger.Debug("drivers skipped due to white/blacklist", "skipped_drivers", skippedDrivers) } return nil } @@ -139,13 +144,16 @@ func (fp *FingerprintManager) Run() error { // setupFingerprints is used to fingerprint the node to see if these attributes are // supported func (fm *FingerprintManager) setupFingerprinters(fingerprints []string) error { + //FIXME Update fingerprinters to hclog + tempLogger := log.New(os.Stdout, "", log.LstdFlags|log.Lmicroseconds) + var appliedFingerprints []string for _, name := range fingerprints { - f, err := fingerprint.NewFingerprint(name, fm.logger) + f, err := fingerprint.NewFingerprint(name, tempLogger) if err != nil { - fm.logger.Printf("[ERR] client.fingerprint_manager: fingerprinting for %v failed: %+v", name, err) + fm.logger.Error("error fingerprinting", "error", err, "fingerprinter", name) return err } @@ -165,15 +173,18 @@ func (fm *FingerprintManager) setupFingerprinters(fingerprints []string) error { } } - fm.logger.Printf("[DEBUG] client.fingerprint_manager: detected fingerprints %v", appliedFingerprints) + fm.logger.Debug("detected fingerprints", "node_attrs", appliedFingerprints) return nil } // setupDrivers is used to fingerprint the node to see if these drivers are // supported func (fm *FingerprintManager) setupDrivers(drivers []string) error { + //FIXME Update fingerprinters to hclog + tempLogger := log.New(os.Stdout, "", log.LstdFlags|log.Lmicroseconds) + var availDrivers []string - driverCtx := driver.NewDriverContext("", "", "", "", fm.getConfig(), fm.getNode(), fm.logger, nil) + driverCtx := driver.NewDriverContext("", "", "", "", fm.getConfig(), fm.getNode(), tempLogger, nil) for _, name := range drivers { d, err := driver.NewDriver(name, driverCtx) @@ -189,7 +200,7 @@ func (fm *FingerprintManager) setupDrivers(drivers []string) error { // where health checks are enabled. detected, err := fm.fingerprintDriver(name, d, true) if err != nil { - fm.logger.Printf("[DEBUG] client.fingerprint_manager: fingerprinting driver %v failed: %+v", name, err) + fm.logger.Debug("error fingerprinting driver", "error", err, "fingerprinter", name) return err } @@ -203,13 +214,13 @@ func (fm *FingerprintManager) setupDrivers(drivers []string) error { } } - fm.logger.Printf("[DEBUG] client.fingerprint_manager: detected drivers %v", availDrivers) + fm.logger.Debug("detected drivers", "drivers", availDrivers) return nil } // runFingerprint runs each fingerprinter individually on an ongoing basis func (fm *FingerprintManager) runFingerprint(f fingerprint.Fingerprint, period time.Duration, name string) { - fm.logger.Printf("[DEBUG] client.fingerprint_manager: fingerprinting %s every %v", name, period) + fm.logger.Debug("fingerprinting periodically", "fingerprinter", name, "period", period) timer := time.NewTimer(period) defer timer.Stop() @@ -221,7 +232,7 @@ func (fm *FingerprintManager) runFingerprint(f fingerprint.Fingerprint, period t _, err := fm.fingerprint(name, f) if err != nil { - fm.logger.Printf("[DEBUG] client.fingerprint_manager: periodic fingerprinting for %v failed: %+v", name, err) + fm.logger.Debug("error periodic fingerprinting", "error", err, "fingerprinter", name) continue } @@ -272,7 +283,7 @@ func (fm *FingerprintManager) watchDriver(d driver.Driver, name string) { ticker := time.NewTicker(fingerprintPeriod) fingerprintTicker = ticker.C defer ticker.Stop() - fm.logger.Printf("[DEBUG] client.fingerprint_manager: fingerprinting driver %s every %v", name, fingerprintPeriod) + fm.logger.Debug("periodically fingerprinting driver", "name", name, "period", fingerprintPeriod) } var isHealthCheckPeriodic bool @@ -282,13 +293,13 @@ func (fm *FingerprintManager) watchDriver(d driver.Driver, name string) { var healthCheckResp cstructs.HealthCheckIntervalResponse if err := hc.GetHealthCheckInterval(req, &healthCheckResp); err != nil { - fm.logger.Printf("[ERR] client.fingerprint_manager: error getting health check interval for driver %s: %v", name, err) + fm.logger.Error("error getting health check interval for driver", "error", err, "driver", name) } else if healthCheckResp.Eligible { isHealthCheckPeriodic = true ticker := time.NewTicker(healthCheckResp.Period) healthTicker = ticker.C defer ticker.Stop() - fm.logger.Printf("[DEBUG] client.fingerprint_manager: health checking driver %s every %v", name, healthCheckResp.Period) + fm.logger.Debug("periodically health checking driver", "driver", name, "period", healthCheckResp.Period) } } @@ -299,7 +310,7 @@ func (fm *FingerprintManager) watchDriver(d driver.Driver, name string) { return case <-fingerprintTicker: if _, err := fm.fingerprintDriver(name, d, isHealthCheckPeriodic); err != nil { - fm.logger.Printf("[DEBUG] client.fingerprint_manager: periodic fingerprinting for driver %v failed: %+v", name, err) + fm.logger.Debug("error periodic fingerprinting driver", "error", err, "driver", name) } fm.nodeLock.Lock() @@ -316,7 +327,7 @@ func (fm *FingerprintManager) watchDriver(d driver.Driver, name string) { case <-healthTicker: if driverEverDetected { if err := fm.runDriverHealthCheck(name, hc); err != nil { - fm.logger.Printf("[DEBUG] client.fingerprint_manager: health checking for %v failed: %v", name, err) + fm.logger.Debug("error health checking", "error", err, "driver", name) } } } diff --git a/client/gc.go b/client/gc.go index 4d4eb7f17..c08c43bdb 100644 --- a/client/gc.go +++ b/client/gc.go @@ -3,10 +3,10 @@ package client import ( "container/heap" "fmt" - "log" "sync" "time" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/stats" "github.com/hashicorp/nomad/nomad/structs" ) @@ -57,16 +57,17 @@ type AllocGarbageCollector struct { // triggerCh is ticked by the Trigger method to cause a GC triggerCh chan struct{} - logger *log.Logger + logger hclog.Logger } // NewAllocGarbageCollector returns a garbage collector for terminated // allocations on a node. Must call Run() in a goroutine enable periodic // garbage collection. -func NewAllocGarbageCollector(logger *log.Logger, statsCollector stats.NodeStatsCollector, ac AllocCounter, config *GCConfig) *AllocGarbageCollector { +func NewAllocGarbageCollector(logger hclog.Logger, statsCollector stats.NodeStatsCollector, ac AllocCounter, config *GCConfig) *AllocGarbageCollector { + logger = logger.Named("gc") // Require at least 1 to make progress if config.ParallelDestroys <= 0 { - logger.Printf("[WARN] client.gc: garbage collector defaulting parallelism to 1 due to invalid input value of %d", config.ParallelDestroys) + logger.Warn("garbage collector defaulting parallelism to 1 due to invalid input value", "gc_parallel_destroys", config.ParallelDestroys) config.ParallelDestroys = 1 } @@ -97,7 +98,7 @@ func (a *AllocGarbageCollector) Run() { } if err := a.keepUsageBelowThreshold(); err != nil { - a.logger.Printf("[ERR] client.gc: error garbage collecting allocation: %v", err) + a.logger.Error("error garbage collecting allocations", "error", err) } } } @@ -129,7 +130,7 @@ func (a *AllocGarbageCollector) keepUsageBelowThreshold() error { // See if we are below thresholds for used disk space and inode usage diskStats := a.statsCollector.Stats().AllocDirStats reason := "" - level := "WARN" + logf := a.logger.Warn liveAllocs := a.allocCounter.NumAllocs() @@ -143,7 +144,7 @@ func (a *AllocGarbageCollector) keepUsageBelowThreshold() error { case liveAllocs > a.config.MaxAllocs: // if we're unable to gc, don't WARN until at least 2x over limit if liveAllocs < (a.config.MaxAllocs * 2) { - level = "INFO" + logf = a.logger.Info } reason = fmt.Sprintf("number of allocations (%d) is over the limit (%d)", liveAllocs, a.config.MaxAllocs) } @@ -156,7 +157,7 @@ func (a *AllocGarbageCollector) keepUsageBelowThreshold() error { // Collect an allocation gcAlloc := a.allocRunners.Pop() if gcAlloc == nil { - a.logger.Printf("[%s] client.gc: garbage collection due to %s skipped because no terminal allocations", level, reason) + logf("garbage collection skipped because no terminal allocations", "reason", reason) break } @@ -170,7 +171,7 @@ func (a *AllocGarbageCollector) keepUsageBelowThreshold() error { // lock to restrict parallelism and then destroy the alloc runner, returning // once the allocation has been destroyed. func (a *AllocGarbageCollector) destroyAllocRunner(allocID string, ar AllocRunner, reason string) { - a.logger.Printf("[INFO] client.gc: garbage collecting allocation %s due to %s", allocID, reason) + a.logger.Info("garbage collecting allocation", "alloc_id", allocID, "reason", reason) // Acquire the destroy lock select { @@ -186,7 +187,7 @@ func (a *AllocGarbageCollector) destroyAllocRunner(allocID string, ar AllocRunne case <-a.shutdownCh: } - a.logger.Printf("[DEBUG] client.gc: garbage collected %s", allocID) + a.logger.Debug("garbage collected %s", "alloc_id", allocID) // Release the lock <-a.destroyCh @@ -199,13 +200,14 @@ func (a *AllocGarbageCollector) Stop() { // Collect garbage collects a single allocation on a node. Returns true if // alloc was found and garbage collected; otherwise false. func (a *AllocGarbageCollector) Collect(allocID string) bool { - if gcAlloc := a.allocRunners.Remove(allocID); gcAlloc != nil { - a.destroyAllocRunner(allocID, gcAlloc.allocRunner, "forced collection") - return true + gcAlloc := a.allocRunners.Remove(allocID) + if gcAlloc == nil { + a.logger.Debug("alloc was already garbage collected", "alloc_id", allocID) + return false } - a.logger.Printf("[DEBUG] client.gc: alloc %s is invalid or was already garbage collected", allocID) - return false + a.destroyAllocRunner(allocID, gcAlloc.allocRunner, "forced collection") + return true } // CollectAll garbage collects all terminated allocations on a node @@ -335,7 +337,7 @@ func (a *AllocGarbageCollector) MakeRoomFor(allocations []*structs.Allocation) e // MarkForCollection starts tracking an allocation for Garbage Collection func (a *AllocGarbageCollector) MarkForCollection(allocID string, ar AllocRunner) { if a.allocRunners.Push(allocID, ar) { - a.logger.Printf("[INFO] client.gc: marking allocation %v for GC", ar.Alloc().ID) + a.logger.Info("marking allocation for GC", "alloc_id", allocID) } } diff --git a/client/rpc.go b/client/rpc.go index 151421c66..730259e52 100644 --- a/client/rpc.go +++ b/client/rpc.go @@ -72,7 +72,7 @@ TRY: } // Move off to another server, and see if we can retry. - c.logger.Printf("[ERR] nomad: %q RPC failed to server %s: %v", method, server.Addr, rpcErr) + c.rpcLogger.Error("error performing RPC to server", "error", rpcErr, "rpc", method, "server", server.Addr) c.servers.NotifyFailedServer(server) if retry := canRetry(args, rpcErr); !retry { return rpcErr @@ -119,7 +119,7 @@ func (c *Client) RemoteStreamingRpcHandler(method string) (structs.StreamingRpcH conn, err := c.streamingRpcConn(server, method) if err != nil { // Move off to another server - c.logger.Printf("[ERR] nomad: %q RPC failed to server %s: %v", method, server.Addr, err) + c.rpcLogger.Error("error performing RPC to server", "error", err, "rpc", method, "server", server.Addr) c.servers.NotifyFailedServer(server) return nil, err } @@ -261,7 +261,7 @@ func (c *Client) listenConn(s *yamux.Session) { return } - c.logger.Printf("[ERR] client.rpc: failed to accept RPC conn: %v", err) + c.rpcLogger.Error("[ERR] client.rpc: failed to accept RPC conn: %v", err) continue } @@ -277,7 +277,7 @@ func (c *Client) handleConn(conn net.Conn) { buf := make([]byte, 1) if _, err := conn.Read(buf); err != nil { if err != io.EOF { - c.logger.Printf("[ERR] client.rpc: failed to read byte: %v", err) + c.rpcLogger.Error("error reading byte", "error", err) } conn.Close() return @@ -292,7 +292,7 @@ func (c *Client) handleConn(conn net.Conn) { c.handleStreamingConn(conn) default: - c.logger.Printf("[ERR] client.rpc: unrecognized RPC byte: %v", buf[0]) + c.rpcLogger.Error("unrecognized RPC byte", "byte", buf[0]) conn.Close() return } @@ -311,7 +311,7 @@ func (c *Client) handleNomadConn(conn net.Conn) { if err := c.rpcServer.ServeRequest(rpcCodec); err != nil { if err != io.EOF && !strings.Contains(err.Error(), "closed") { - c.logger.Printf("[ERR] client.rpc: RPC error: %v (%v)", err, conn) + c.rpcLogger.Error("error performing RPC", "error", err, "addr", conn.RemoteAddr()) metrics.IncrCounter([]string{"client", "rpc", "request_error"}, 1) } return @@ -329,7 +329,7 @@ func (c *Client) handleStreamingConn(conn net.Conn) { decoder := codec.NewDecoder(conn, structs.MsgpackHandle) if err := decoder.Decode(&header); err != nil { if err != io.EOF && !strings.Contains(err.Error(), "closed") { - c.logger.Printf("[ERR] client.rpc: Streaming RPC error: %v (%v)", err, conn) + c.rpcLogger.Error("error performing streaming RPC", "error", err, "addr", conn.RemoteAddr()) metrics.IncrCounter([]string{"client", "streaming_rpc", "request_error"}, 1) } @@ -339,7 +339,7 @@ func (c *Client) handleStreamingConn(conn net.Conn) { ack := structs.StreamingRpcAck{} handler, err := c.streamingRpcs.GetHandler(header.Method) if err != nil { - c.logger.Printf("[ERR] client.rpc: Streaming RPC error: %v (%v)", err, conn) + c.rpcLogger.Error("streaming RPC error", "addr", conn.RemoteAddr(), "error", err) metrics.IncrCounter([]string{"client", "streaming_rpc", "request_error"}, 1) ack.Error = err.Error() } diff --git a/client/servers/manager.go b/client/servers/manager.go index 31d12b831..fea945dc5 100644 --- a/client/servers/manager.go +++ b/client/servers/manager.go @@ -4,7 +4,6 @@ package servers import ( - "log" "math/rand" "net" "sort" @@ -13,6 +12,7 @@ import ( "time" "github.com/hashicorp/consul/lib" + hclog "github.com/hashicorp/go-hclog" ) const ( @@ -153,8 +153,6 @@ type Manager struct { // shutdownCh is a copy of the channel in Nomad.Client shutdownCh chan struct{} - logger *log.Logger - // numNodes is used to estimate the approximate number of nodes in // a cluster and limit the rate at which it rebalances server // connections. This should be read and set using atomic. @@ -164,11 +162,14 @@ type Manager struct { // pool. Pinger is an interface that wraps client.ConnPool. connPoolPinger Pinger + logger hclog.Logger + sync.Mutex } // New is the only way to safely create a new Manager struct. -func New(logger *log.Logger, shutdownCh chan struct{}, connPoolPinger Pinger) (m *Manager) { +func New(logger hclog.Logger, shutdownCh chan struct{}, connPoolPinger Pinger) (m *Manager) { + logger = logger.Named("server_mgr") return &Manager{ logger: logger, connPoolPinger: connPoolPinger, @@ -188,7 +189,7 @@ func (m *Manager) Start() { m.refreshServerRebalanceTimer() case <-m.shutdownCh: - m.logger.Printf("[DEBUG] manager: shutting down") + m.logger.Debug("shutting down") return } } @@ -221,7 +222,7 @@ func (m *Manager) FindServer() *Server { defer m.Unlock() if len(m.servers) == 0 { - m.logger.Printf("[WARN] manager: No servers available") + m.logger.Warn("no servers available") return nil } @@ -303,13 +304,13 @@ func (m *Manager) RebalanceServers() { foundHealthyServer = true break } - m.logger.Printf(`[DEBUG] manager: pinging server "%s" failed: %s`, srv, err) + m.logger.Debug("error pinging server", "error", err, "server", srv) servers.cycle() } if !foundHealthyServer { - m.logger.Printf("[DEBUG] manager: No healthy servers during rebalance") + m.logger.Debug("no healthy servers during rebalance") return } diff --git a/client/stats/host.go b/client/stats/host.go index b811ee51c..c3cb90cbb 100644 --- a/client/stats/host.go +++ b/client/stats/host.go @@ -2,12 +2,12 @@ package stats import ( "fmt" - "log" "math" "runtime" "sync" "time" + hclog "github.com/hashicorp/go-hclog" "github.com/shirou/gopsutil/cpu" "github.com/shirou/gopsutil/disk" "github.com/shirou/gopsutil/host" @@ -64,7 +64,6 @@ type NodeStatsCollector interface { type HostStatsCollector struct { numCores int statsCalculator map[string]*HostCpuStatsCalculator - logger *log.Logger hostStats *HostStats hostStatsLock sync.RWMutex allocDir string @@ -72,12 +71,15 @@ type HostStatsCollector struct { // badParts is a set of partitions whose usage cannot be read; used to // squelch logspam. badParts map[string]struct{} + + logger hclog.Logger } // NewHostStatsCollector returns a HostStatsCollector. The allocDir is passed in // so that we can present the disk related statistics for the mountpoint where // the allocation directory lives -func NewHostStatsCollector(logger *log.Logger, allocDir string) *HostStatsCollector { +func NewHostStatsCollector(logger hclog.Logger, allocDir string) *HostStatsCollector { + logger = logger.Named("host_stats") numCores := runtime.NumCPU() statsCalculator := make(map[string]*HostCpuStatsCalculator) collector := &HostStatsCollector{ @@ -175,7 +177,7 @@ func (h *HostStatsCollector) collectDiskStats() ([]*DiskStats, error) { } h.badParts[partition.Mountpoint] = struct{}{} - h.logger.Printf("[WARN] client: error fetching host disk usage stats for %v: %v", partition.Mountpoint, err) + h.logger.Warn("error fetching host disk usage stats", "error", err, "partition", partition.Mountpoint) continue } delete(h.badParts, partition.Mountpoint) @@ -194,7 +196,7 @@ func (h *HostStatsCollector) Stats() *HostStats { if h.hostStats == nil { if err := h.collectLocked(); err != nil { - h.logger.Printf("[WARN] client: error fetching host resource usage stats: %v", err) + h.logger.Warn("error fetching host resource usage stats", "error", err) } } diff --git a/client/vaultclient/vaultclient.go b/client/vaultclient/vaultclient.go index b72bf26dd..e08bd33ea 100644 --- a/client/vaultclient/vaultclient.go +++ b/client/vaultclient/vaultclient.go @@ -3,12 +3,12 @@ package vaultclient import ( "container/heap" "fmt" - "log" "math/rand" "strings" "sync" "time" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/nomad/structs/config" vaultapi "github.com/hashicorp/vault/api" @@ -82,7 +82,7 @@ type vaultClient struct { config *config.VaultConfig lock sync.RWMutex - logger *log.Logger + logger hclog.Logger } // vaultClientRenewalRequest is a request object for renewal of both tokens and @@ -122,14 +122,12 @@ type vaultClientHeap struct { type vaultDataHeapImp []*vaultClientHeapEntry // NewVaultClient returns a new vault client from the given config. -func NewVaultClient(config *config.VaultConfig, logger *log.Logger, tokenDeriver TokenDeriverFunc) (*vaultClient, error) { +func NewVaultClient(config *config.VaultConfig, logger hclog.Logger, tokenDeriver TokenDeriverFunc) (*vaultClient, error) { if config == nil { return nil, fmt.Errorf("nil vault config") } - if logger == nil { - return nil, fmt.Errorf("nil logger") - } + logger = logger.Named("vault") c := &vaultClient{ config: config, @@ -148,14 +146,14 @@ func NewVaultClient(config *config.VaultConfig, logger *log.Logger, tokenDeriver // Get the Vault API configuration apiConf, err := config.ApiConfig() if err != nil { - logger.Printf("[ERR] client.vault: failed to create vault API config: %v", err) + logger.Error("error creating vault API config", "error", err) return nil, err } // Create the Vault API client client, err := vaultapi.NewClient(apiConf) if err != nil { - logger.Printf("[ERR] client.vault: failed to create Vault client. Not retrying: %v", err) + logger.Error("error creating vault client", "error", err) return nil, err } @@ -237,7 +235,7 @@ func (c *vaultClient) DeriveToken(alloc *structs.Allocation, taskNames []string) tokens, err := c.tokenDeriver(alloc, taskNames, c.client) if err != nil { - c.logger.Printf("[ERR] client.vault: failed to derive token for allocation %q and tasks %v: %v", alloc.ID, taskNames, err) + c.logger.Error("error deriving token", "error", err, "alloc_id", alloc.ID, "task_names", taskNames) return nil, err } @@ -299,7 +297,7 @@ func (c *vaultClient) RenewToken(token string, increment int) (<-chan error, err // Perform the renewal of the token and send any error to the dedicated // error channel. if err := c.renew(renewalReq); err != nil { - c.logger.Printf("[ERR] client.vault: renewal of token failed: %v", err) + c.logger.Error("error during renewal of token", "error", err) return nil, err } @@ -336,7 +334,7 @@ func (c *vaultClient) RenewLease(leaseId string, increment int) (<-chan error, e // Renew the secret and send any error to the dedicated error channel if err := c.renew(renewalReq); err != nil { - c.logger.Printf("[ERR] client.vault: renewal of lease failed: %v", err) + c.logger.Error("error during renewal of lease", "error", err) return nil, err } @@ -433,8 +431,9 @@ func (c *vaultClient) renew(req *vaultClientRenewalRequest) error { strings.Contains(renewalErr.Error(), "permission denied")) { fatal = true } else if renewalErr != nil { - c.logger.Printf("[DEBUG] client.vault: req.increment: %d, leaseDuration: %d, duration: %d", req.increment, leaseDuration, duration) - c.logger.Printf("[ERR] client.vault: renewal of lease or token failed due to a non-fatal error. Retrying at %v: %v", next.String(), renewalErr) + c.logger.Debug("renewal error details", "req.increment", req.increment, "lease_duration", leaseDuration, "duration", duration) + c.logger.Error("error during renewal of lease or token failed due to a non-fatal error; retrying", + "error", renewalErr, "period", next) } if c.isTracked(req.id) { @@ -531,12 +530,12 @@ func (c *vaultClient) run() { select { case <-renewalCh: if err := c.renew(renewalReq); err != nil { - c.logger.Printf("[ERR] client.vault: renewal of token failed: %v", err) + c.logger.Error("error renewing token", "error", err) } case <-c.updateCh: continue case <-c.stopCh: - c.logger.Printf("[DEBUG] client.vault: stopped") + c.logger.Debug("stopped") return } } From 0bbf3a93ee525a0c444bc836c525b10972c3600d Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 30 Aug 2018 16:59:16 -0700 Subject: [PATCH 053/152] make AllocBroadcaster easier to use And test thoroughly. --- client/structs/{funcs.go => broadcaster.go} | 59 +++++-- client/structs/broadcaster_test.go | 164 ++++++++++++++++++++ 2 files changed, 206 insertions(+), 17 deletions(-) rename client/structs/{funcs.go => broadcaster.go} (80%) create mode 100644 client/structs/broadcaster_test.go diff --git a/client/structs/funcs.go b/client/structs/broadcaster.go similarity index 80% rename from client/structs/funcs.go rename to client/structs/broadcaster.go index 021e95d2d..88b51ebf8 100644 --- a/client/structs/funcs.go +++ b/client/structs/broadcaster.go @@ -37,17 +37,9 @@ func NewAllocBroadcaster(initial *structs.Allocation) *AllocBroadcaster { } } -// AllocListener implements a listening endpoint for an allocation broadcast -// channel. -type AllocListener struct { - // Ch receives the broadcast messages. - Ch <-chan *structs.Allocation - b *AllocBroadcaster - id int -} - // Send broadcasts an allocation update. Any pending updates are replaced with // this version of the allocation to prevent blocking on slow receivers. +// Returns ErrAllocBroadcasterClosed if called after broadcaster is closed. func (b *AllocBroadcaster) Send(v *structs.Allocation) error { b.m.Lock() defer b.m.Unlock() @@ -72,7 +64,7 @@ func (b *AllocBroadcaster) Send(v *structs.Allocation) error { } // Close closes the channel, disabling the sending of further allocation -// updates. +// updates. Safe to call concurrently and more than once. func (b *AllocBroadcaster) Close() { b.m.Lock() defer b.m.Unlock() @@ -80,11 +72,36 @@ func (b *AllocBroadcaster) Close() { return } - b.alloc = nil - b.closed = true + // Close all listener chans for _, l := range b.listeners { close(l) } + + // Clear all references and mark broadcaster as closed + b.listeners = nil + b.alloc = nil + b.closed = true +} + +// stop an individual listener +func (b *AllocBroadcaster) stop(id int) { + b.m.Lock() + defer b.m.Unlock() + + // If broadcaster has been closed there's nothing more to do. + if b.closed { + return + } + + l, ok := b.listeners[id] + if !ok { + // If this listener has been stopped already there's nothing + // more to do. + return + } + + close(l) + delete(b.listeners, id) } // Listen returns a Listener for the broadcast channel. @@ -114,9 +131,17 @@ func (b *AllocBroadcaster) Listen() *AllocListener { return &AllocListener{ch, b, b.nextId} } -// Close closes the Listener, disabling the receival of further messages. -func (l *AllocListener) Close() { - l.b.m.Lock() - defer l.b.m.Unlock() - delete(l.b.listeners, l.id) +// AllocListener implements a listening endpoint for an allocation broadcast +// channel. +type AllocListener struct { + // Ch receives the broadcast messages. + Ch <-chan *structs.Allocation + b *AllocBroadcaster + id int +} + +// Close closes the Listener, disabling the receival of further messages. Safe +// to call more than once and concurrently with receiving on Ch. +func (l *AllocListener) Close() { + l.b.stop(l.id) } diff --git a/client/structs/broadcaster_test.go b/client/structs/broadcaster_test.go new file mode 100644 index 000000000..19c7ace2d --- /dev/null +++ b/client/structs/broadcaster_test.go @@ -0,0 +1,164 @@ +package structs + +import ( + "fmt" + "testing" + "time" + + "github.com/hashicorp/nomad/nomad/mock" + "github.com/stretchr/testify/require" +) + +// TestAllocBroadcaster_SendRecv asserts the initial and latest sends to a +// broadcaster are received by listeners. +func TestAllocBroadcaster_SendRecv(t *testing.T) { + t.Parallel() + + alloc := mock.Alloc() + alloc.AllocModifyIndex = 10 + + b := NewAllocBroadcaster(alloc.Copy()) + defer b.Close() + + // Create a listener and get the initial alloc + l := b.Listen() + defer l.Close() + initial := <-l.Ch + require.Equal(t, alloc.AllocModifyIndex, initial.AllocModifyIndex) + + // Increment the index and send a new copy + alloc.AllocModifyIndex = 20 + require.NoError(t, b.Send(alloc.Copy())) + recvd := <-l.Ch + require.Equal(t, alloc.AllocModifyIndex, recvd.AllocModifyIndex) + + // Send two now copies and assert only the last was received + alloc.AllocModifyIndex = 30 + require.NoError(t, b.Send(alloc.Copy())) + alloc.AllocModifyIndex = 40 + require.NoError(t, b.Send(alloc.Copy())) + + recvd = <-l.Ch + require.Equal(t, alloc.AllocModifyIndex, recvd.AllocModifyIndex) +} + +// TestAllocBroadcaster_RecvBlocks asserts listeners are blocked until a send occurs. +func TestAllocBroadcaster_RecvBlocks(t *testing.T) { + t.Parallel() + + alloc := mock.Alloc() + b := NewAllocBroadcaster(alloc.Copy()) + defer b.Close() + + l1 := b.Listen() + defer l1.Close() + + l2 := b.Listen() + defer l2.Close() + + // Every listener should get the initial alloc even when there hasn't + // been a Send() + require.NotNil(t, <-l1.Ch) + require.NotNil(t, <-l2.Ch) + + done := make(chan int, 2) + + // Subsequent listens should block until a subsequent send + go func() { + <-l1.Ch + done <- 1 + }() + + go func() { + <-l2.Ch + done <- 1 + }() + + select { + case <-done: + t.Fatalf("unexpected receive by a listener") + case <-time.After(10 * time.Millisecond): + } + + // Do a Send and expect both listeners to receive it + b.Send(alloc) + <-done + <-done +} + +// TestAllocBroadcaster_Concurrency asserts that the broadcaster behaves +// correctly with concurrent listeners being added and closed. +func TestAllocBroadcaster_Concurrency(t *testing.T) { + t.Parallel() + + alloc := mock.Alloc() + b := NewAllocBroadcaster(alloc.Copy()) + defer b.Close() + + errs := make(chan error, 10) + listeners := make([]*AllocListener, 10) + for i := 0; i < len(listeners); i++ { + l := b.Listen() + defer l.Close() + + listeners[i] = l + go func(index uint64, listener *AllocListener) { + defer listener.Close() + for { + a, ok := <-listener.Ch + if !ok { + return + } + + if a.AllocModifyIndex < index { + errs <- fmt.Errorf("index=%d < %d", a.AllocModifyIndex, index) + return + } + index = a.AllocModifyIndex + } + }(alloc.AllocModifyIndex, l) + } + + for i := 0; i < 100; i++ { + alloc.AllocModifyIndex++ + require.NoError(t, b.Send(alloc.Copy())) + } + + if len(errs) > 0 { + t.Fatalf("%d listener errors. First error:\n%v", len(errs), <-errs) + } + + // Closing a couple shouldn't cause errors + listeners[0].Close() + listeners[1].Close() + + for i := 0; i < 100; i++ { + alloc.AllocModifyIndex++ + require.NoError(t, b.Send(alloc.Copy())) + } + + if len(errs) > 0 { + t.Fatalf("%d listener errors. First error:\n%v", len(errs), <-errs) + } + + // Closing the broadcaster *should* error + b.Close() + require.Equal(t, ErrAllocBroadcasterClosed, b.Send(alloc)) + + // All Listeners should be closed + for _, l := range listeners { + select { + case _, ok := <-l.Ch: + if ok { + // This check can beat the goroutine above to + // recv'ing the final update. Listener must be + // closed on next recv. + if _, ok := <-l.Ch; ok { + t.Fatalf("expected listener to be closed") + } + } + default: + t.Fatalf("expected listener to be closed; not blocking") + } + } +} From 5c5c6dc41b00e3a1b7a34f9366a65a61dffff208 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 30 Aug 2018 17:13:00 -0700 Subject: [PATCH 054/152] refactor ar hooks into their own files minimize passed dependencies to ease testing --- client/allocrunnerv2/alloc_runner_hooks.go | 77 +--------------------- client/allocrunnerv2/allocdir_hook.go | 35 ++++++++++ client/allocrunnerv2/migrate_hook.go | 56 ++++++++++++++++ 3 files changed, 92 insertions(+), 76 deletions(-) create mode 100644 client/allocrunnerv2/allocdir_hook.go create mode 100644 client/allocrunnerv2/migrate_hook.go diff --git a/client/allocrunnerv2/alloc_runner_hooks.go b/client/allocrunnerv2/alloc_runner_hooks.go index 4ac7b9ab8..198f3229f 100644 --- a/client/allocrunnerv2/alloc_runner_hooks.go +++ b/client/allocrunnerv2/alloc_runner_hooks.go @@ -7,9 +7,7 @@ import ( log "github.com/hashicorp/go-hclog" multierror "github.com/hashicorp/go-multierror" - "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - "github.com/hashicorp/nomad/client/allocwatcher" ) // initRunnerHooks intializes the runners hooks. @@ -19,7 +17,7 @@ func (ar *allocRunner) initRunnerHooks() { // Create the alloc directory hook. This is run first to ensure the // directoy path exists for other hooks. ar.runnerHooks = []interfaces.RunnerHook{ - newAllocDirHook(hookLogger, ar), + newAllocDirHook(hookLogger, ar.allocDir), newDiskMigrationHook(hookLogger, ar.prevAllocWatcher, ar.allocDir), } } @@ -141,79 +139,6 @@ func (ar *allocRunner) destroy() error { return nil } -// allocDirHook creates and destroys the root directory and shared directories -// for an allocation. -type allocDirHook struct { - runner *allocRunner - logger log.Logger -} - -func newAllocDirHook(logger log.Logger, runner *allocRunner) *allocDirHook { - ad := &allocDirHook{ - runner: runner, - } - ad.logger = logger.Named(ad.Name()) - return ad -} - -func (h *allocDirHook) Name() string { - return "alloc_dir" -} - -func (h *allocDirHook) Prerun(context.Context) error { - return h.runner.allocDir.Build() -} - -func (h *allocDirHook) Destroy() error { - return h.runner.allocDir.Destroy() -} - -// diskMigrationHook migrates ephemeral disk volumes. Depends on alloc dir -// being built but must be run before anything else manipulates the alloc dir. -type diskMigrationHook struct { - allocDir *allocdir.AllocDir - allocWatcher allocwatcher.PrevAllocWatcher - logger log.Logger -} - -func newDiskMigrationHook(logger log.Logger, allocWatcher allocwatcher.PrevAllocWatcher, allocDir *allocdir.AllocDir) *diskMigrationHook { - h := &diskMigrationHook{ - allocDir: allocDir, - allocWatcher: allocWatcher, - } - h.logger = logger.Named(h.Name()) - return h -} - -func (h *diskMigrationHook) Name() string { - return "migrate_disk" -} - -func (h *diskMigrationHook) Prerun(ctx context.Context) error { - // Wait for a previous alloc - if any - to terminate - if err := h.allocWatcher.Wait(ctx); err != nil { - return err - } - - // Wait for data to be migrated from a previous alloc if applicable - if err := h.allocWatcher.Migrate(ctx, h.allocDir); err != nil { - if err == context.Canceled { - return err - } - - // Soft-fail on migration errors - h.logger.Warn("error migrating data from previous alloc", "error", err) - - // Recreate alloc dir to ensure a clean slate - h.allocDir.Destroy() - if err := h.allocDir.Build(); err != nil { - return fmt.Errorf("failed to clean task directories after failed migration: %v", err) - } - } - - return nil -} - // TODO type allocHealthWatcherHook struct { runner *allocRunner diff --git a/client/allocrunnerv2/allocdir_hook.go b/client/allocrunnerv2/allocdir_hook.go new file mode 100644 index 000000000..f4331a9da --- /dev/null +++ b/client/allocrunnerv2/allocdir_hook.go @@ -0,0 +1,35 @@ +package allocrunnerv2 + +import ( + "context" + + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocdir" +) + +// allocDirHook creates and destroys the root directory and shared directories +// for an allocation. +type allocDirHook struct { + allocDir *allocdir.AllocDir + logger log.Logger +} + +func newAllocDirHook(logger log.Logger, allocDir *allocdir.AllocDir) *allocDirHook { + ad := &allocDirHook{ + allocDir: allocDir, + } + ad.logger = logger.Named(ad.Name()) + return ad +} + +func (h *allocDirHook) Name() string { + return "alloc_dir" +} + +func (h *allocDirHook) Prerun(context.Context) error { + return h.allocDir.Build() +} + +func (h *allocDirHook) Destroy() error { + return h.allocDir.Destroy() +} diff --git a/client/allocrunnerv2/migrate_hook.go b/client/allocrunnerv2/migrate_hook.go new file mode 100644 index 000000000..c442b693c --- /dev/null +++ b/client/allocrunnerv2/migrate_hook.go @@ -0,0 +1,56 @@ +package allocrunnerv2 + +import ( + "context" + "fmt" + + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocdir" + "github.com/hashicorp/nomad/client/allocwatcher" +) + +// diskMigrationHook migrates ephemeral disk volumes. Depends on alloc dir +// being built but must be run before anything else manipulates the alloc dir. +type diskMigrationHook struct { + allocDir *allocdir.AllocDir + allocWatcher allocwatcher.PrevAllocWatcher + logger log.Logger +} + +func newDiskMigrationHook(logger log.Logger, allocWatcher allocwatcher.PrevAllocWatcher, allocDir *allocdir.AllocDir) *diskMigrationHook { + h := &diskMigrationHook{ + allocDir: allocDir, + allocWatcher: allocWatcher, + } + h.logger = logger.Named(h.Name()) + return h +} + +func (h *diskMigrationHook) Name() string { + return "migrate_disk" +} + +func (h *diskMigrationHook) Prerun(ctx context.Context) error { + // Wait for a previous alloc - if any - to terminate + if err := h.allocWatcher.Wait(ctx); err != nil { + return err + } + + // Wait for data to be migrated from a previous alloc if applicable + if err := h.allocWatcher.Migrate(ctx, h.allocDir); err != nil { + if err == context.Canceled { + return err + } + + // Soft-fail on migration errors + h.logger.Warn("error migrating data from previous alloc", "error", err) + + // Recreate alloc dir to ensure a clean slate + h.allocDir.Destroy() + if err := h.allocDir.Build(); err != nil { + return fmt.Errorf("failed to clean task directories after failed migration: %v", err) + } + } + + return nil +} From 4136e59f7962bdcb1620882a4fc6eeaa24f109f3 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 30 Aug 2018 14:33:50 -0700 Subject: [PATCH 055/152] arv2: implement alloc health watching Also remove initial alloc from broadcaster as it just caused useless extra processing. --- client/allochealth/tracker.go | 472 ++++++++++++++++++ client/allocrunnerv2/alloc_runner.go | 73 ++- client/allocrunnerv2/alloc_runner_hooks.go | 86 ++-- client/allocrunnerv2/health_hook.go | 247 +++++++++ .../interfaces/runner_lifecycle.go | 7 +- client/allocrunnerv2/state/state.go | 26 + .../allocrunnerv2/taskrunner/task_runner.go | 3 +- client/structs/broadcaster.go | 22 +- client/structs/broadcaster_test.go | 33 +- 9 files changed, 874 insertions(+), 95 deletions(-) create mode 100644 client/allochealth/tracker.go create mode 100644 client/allocrunnerv2/health_hook.go diff --git a/client/allochealth/tracker.go b/client/allochealth/tracker.go new file mode 100644 index 000000000..cc865a605 --- /dev/null +++ b/client/allochealth/tracker.go @@ -0,0 +1,472 @@ +package allochealth + +import ( + "context" + "fmt" + "strings" + "sync" + "time" + + "github.com/hashicorp/consul/api" + hclog "github.com/hashicorp/go-hclog" + cconsul "github.com/hashicorp/nomad/client/consul" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/command/agent/consul" + "github.com/hashicorp/nomad/nomad/structs" +) + +const ( + // allocHealthEventSource is the source used for emitting task events + allocHealthEventSource = "Alloc Unhealthy" + + // consulCheckLookupInterval is the interval at which we check if the + // Consul checks are healthy or unhealthy. + consulCheckLookupInterval = 500 * time.Millisecond +) + +// Tracker tracks the health of an allocation and makes health events watchable +// via channels. +type Tracker struct { + // ctx and cancelFn is used to shutdown the tracker + ctx context.Context + cancelFn context.CancelFunc + + // alloc is the alloc we are tracking + alloc *structs.Allocation + + // tg is the task group we are tracking + tg *structs.TaskGroup + + // minHealthyTime is the duration an alloc must remain healthy to be + // considered healthy + minHealthyTime time.Duration + + // useChecks specifies whether to use Consul healh checks or not + useChecks bool + + // consulCheckCount is the number of checks the task group will attempt to + // register + consulCheckCount int + + // allocUpdates is a listener for retrieving new alloc updates + allocUpdates *cstructs.AllocListener + + // consulClient is used to look up the state of the task's checks + consulClient cconsul.ConsulServiceAPI + + // healthy is used to signal whether we have determined the allocation to be + // healthy or unhealthy + healthy chan bool + + // allocStopped is triggered when the allocation is stopped and tracking is + // not needed + allocStopped chan struct{} + + // l is used to lock shared fields listed below + l sync.Mutex + + // tasksHealthy marks whether all the tasks have met their health check + // (disregards Consul) + tasksHealthy bool + + // allocFailed marks whether the allocation failed + allocFailed bool + + // checksHealthy marks whether all the task's Consul checks are healthy + checksHealthy bool + + // taskHealth contains the health state for each task + taskHealth map[string]*taskHealthState + + logger hclog.Logger +} + +// NewTracker returns a health tracker for the given allocation. An alloc +// listener and consul API object are given so that the watcher can detect +// health changes. +func NewTracker(parentCtx context.Context, logger hclog.Logger, alloc *structs.Allocation, + allocUpdates *cstructs.AllocListener, consulClient cconsul.ConsulServiceAPI, + minHealthyTime time.Duration, useChecks bool) *Tracker { + + // Do not create a named sub-logger as the hook controlling + // this struct should pass in an appropriately named + // sub-logger. + t := &Tracker{ + healthy: make(chan bool, 1), + allocStopped: make(chan struct{}), + alloc: alloc, + tg: alloc.Job.LookupTaskGroup(alloc.TaskGroup), + minHealthyTime: minHealthyTime, + useChecks: useChecks, + allocUpdates: allocUpdates, + consulClient: consulClient, + logger: logger, + } + + t.taskHealth = make(map[string]*taskHealthState, len(t.tg.Tasks)) + for _, task := range t.tg.Tasks { + t.taskHealth[task.Name] = &taskHealthState{task: task} + } + + for _, task := range t.tg.Tasks { + for _, s := range task.Services { + t.consulCheckCount += len(s.Checks) + } + } + + t.ctx, t.cancelFn = context.WithCancel(parentCtx) + return t +} + +// Start starts the watcher. +func (t *Tracker) Start() { + go t.watchTaskEvents() + if t.useChecks { + go t.watchConsulEvents() + } +} + +// HealthyCh returns a channel that will emit a boolean indicating the health of +// the allocation. +func (t *Tracker) HealthyCh() <-chan bool { + return t.healthy +} + +// AllocStoppedCh returns a channel that will be fired if the allocation is +// stopped. This means that health will not be set. +func (t *Tracker) AllocStoppedCh() <-chan struct{} { + return t.allocStopped +} + +// TaskEvents returns a map of events by task. This should only be called after +// health has been determined. Only tasks that have contributed to the +// allocation being unhealthy will have an event. +func (t *Tracker) TaskEvents() map[string]*structs.TaskEvent { + t.l.Lock() + defer t.l.Unlock() + + // Nothing to do since the failure wasn't task related + if t.allocFailed { + return nil + } + + deadline, _ := t.ctx.Deadline() + events := make(map[string]*structs.TaskEvent, len(t.tg.Tasks)) + + // Go through are task information and build the event map + for task, state := range t.taskHealth { + useChecks := t.tg.Update.HealthCheck == structs.UpdateStrategyHealthCheck_Checks + if e, ok := state.event(deadline, t.tg.Update.MinHealthyTime, useChecks); ok { + events[task] = structs.NewTaskEvent(allocHealthEventSource).SetMessage(e) + } + } + + return events +} + +// setTaskHealth is used to set the tasks health as healthy or unhealthy. If the +// allocation is terminal, health is immediately broadcasted. +func (t *Tracker) setTaskHealth(healthy, terminal bool) { + t.l.Lock() + defer t.l.Unlock() + t.tasksHealthy = healthy + + // If we are marked healthy but we also require Consul to be healthy and it + // isn't yet, return, unless the task is terminal + requireConsul := t.useChecks && t.consulCheckCount > 0 + if !terminal && healthy && requireConsul && !t.checksHealthy { + return + } + + select { + case t.healthy <- healthy: + default: + } + + // Shutdown the tracker + t.cancelFn() +} + +// setCheckHealth is used to mark the checks as either healthy or unhealthy. +func (t *Tracker) setCheckHealth(healthy bool) { + t.l.Lock() + defer t.l.Unlock() + t.checksHealthy = healthy + + // Only signal if we are healthy and so is the tasks + if !healthy || !t.tasksHealthy { + return + } + + select { + case t.healthy <- healthy: + default: + } + + // Shutdown the tracker + t.cancelFn() +} + +// markAllocStopped is used to mark the allocation as having stopped. +func (t *Tracker) markAllocStopped() { + close(t.allocStopped) + t.cancelFn() +} + +// watchTaskEvents is a long lived watcher that watches for the health of the +// allocation's tasks. +func (t *Tracker) watchTaskEvents() { + alloc := t.alloc + allStartedTime := time.Time{} + healthyTimer := time.NewTimer(0) + if !healthyTimer.Stop() { + select { + case <-healthyTimer.C: + default: + } + } + + for { + // If the alloc is being stopped by the server just exit + switch alloc.DesiredStatus { + case structs.AllocDesiredStatusStop, structs.AllocDesiredStatusEvict: + t.logger.Trace("desired status is terminal for alloc", "alloc_id", alloc.ID, "desired_status", alloc.DesiredStatus) + t.markAllocStopped() + return + } + + // Store the task states + t.l.Lock() + for task, state := range alloc.TaskStates { + t.taskHealth[task].state = state + } + t.l.Unlock() + + // Detect if the alloc is unhealthy or if all tasks have started yet + latestStartTime := time.Time{} + for _, state := range alloc.TaskStates { + // One of the tasks has failed so we can exit watching + if state.Failed || !state.FinishedAt.IsZero() { + t.setTaskHealth(false, true) + return + } + + if state.State != structs.TaskStateRunning { + latestStartTime = time.Time{} + break + } else if state.StartedAt.After(latestStartTime) { + latestStartTime = state.StartedAt + } + } + + // If the alloc is marked as failed by the client but none of the + // individual tasks failed, that means something failed at the alloc + // level. + if alloc.ClientStatus == structs.AllocClientStatusFailed { + t.l.Lock() + t.allocFailed = true + t.l.Unlock() + t.setTaskHealth(false, true) + return + } + + if !latestStartTime.Equal(allStartedTime) { + // Avoid the timer from firing at the old start time + if !healthyTimer.Stop() { + select { + case <-healthyTimer.C: + default: + } + } + + // Set the timer since all tasks are started + if !latestStartTime.IsZero() { + allStartedTime = latestStartTime + healthyTimer.Reset(t.minHealthyTime) + } + } + + select { + case <-t.ctx.Done(): + return + case newAlloc, ok := <-t.allocUpdates.Ch: + if !ok { + return + } + alloc = newAlloc + case <-healthyTimer.C: + t.setTaskHealth(true, false) + } + } +} + +// watchConsulEvents iis a long lived watcher that watches for the health of the +// allocation's Consul checks. +func (t *Tracker) watchConsulEvents() { + // checkTicker is the ticker that triggers us to look at the checks in + // Consul + checkTicker := time.NewTicker(consulCheckLookupInterval) + defer checkTicker.Stop() + + // healthyTimer fires when the checks have been healthy for the + // MinHealthyTime + healthyTimer := time.NewTimer(0) + if !healthyTimer.Stop() { + select { + case <-healthyTimer.C: + default: + } + } + + // primed marks whether the healthy timer has been set + primed := false + + // Store whether the last Consul checks call was successful or not + consulChecksErr := false + + // allocReg are the registered objects in Consul for the allocation + var allocReg *consul.AllocRegistration + +OUTER: + for { + select { + case <-t.ctx.Done(): + return + case <-checkTicker.C: + newAllocReg, err := t.consulClient.AllocRegistrations(t.alloc.ID) + if err != nil { + if !consulChecksErr { + consulChecksErr = true + t.logger.Warn("error looking up Consul registrations for allocation", "error", err, "alloc_id", t.alloc.ID) + } + continue OUTER + } else { + consulChecksErr = false + allocReg = newAllocReg + } + case <-healthyTimer.C: + t.setCheckHealth(true) + } + + if allocReg == nil { + continue + } + + // Store the task registrations + t.l.Lock() + for task, reg := range allocReg.Tasks { + t.taskHealth[task].taskRegistrations = reg + } + t.l.Unlock() + + // Detect if all the checks are passing + passed := true + + CHECKS: + for _, treg := range allocReg.Tasks { + for _, sreg := range treg.Services { + for _, check := range sreg.Checks { + if check.Status == api.HealthPassing { + continue + } + + passed = false + t.setCheckHealth(false) + break CHECKS + } + } + } + + if !passed { + // Reset the timer since we have transitioned back to unhealthy + if primed { + if !healthyTimer.Stop() { + select { + case <-healthyTimer.C: + default: + } + } + primed = false + } + } else if !primed { + // Reset the timer to fire after MinHealthyTime + if !healthyTimer.Stop() { + select { + case <-healthyTimer.C: + default: + } + } + + primed = true + healthyTimer.Reset(t.minHealthyTime) + } + } +} + +// taskHealthState captures all known health information about a task. It is +// largely used to determine if the task has contributed to the allocation being +// unhealthy. +type taskHealthState struct { + task *structs.Task + state *structs.TaskState + taskRegistrations *consul.TaskRegistration +} + +// event takes the deadline time for the allocation to be healthy and the update +// strategy of the group. It returns true if the task has contributed to the +// allocation being unhealthy and if so, an event description of why. +func (t *taskHealthState) event(deadline time.Time, minHealthyTime time.Duration, useChecks bool) (string, bool) { + requireChecks := false + desiredChecks := 0 + for _, s := range t.task.Services { + if nc := len(s.Checks); nc > 0 { + requireChecks = true + desiredChecks += nc + } + } + requireChecks = requireChecks && useChecks + + if t.state != nil { + if t.state.Failed { + return "Unhealthy because of failed task", true + } + if t.state.State != structs.TaskStateRunning { + return "Task not running by deadline", true + } + + // We are running so check if we have been running long enough + if t.state.StartedAt.Add(minHealthyTime).After(deadline) { + return fmt.Sprintf("Task not running for min_healthy_time of %v by deadline", minHealthyTime), true + } + } + + if t.taskRegistrations != nil { + var notPassing []string + passing := 0 + + OUTER: + for _, sreg := range t.taskRegistrations.Services { + for _, check := range sreg.Checks { + if check.Status != api.HealthPassing { + notPassing = append(notPassing, sreg.Service.Service) + continue OUTER + } else { + passing++ + } + } + } + + if len(notPassing) != 0 { + return fmt.Sprintf("Services not healthy by deadline: %s", strings.Join(notPassing, ", ")), true + } + + if passing != desiredChecks { + return fmt.Sprintf("Only %d out of %d checks registered and passing", passing, desiredChecks), true + } + + } else if requireChecks { + return "Service checks not registered", true + } + + return "", false +} diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index ec8fd3c46..9bf7167c4 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -24,14 +24,6 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) -const ( - // updateChCap is the capacity of AllocRunner's updateCh. It must be 1 - // as we only want to process the latest update, so if there's already - // a pending update it will be removed from the chan before adding the - // newer update. - updateChCap = 1 -) - // allocRunner is used to run all the tasks in a given allocation type allocRunner struct { // id is the ID of the allocation. Can be accessed without a lock @@ -82,11 +74,6 @@ type allocRunner struct { tasks map[string]*taskrunner.TaskRunner tasksLock sync.RWMutex - // updateCh receives allocation updates via the Update method. Must - // have buffer size 1 in order to support dropping pending updates when - // a newer allocation is received. - updateCh chan *structs.Allocation - // allocBroadcaster sends client allocation updates to all listeners allocBroadcaster *cstructs.AllocBroadcaster @@ -111,16 +98,15 @@ func NewAllocRunner(config *Config) (*allocRunner, error) { vaultClient: config.Vault, tasks: make(map[string]*taskrunner.TaskRunner, len(tg.Tasks)), waitCh: make(chan struct{}), - updateCh: make(chan *structs.Allocation, updateChCap), state: &state.State{}, stateDB: config.StateDB, stateUpdater: config.StateUpdater, - allocBroadcaster: cstructs.NewAllocBroadcaster(alloc), + allocBroadcaster: cstructs.NewAllocBroadcaster(), prevAllocWatcher: config.PrevAllocWatcher, } // Create the logger based on the allocation ID - ar.logger = config.Logger.With("alloc_id", alloc.ID) + ar.logger = config.Logger.Named("alloc_runner").With("alloc_id", alloc.ID) // Create alloc dir ar.allocDir = allocdir.NewAllocDir(ar.logger, filepath.Join(config.ClientConfig.AllocDir, alloc.ID)) @@ -251,6 +237,56 @@ func (ar *allocRunner) Restore() error { return nil } +// SetHealth allows the health watcher hook to set the alloc's +// deployment/migration health and emit task events. +// +// Only for use by health hook. +func (ar *allocRunner) SetHealth(healthy, isDeploy bool, trackerTaskEvents map[string]*structs.TaskEvent) { + // Updating alloc deployment state is tricky because it may be nil, but + // if it's not then we need to maintain the values of Canary and + // ModifyIndex as they're only mutated by the server. + ar.stateLock.Lock() + ar.state.SetDeploymentStatus(time.Now(), healthy) + ar.stateLock.Unlock() + + // If deployment is unhealthy emit task events explaining why + ar.tasksLock.RLock() + if !healthy && isDeploy { + for task, event := range trackerTaskEvents { + if tr, ok := ar.tasks[task]; ok { + tr.EmitEvent(event) + } + } + } + + // Gather the state of the other tasks + states := make(map[string]*structs.TaskState, len(ar.tasks)) + for name, tr := range ar.tasks { + states[name] = tr.TaskState() + } + ar.tasksLock.RUnlock() + + // Build the client allocation + calloc := ar.clientAlloc(states) + + // Update the server + ar.stateUpdater.AllocStateUpdated(calloc) + + // Broadcast client alloc to listeners + ar.allocBroadcaster.Send(calloc) +} + +// ClearHealth allows the health watcher hook to clear the alloc's deployment +// health if the deployment id changes. It does not update the server as the +// status is only cleared when already receiving an update from the server. +// +// Only for use by health hook. +func (ar *allocRunner) ClearHealth() { + ar.stateLock.Lock() + ar.state.ClearDeploymentStatus() + ar.stateLock.Unlock() +} + // TaskStateUpdated is called by TaskRunner when a task's state has been // updated. This hook is used to compute changes to the alloc's ClientStatus // and to update the server with the new state. @@ -408,7 +444,10 @@ func (ar *allocRunner) Update(update *structs.Allocation) { // Update ar.alloc ar.setAlloc(update) - //TODO Run AR Update hooks + // Run hooks + if err := ar.update(update); err != nil { + ar.logger.Error("error running update hooks", "error", err) + } // Update task runners for _, tr := range ar.tasks { diff --git a/client/allocrunnerv2/alloc_runner_hooks.go b/client/allocrunnerv2/alloc_runner_hooks.go index 198f3229f..a236f7fa9 100644 --- a/client/allocrunnerv2/alloc_runner_hooks.go +++ b/client/allocrunnerv2/alloc_runner_hooks.go @@ -5,9 +5,9 @@ import ( "fmt" "time" - log "github.com/hashicorp/go-hclog" multierror "github.com/hashicorp/go-multierror" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/nomad/structs" ) // initRunnerHooks intializes the runners hooks. @@ -19,6 +19,7 @@ func (ar *allocRunner) initRunnerHooks() { ar.runnerHooks = []interfaces.RunnerHook{ newAllocDirHook(hookLogger, ar.allocDir), newDiskMigrationHook(hookLogger, ar.prevAllocWatcher, ar.allocDir), + newAllocHealthWatcherHook(hookLogger, ar, ar.consulClient), } } @@ -63,6 +64,48 @@ func (ar *allocRunner) prerun() error { return nil } +// update runs the alloc runner update hooks. Update hooks are run +// asynchronously with all other alloc runner operations. +func (ar *allocRunner) update(update *structs.Allocation) error { + if ar.logger.IsTrace() { + start := time.Now() + ar.logger.Trace("running update hooks", "start", start) + defer func() { + end := time.Now() + ar.logger.Trace("finished update hooks", "end", end, "duration", end.Sub(start)) + }() + } + + req := &interfaces.RunnerUpdateRequest{ + Alloc: update, + } + + for _, hook := range ar.runnerHooks { + h, ok := hook.(interfaces.RunnerUpdateHook) + if !ok { + continue + } + + name := h.Name() + var start time.Time + if ar.logger.IsTrace() { + start = time.Now() + ar.logger.Trace("running pre-run hook", "name", name, "start", start) + } + + if err := h.Update(req); err != nil { + return fmt.Errorf("hook %q failed: %v", name, err) + } + + if ar.logger.IsTrace() { + end := time.Now() + ar.logger.Trace("finished update hooks", "name", name, "end", end, "duration", end.Sub(start)) + } + } + + return nil +} + // postrun is used to run the runners postrun hooks. func (ar *allocRunner) postrun() error { if ar.logger.IsTrace() { @@ -138,44 +181,3 @@ func (ar *allocRunner) destroy() error { return nil } - -// TODO -type allocHealthWatcherHook struct { - runner *allocRunner - logger log.Logger - ctx context.Context - cancelFn context.CancelFunc -} - -func newAllocHealthWatcherHook(runner *allocRunner, logger log.Logger) *allocHealthWatcherHook { - ctx, cancelFn := context.WithCancel(context.Background()) - ad := &allocHealthWatcherHook{ - runner: runner, - ctx: ctx, - cancelFn: cancelFn, - } - - ad.logger = logger.Named(ad.Name()) - return ad -} - -func (h *allocHealthWatcherHook) Name() string { - return "alloc_health_watcher" -} - -func (h *allocHealthWatcherHook) Prerun() error { - return nil -} - -func (h *allocHealthWatcherHook) Update() error { - // Cancel the old watcher and create a new one - h.cancelFn() - - // TODO create the new one - return nil -} - -func (h *allocHealthWatcherHook) Destroy() error { - h.cancelFn() - return nil -} diff --git a/client/allocrunnerv2/health_hook.go b/client/allocrunnerv2/health_hook.go new file mode 100644 index 000000000..fb2857d62 --- /dev/null +++ b/client/allocrunnerv2/health_hook.go @@ -0,0 +1,247 @@ +package allocrunnerv2 + +import ( + "context" + "fmt" + "sync" + "time" + + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allochealth" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/consul" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/nomad/structs" +) + +// healthMutator is able to set/clear alloc health as well as listen for alloc +// changes. +type healthMutator interface { + // Alloc returns the original alloc + Alloc() *structs.Allocation + + // Set health via the mutator + SetHealth(healthy, isDeploy bool, taskEvents map[string]*structs.TaskEvent) + + // Clear health when the deployment ID changes + ClearHealth() + + // Listener listens for alloc updates + Listener() *cstructs.AllocListener +} + +// allocHealthWatcherHook is responsible for watching an allocation's task +// status and (optionally) Consul health check status to determine if the +// allocation is health or unhealthy. Used by deployments and migrations. +type allocHealthWatcherHook struct { + runner healthMutator + + // consul client used to monitor health checks + consul consul.ConsulServiceAPI + + // listener is given to trackers to listen for alloc updates and closed + // when the alloc is destroyed. + listener *cstructs.AllocListener + + // hookLock is held by hook methods to prevent concurrent access by + // Update and synchronous hooks. + hookLock sync.Mutex + + // watchLock is held by the health watching/setting goroutine to ensure + // only one health watching goroutine is running at a time. + watchLock sync.Mutex + + // ranOnce is set once Prerun or Update have run at least once. This + // prevents Prerun from running if an Update has already been + // processed. Must hold hookLock to access. + ranOnce bool + + // cancelFn stops the health watching/setting goroutine. Grab the + // watchLock to block until it exits. + cancelFn context.CancelFunc + + // alloc set by new func or Update. Must hold hookLock to access. + alloc *structs.Allocation + + // isDeploy is true if monitoring a deployment. Set in init(). Must + // hold hookLock to access. + isDeploy bool + + logger log.Logger +} + +func newAllocHealthWatcherHook(logger log.Logger, runner healthMutator, consul consul.ConsulServiceAPI) interfaces.RunnerHook { + alloc := runner.Alloc() + + // Neither deployments nor migrations care about the health of + // non-service jobs so never watch their health + if alloc.Job.Type != structs.JobTypeService { + return noopAllocHealthWatcherHook{} + } + + h := &allocHealthWatcherHook{ + runner: runner, + alloc: alloc, + cancelFn: func() {}, // initialize to prevent nil func panics + consul: consul, + listener: runner.Listener(), + } + + h.logger = logger.Named(h.Name()) + return h +} + +func (h *allocHealthWatcherHook) Name() string { + return "alloc_health_watcher" +} + +// init starts the allochealth.Tracker and watchHealth goroutine on either +// Prerun or Update. Caller must set/update alloc and logger fields. +// +// Not threadsafe so the caller should lock since Updates occur concurrently. +func (h *allocHealthWatcherHook) init() error { + // No need to watch health as it's already set + if h.alloc.DeploymentStatus.HasHealth() { + return nil + } + + tg := h.alloc.Job.LookupTaskGroup(h.alloc.TaskGroup) + if tg == nil { + return fmt.Errorf("task group %q does not exist in job %q", h.alloc.TaskGroup, h.alloc.Job.ID) + } + + h.isDeploy = h.alloc.DeploymentID != "" + + // No need to watch allocs for deployments that rely on operators + // manually setting health + if h.isDeploy && (tg.Update == nil || tg.Update.HealthCheck == structs.UpdateStrategyHealthCheck_Manual) { + return nil + } + + // Define the deadline, health method, min healthy time from the + // deployment if this is a deployment; otherwise from the migration + // strategy. + deadline, useChecks, minHealthyTime := getHealthParams(time.Now(), tg, h.isDeploy) + + // Create a context that is canceled when the tracker should shutdown + // or the deadline is reached. + ctx := context.Background() + ctx, h.cancelFn = context.WithDeadline(ctx, deadline) + + // Create a new tracker, start it, and watch for health results. + tracker := allochealth.NewTracker(ctx, h.logger, h.alloc, + h.listener, h.consul, minHealthyTime, useChecks) + tracker.Start() + go h.watchHealth(ctx, tracker) + return nil +} + +func (h *allocHealthWatcherHook) Prerun(context.Context) error { + h.hookLock.Lock() + defer h.hookLock.Unlock() + + if h.ranOnce { + // An Update beat Prerun to running the watcher; noop + return nil + } + + h.ranOnce = true + return h.init() +} + +func (h *allocHealthWatcherHook) Update(req *interfaces.RunnerUpdateRequest) error { + h.hookLock.Lock() + defer h.hookLock.Unlock() + + // Prevent Prerun from running after an Update + h.ranOnce = true + + // Cancel the old watcher and create a new one + h.cancelFn() + + // Acquire the watchLock to ensure the old watcher has exited before + // continuing. Kind of an ugly/easy-to-reuse done chan. + h.watchLock.Lock() + h.watchLock.Unlock() + + // Deployment has changed, reset status + if req.Alloc.DeploymentID != h.alloc.DeploymentID { + h.runner.ClearHealth() + } + + // Update alloc + h.alloc = req.Alloc + + return h.init() +} + +func (h *allocHealthWatcherHook) Destroy() error { + h.hookLock.Lock() + defer h.hookLock.Unlock() + + h.cancelFn() + h.listener.Close() + + // Acquire the watchLock to ensure any existing watcher has exited + // before exiting. Kind of an ugly/easy-to-reuse done chan. + h.watchLock.Lock() + h.watchLock.Unlock() + + return nil +} + +// watchHealth watches alloc health until it is set, the alloc is stopped, or +// the context is canceled. watchHealth will be canceled and restarted on +// Updates so calls are serialized with a lock. +func (h *allocHealthWatcherHook) watchHealth(ctx context.Context, tracker *allochealth.Tracker) { + h.watchLock.Lock() + defer h.watchLock.Unlock() + + select { + case <-ctx.Done(): + return + + case <-tracker.AllocStoppedCh(): + return + + case healthy := <-tracker.HealthyCh(): + // If this is an unhealthy deployment emit events for tasks + var taskEvents map[string]*structs.TaskEvent + if !healthy && h.isDeploy { + taskEvents = tracker.TaskEvents() + } + + h.runner.SetHealth(healthy, h.isDeploy, taskEvents) + } +} + +// getHealthParams returns the health watcher parameters which vary based on +// whether this allocation is in a deployment or migration. +func getHealthParams(now time.Time, tg *structs.TaskGroup, isDeploy bool) (deadline time.Time, useChecks bool, minHealthyTime time.Duration) { + if isDeploy { + deadline = now.Add(tg.Update.HealthyDeadline) + minHealthyTime = tg.Update.MinHealthyTime + useChecks = tg.Update.HealthCheck == structs.UpdateStrategyHealthCheck_Checks + } else { + strategy := tg.Migrate + if strategy == nil { + // For backwards compat with pre-0.8 allocations that + // don't have a migrate strategy set. + strategy = structs.DefaultMigrateStrategy() + } + + deadline = now.Add(strategy.HealthyDeadline) + minHealthyTime = strategy.MinHealthyTime + useChecks = strategy.HealthCheck == structs.MigrateStrategyHealthChecks + } + return +} + +// noopAllocHealthWatcherHook is an empty hook implementation returned by +// newAllocHealthWatcherHook when an allocation will never need its health +// monitored. +type noopAllocHealthWatcherHook struct{} + +func (noopAllocHealthWatcherHook) Name() string { + return "alloc_health_watcher" +} diff --git a/client/allocrunnerv2/interfaces/runner_lifecycle.go b/client/allocrunnerv2/interfaces/runner_lifecycle.go index 63164dce9..f3150e60c 100644 --- a/client/allocrunnerv2/interfaces/runner_lifecycle.go +++ b/client/allocrunnerv2/interfaces/runner_lifecycle.go @@ -4,6 +4,7 @@ import ( "context" "github.com/hashicorp/nomad/client/allocrunnerv2/state" + "github.com/hashicorp/nomad/nomad/structs" ) // RunnnerHook is a lifecycle hook into the life cycle of an allocation runner. @@ -28,7 +29,11 @@ type RunnerDestroyHook interface { type RunnerUpdateHook interface { RunnerHook - Update() error + Update(*RunnerUpdateRequest) error +} + +type RunnerUpdateRequest struct { + Alloc *structs.Allocation } // XXX Not sure yet diff --git a/client/allocrunnerv2/state/state.go b/client/allocrunnerv2/state/state.go index 73310b7c1..85ec7c3da 100644 --- a/client/allocrunnerv2/state/state.go +++ b/client/allocrunnerv2/state/state.go @@ -1,6 +1,8 @@ package state import ( + "time" + "github.com/hashicorp/nomad/nomad/structs" ) @@ -17,3 +19,27 @@ type State struct { // DeploymentStatus captures the status of the deployment DeploymentStatus *structs.AllocDeploymentStatus } + +// SetDeploymentStatus is a helper for updating the client-controlled +// DeploymentStatus fields: Healthy and Timestamp. The Canary and ModifyIndex +// fields should only be updated by the server. +func (s *State) SetDeploymentStatus(timestamp time.Time, healthy bool) { + if s.DeploymentStatus == nil { + s.DeploymentStatus = &structs.AllocDeploymentStatus{} + } + + s.DeploymentStatus.Healthy = &healthy + s.DeploymentStatus.Timestamp = timestamp +} + +// ClearDeploymentStatus is a helper to clear the client-controlled +// DeploymentStatus fields: Healthy and Timestamp. The Canary and ModifyIndex +// fields should only be updated by the server. +func (s *State) ClearDeploymentStatus() { + if s.DeploymentStatus == nil { + return + } + + s.DeploymentStatus.Healthy = nil + s.DeploymentStatus.Timestamp = time.Time{} +} diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 634a4f804..5bdc21c79 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -483,7 +483,8 @@ func (tr *TaskRunner) Restore() error { return nil } -// SetState sets the task runners allocation state. +// SetState sets the task runners allocation state and triggers a server +// update. func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { // Update the local state stateCopy := tr.setStateLocal(state, event) diff --git a/client/structs/broadcaster.go b/client/structs/broadcaster.go index 88b51ebf8..b3aa2fb3d 100644 --- a/client/structs/broadcaster.go +++ b/client/structs/broadcaster.go @@ -23,18 +23,14 @@ var ErrAllocBroadcasterClosed = errors.New("alloc broadcaster closed") // receive the latest allocation update -- never a stale version. type AllocBroadcaster struct { m sync.Mutex - alloc *structs.Allocation listeners map[int]chan *structs.Allocation // lazy init nextId int closed bool } -// NewAllocBroadcaster returns a new AllocBroadcaster with the given initial -// allocation. -func NewAllocBroadcaster(initial *structs.Allocation) *AllocBroadcaster { - return &AllocBroadcaster{ - alloc: initial, - } +// NewAllocBroadcaster returns a new AllocBroadcaster. +func NewAllocBroadcaster() *AllocBroadcaster { + return &AllocBroadcaster{} } // Send broadcasts an allocation update. Any pending updates are replaced with @@ -47,9 +43,6 @@ func (b *AllocBroadcaster) Send(v *structs.Allocation) error { return ErrAllocBroadcasterClosed } - // Update alloc on broadcaster to send to newly created listeners - b.alloc = v - // Send alloc to already created listeners for _, l := range b.listeners { select { @@ -64,7 +57,8 @@ func (b *AllocBroadcaster) Send(v *structs.Allocation) error { } // Close closes the channel, disabling the sending of further allocation -// updates. Safe to call concurrently and more than once. +// updates. Pending updates are still received by listeners. Safe to call +// concurrently and more than once. func (b *AllocBroadcaster) Close() { b.m.Lock() defer b.m.Unlock() @@ -79,7 +73,6 @@ func (b *AllocBroadcaster) Close() { // Clear all references and mark broadcaster as closed b.listeners = nil - b.alloc = nil b.closed = true } @@ -118,12 +111,9 @@ func (b *AllocBroadcaster) Listen() *AllocListener { ch := make(chan *structs.Allocation, listenerCap) + // Broadcaster is already closed, close this listener if b.closed { - // Broadcaster is already closed, close this listener close(ch) - } else { - // Send the current allocation to the listener - ch <- b.alloc } b.listeners[b.nextId] = ch diff --git a/client/structs/broadcaster_test.go b/client/structs/broadcaster_test.go index 19c7ace2d..ac91bf226 100644 --- a/client/structs/broadcaster_test.go +++ b/client/structs/broadcaster_test.go @@ -9,25 +9,27 @@ import ( "github.com/stretchr/testify/require" ) -// TestAllocBroadcaster_SendRecv asserts the initial and latest sends to a -// broadcaster are received by listeners. +// TestAllocBroadcaster_SendRecv asserts the latest sends to a broadcaster are +// received by listeners. func TestAllocBroadcaster_SendRecv(t *testing.T) { t.Parallel() - alloc := mock.Alloc() - alloc.AllocModifyIndex = 10 - - b := NewAllocBroadcaster(alloc.Copy()) + b := NewAllocBroadcaster() defer b.Close() - // Create a listener and get the initial alloc + // Create a listener and assert it blocks until an update l := b.Listen() defer l.Close() - initial := <-l.Ch - require.Equal(t, alloc.AllocModifyIndex, initial.AllocModifyIndex) + select { + case <-l.Ch: + t.Fatalf("unexpected initial alloc") + case <-time.After(10 * time.Millisecond): + // Ok! Ch is empty until a Send + } - // Increment the index and send a new copy - alloc.AllocModifyIndex = 20 + // Send an update + alloc := mock.Alloc() + alloc.AllocModifyIndex = 10 require.NoError(t, b.Send(alloc.Copy())) recvd := <-l.Ch require.Equal(t, alloc.AllocModifyIndex, recvd.AllocModifyIndex) @@ -47,7 +49,7 @@ func TestAllocBroadcaster_RecvBlocks(t *testing.T) { t.Parallel() alloc := mock.Alloc() - b := NewAllocBroadcaster(alloc.Copy()) + b := NewAllocBroadcaster() defer b.Close() l1 := b.Listen() @@ -56,11 +58,6 @@ func TestAllocBroadcaster_RecvBlocks(t *testing.T) { l2 := b.Listen() defer l2.Close() - // Every listener should get the initial alloc even when there hasn't - // been a Send() - require.NotNil(t, <-l1.Ch) - require.NotNil(t, <-l2.Ch) - done := make(chan int, 2) // Subsequent listens should block until a subsequent send @@ -92,7 +89,7 @@ func TestAllocBroadcaster_Concurrency(t *testing.T) { t.Parallel() alloc := mock.Alloc() - b := NewAllocBroadcaster(alloc.Copy()) + b := NewAllocBroadcaster() defer b.Close() errs := make(chan error, 10) From 4d5aaac6d2df52cea3864e39c5ff9af2b72f3699 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 6 Sep 2018 17:46:32 -0700 Subject: [PATCH 056/152] fix detection of task transitioning to running --- client/allocrunnerv2/taskrunner/task_runner.go | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 5bdc21c79..7d62e9c7c 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -486,6 +486,7 @@ func (tr *TaskRunner) Restore() error { // SetState sets the task runners allocation state and triggers a server // update. func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { + tr.logger.Debug("setting task state", "state", state, "event", event.Type) // Update the local state stateCopy := tr.setStateLocal(state, event) @@ -505,6 +506,7 @@ func (tr *TaskRunner) setStateLocal(state string, event *structs.TaskEvent) *str } // Update the task state + oldState := tr.state.State taskState := tr.state taskState.State = state @@ -515,7 +517,7 @@ func (tr *TaskRunner) setStateLocal(state string, event *structs.TaskEvent) *str switch state { case structs.TaskStateRunning: // Capture the start time if it is just starting - if taskState.State != structs.TaskStateRunning { + if oldState != structs.TaskStateRunning { taskState.StartedAt = time.Now().UTC() if !tr.clientConfig.DisableTaggedMetrics { metrics.IncrCounterWithLabels([]string{"client", "allocs", "running"}, 1, tr.baseLabels) From 93f3ac9ed67e765c707b4edba9c80067d9f08e4a Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 10 Sep 2018 17:07:59 -0700 Subject: [PATCH 057/152] ar: create health setting shim for health watcher --- client/allocrunnerv2/alloc_runner.go | 50 --------------- client/allocrunnerv2/alloc_runner_hooks.go | 61 ++++++++++++++++++- client/allocrunnerv2/health_hook.go | 31 ++++------ .../allocrunnerv2/taskrunner/task_runner.go | 2 +- 4 files changed, 73 insertions(+), 71 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 9bf7167c4..dc65e180b 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -237,56 +237,6 @@ func (ar *allocRunner) Restore() error { return nil } -// SetHealth allows the health watcher hook to set the alloc's -// deployment/migration health and emit task events. -// -// Only for use by health hook. -func (ar *allocRunner) SetHealth(healthy, isDeploy bool, trackerTaskEvents map[string]*structs.TaskEvent) { - // Updating alloc deployment state is tricky because it may be nil, but - // if it's not then we need to maintain the values of Canary and - // ModifyIndex as they're only mutated by the server. - ar.stateLock.Lock() - ar.state.SetDeploymentStatus(time.Now(), healthy) - ar.stateLock.Unlock() - - // If deployment is unhealthy emit task events explaining why - ar.tasksLock.RLock() - if !healthy && isDeploy { - for task, event := range trackerTaskEvents { - if tr, ok := ar.tasks[task]; ok { - tr.EmitEvent(event) - } - } - } - - // Gather the state of the other tasks - states := make(map[string]*structs.TaskState, len(ar.tasks)) - for name, tr := range ar.tasks { - states[name] = tr.TaskState() - } - ar.tasksLock.RUnlock() - - // Build the client allocation - calloc := ar.clientAlloc(states) - - // Update the server - ar.stateUpdater.AllocStateUpdated(calloc) - - // Broadcast client alloc to listeners - ar.allocBroadcaster.Send(calloc) -} - -// ClearHealth allows the health watcher hook to clear the alloc's deployment -// health if the deployment id changes. It does not update the server as the -// status is only cleared when already receiving an update from the server. -// -// Only for use by health hook. -func (ar *allocRunner) ClearHealth() { - ar.stateLock.Lock() - ar.state.ClearDeploymentStatus() - ar.stateLock.Unlock() -} - // TaskStateUpdated is called by TaskRunner when a task's state has been // updated. This hook is used to compute changes to the alloc's ClientStatus // and to update the server with the new state. diff --git a/client/allocrunnerv2/alloc_runner_hooks.go b/client/allocrunnerv2/alloc_runner_hooks.go index a236f7fa9..d5be55f6a 100644 --- a/client/allocrunnerv2/alloc_runner_hooks.go +++ b/client/allocrunnerv2/alloc_runner_hooks.go @@ -10,16 +10,75 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) +// allocHealthSetter is a shim to allow the alloc health watcher hook to set +// and clear the alloc health without full access to the alloc runner state +type allocHealthSetter struct { + ar *allocRunner +} + +// ClearHealth allows the health watcher hook to clear the alloc's deployment +// health if the deployment id changes. It does not update the server as the +// status is only cleared when already receiving an update from the server. +// +// Only for use by health hook. +func (a *allocHealthSetter) ClearHealth() { + a.ar.stateLock.Lock() + a.ar.state.ClearDeploymentStatus() + a.ar.stateLock.Unlock() +} + +// SetHealth allows the health watcher hook to set the alloc's +// deployment/migration health and emit task events. +// +// Only for use by health hook. +func (a *allocHealthSetter) SetHealth(healthy, isDeploy bool, trackerTaskEvents map[string]*structs.TaskEvent) { + // Updating alloc deployment state is tricky because it may be nil, but + // if it's not then we need to maintain the values of Canary and + // ModifyIndex as they're only mutated by the server. + a.ar.stateLock.Lock() + a.ar.state.SetDeploymentStatus(time.Now(), healthy) + a.ar.stateLock.Unlock() + + // If deployment is unhealthy emit task events explaining why + a.ar.tasksLock.RLock() + if !healthy && isDeploy { + for task, event := range trackerTaskEvents { + if tr, ok := a.ar.tasks[task]; ok { + tr.EmitEvent(event) + } + } + } + + // Gather the state of the other tasks + states := make(map[string]*structs.TaskState, len(a.ar.tasks)) + for name, tr := range a.ar.tasks { + states[name] = tr.TaskState() + } + a.ar.tasksLock.RUnlock() + + // Build the client allocation + calloc := a.ar.clientAlloc(states) + + // Update the server + a.ar.stateUpdater.AllocStateUpdated(calloc) + + // Broadcast client alloc to listeners + a.ar.allocBroadcaster.Send(calloc) +} + // initRunnerHooks intializes the runners hooks. func (ar *allocRunner) initRunnerHooks() { hookLogger := ar.logger.Named("runner_hook") + // create health setting shim + hs := &allocHealthSetter{ar} + // Create the alloc directory hook. This is run first to ensure the // directoy path exists for other hooks. ar.runnerHooks = []interfaces.RunnerHook{ newAllocDirHook(hookLogger, ar.allocDir), newDiskMigrationHook(hookLogger, ar.prevAllocWatcher, ar.allocDir), - newAllocHealthWatcherHook(hookLogger, ar, ar.consulClient), + newAllocHealthWatcherHook(hookLogger, ar.Alloc(), hs, ar.Listener(), ar.consulClient), } } diff --git a/client/allocrunnerv2/health_hook.go b/client/allocrunnerv2/health_hook.go index fb2857d62..748adfdb0 100644 --- a/client/allocrunnerv2/health_hook.go +++ b/client/allocrunnerv2/health_hook.go @@ -14,27 +14,20 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) -// healthMutator is able to set/clear alloc health as well as listen for alloc -// changes. -type healthMutator interface { - // Alloc returns the original alloc - Alloc() *structs.Allocation - +// healthMutator is able to set/clear alloc health. +type healthSetter interface { // Set health via the mutator SetHealth(healthy, isDeploy bool, taskEvents map[string]*structs.TaskEvent) // Clear health when the deployment ID changes ClearHealth() - - // Listener listens for alloc updates - Listener() *cstructs.AllocListener } // allocHealthWatcherHook is responsible for watching an allocation's task // status and (optionally) Consul health check status to determine if the // allocation is health or unhealthy. Used by deployments and migrations. type allocHealthWatcherHook struct { - runner healthMutator + healthSetter healthSetter // consul client used to monitor health checks consul consul.ConsulServiceAPI @@ -70,8 +63,8 @@ type allocHealthWatcherHook struct { logger log.Logger } -func newAllocHealthWatcherHook(logger log.Logger, runner healthMutator, consul consul.ConsulServiceAPI) interfaces.RunnerHook { - alloc := runner.Alloc() +func newAllocHealthWatcherHook(logger log.Logger, alloc *structs.Allocation, hs healthSetter, + listener *cstructs.AllocListener, consul consul.ConsulServiceAPI) interfaces.RunnerHook { // Neither deployments nor migrations care about the health of // non-service jobs so never watch their health @@ -80,11 +73,11 @@ func newAllocHealthWatcherHook(logger log.Logger, runner healthMutator, consul c } h := &allocHealthWatcherHook{ - runner: runner, - alloc: alloc, - cancelFn: func() {}, // initialize to prevent nil func panics - consul: consul, - listener: runner.Listener(), + alloc: alloc, + cancelFn: func() {}, // initialize to prevent nil func panics + consul: consul, + healthSetter: hs, + listener: listener, } h.logger = logger.Named(h.Name()) @@ -166,7 +159,7 @@ func (h *allocHealthWatcherHook) Update(req *interfaces.RunnerUpdateRequest) err // Deployment has changed, reset status if req.Alloc.DeploymentID != h.alloc.DeploymentID { - h.runner.ClearHealth() + h.healthSetter.ClearHealth() } // Update alloc @@ -211,7 +204,7 @@ func (h *allocHealthWatcherHook) watchHealth(ctx context.Context, tracker *alloc taskEvents = tracker.TaskEvents() } - h.runner.SetHealth(healthy, h.isDeploy, taskEvents) + h.healthSetter.SetHealth(healthy, h.isDeploy, taskEvents) } } diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 7d62e9c7c..6593be3e4 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -561,7 +561,7 @@ func (tr *TaskRunner) setStateLocal(state string, event *structs.TaskEvent) *str } // EmitEvent appends a new TaskEvent to this task's TaskState. The actual -// TaskState.State (pending, running, dead) is *not* updated. Use SetState to +// TaskState.State (pending, running, dead) is not changed. Use SetState to // transition states. // Events are persisted locally but errors are simply logged. func (tr *TaskRunner) EmitEvent(event *structs.TaskEvent) { From a7b427718c39ae40a78d368b874d6d213c5b30d7 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 10 Sep 2018 17:34:45 -0700 Subject: [PATCH 058/152] tr: refactor EmitEvents into Emit+Append * UpdateState: set state, append event, persist, update servers * EmitEvent: append event, persist, update servers * AppendEvent: append event, persist AppendEvent may not even have to persist, but for the sake of correctness I'm going with that for now. --- client/allocrunnerv2/taskrunner/lifecycle.go | 10 ++-- .../allocrunnerv2/taskrunner/task_runner.go | 48 +++++++++++++------ 2 files changed, 39 insertions(+), 19 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/lifecycle.go b/client/allocrunnerv2/taskrunner/lifecycle.go index 521c0157d..48c477dd0 100644 --- a/client/allocrunnerv2/taskrunner/lifecycle.go +++ b/client/allocrunnerv2/taskrunner/lifecycle.go @@ -16,8 +16,8 @@ func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, fai return ErrTaskNotRunning } - // Emit the event - tr.EmitEvent(event) + // Append the event (it will get emitted by UpdateState) + tr.AppendEvent(event) // Tell the restart tracker that a restart triggered the exit tr.restartTracker.SetRestartTriggered(failure) @@ -66,8 +66,8 @@ func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error return ErrTaskNotRunning } - // Emit the event - tr.EmitEvent(event) + // Append the event (it will get emitted by UpdateState) + tr.AppendEvent(event) // Run the hooks prior to killing the task tr.kill() @@ -89,7 +89,7 @@ func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error } // Store that the task has been destroyed and any associated error. - tr.SetState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskKilled).SetKillError(destroyErr)) + tr.UpdateState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskKilled).SetKillError(destroyErr)) if destroyErr != nil { return destroyErr diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 6593be3e4..3118626ef 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -362,12 +362,12 @@ func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { case structs.TaskNotRestarting, structs.TaskTerminated: tr.logger.Info("not restarting task", "reason", reason) if state == structs.TaskNotRestarting { - tr.SetState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskNotRestarting).SetRestartReason(reason).SetFailsTask()) + tr.UpdateState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskNotRestarting).SetRestartReason(reason).SetFailsTask()) } return false, 0 case structs.TaskRestarting: tr.logger.Info("restarting task", "reason", reason, "delay", when) - tr.SetState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskRestarting).SetRestartDelay(when).SetRestartReason(reason)) + tr.UpdateState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskRestarting).SetRestartDelay(when).SetRestartReason(reason)) return true, 0 default: tr.logger.Error("restart tracker returned unknown state", "state", state) @@ -399,7 +399,7 @@ func (tr *TaskRunner) runDriver() error { tr.setDriverHandle(sresp.Handle) // Emit an event that we started - tr.SetState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) + tr.UpdateState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) return nil } @@ -483,9 +483,9 @@ func (tr *TaskRunner) Restore() error { return nil } -// SetState sets the task runners allocation state and triggers a server +// UpdateState sets the task runners allocation state and triggers a server // update. -func (tr *TaskRunner) SetState(state string, event *structs.TaskEvent) { +func (tr *TaskRunner) UpdateState(state string, event *structs.TaskEvent) { tr.logger.Debug("setting task state", "state", state, "event", event.Type) // Update the local state stateCopy := tr.setStateLocal(state, event) @@ -502,7 +502,7 @@ func (tr *TaskRunner) setStateLocal(state string, event *structs.TaskEvent) *str //XXX REMOVE ME AFTER TESTING if state == "" { - panic("SetState must not be called with an empty state") + panic("UpdateState must not be called with an empty state") } // Update the task state @@ -511,7 +511,7 @@ func (tr *TaskRunner) setStateLocal(state string, event *structs.TaskEvent) *str taskState.State = state // Append the event - tr.emitEventImpl(event) + tr.appendEvent(event) // Handle the state transition. switch state { @@ -561,16 +561,37 @@ func (tr *TaskRunner) setStateLocal(state string, event *structs.TaskEvent) *str } // EmitEvent appends a new TaskEvent to this task's TaskState. The actual -// TaskState.State (pending, running, dead) is not changed. Use SetState to +// TaskState.State (pending, running, dead) is not changed. Use UpdateState to // transition states. -// Events are persisted locally but errors are simply logged. +// Events are persisted locally and sent to the server, but errors are simply +// logged. Use AppendEvent to simply add a new event. func (tr *TaskRunner) EmitEvent(event *structs.TaskEvent) { tr.stateLock.Lock() defer tr.stateLock.Unlock() - tr.emitEventImpl(event) + tr.appendEvent(event) + + if err := tr.stateDB.PutTaskState(tr.allocID, tr.taskName, tr.state); err != nil { + // Only a warning because the next event/state-transition will + // try to persist it again. + tr.logger.Warn("error persisting event", "error", err, "event", event) + } + + // Notify the alloc runner of the event + tr.stateUpdater.TaskStateUpdated(tr.taskName, tr.state.Copy()) +} + +// AppendEvent appends a new TaskEvent to this task's TaskState. The actual +// TaskState.State (pending, running, dead) is not changed. Use UpdateState to +// transition states. +// Events are persisted locally and errors are simply logged. Use EmitEvent +// also update AllocRunner. +func (tr *TaskRunner) AppendEvent(event *structs.TaskEvent) { + tr.stateLock.Lock() + defer tr.stateLock.Unlock() + + tr.appendEvent(event) - //XXX EmitEvents do not change TaskState.State so batch? if err := tr.stateDB.PutTaskState(tr.allocID, tr.taskName, tr.state); err != nil { // Only a warning because the next event/state-transition will // try to persist it again. @@ -578,9 +599,8 @@ func (tr *TaskRunner) EmitEvent(event *structs.TaskEvent) { } } -// emitEventImpl is the implementation of EmitEvent without the locking so it -// can be used from SetState. -func (tr *TaskRunner) emitEventImpl(event *structs.TaskEvent) error { +// appendEvent to task's event slice. Caller must acquire stateLock. +func (tr *TaskRunner) appendEvent(event *structs.TaskEvent) error { // Ensure the event is populated with human readable strings event.PopulateEventDisplayMessage() From a73162c97766027b6a5ffa0b1387d1852e2d074f Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 10 Sep 2018 17:39:34 -0700 Subject: [PATCH 059/152] ar: use multierror in update hook loop Make it match TaskRunner update hook behavior --- client/allocrunnerv2/alloc_runner_hooks.go | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner_hooks.go b/client/allocrunnerv2/alloc_runner_hooks.go index d5be55f6a..dda27c5e5 100644 --- a/client/allocrunnerv2/alloc_runner_hooks.go +++ b/client/allocrunnerv2/alloc_runner_hooks.go @@ -109,7 +109,7 @@ func (ar *allocRunner) prerun() error { } if err := pre.Prerun(context.TODO()); err != nil { - return fmt.Errorf("hook %q failed: %v", name, err) + return fmt.Errorf("pre-run hook %q failed: %v", name, err) } //TODO Persist hook state locally @@ -139,6 +139,7 @@ func (ar *allocRunner) update(update *structs.Allocation) error { Alloc: update, } + var merr multierror.Error for _, hook := range ar.runnerHooks { h, ok := hook.(interfaces.RunnerUpdateHook) if !ok { @@ -153,7 +154,7 @@ func (ar *allocRunner) update(update *structs.Allocation) error { } if err := h.Update(req); err != nil { - return fmt.Errorf("hook %q failed: %v", name, err) + merr.Errors = append(merr.Errors, fmt.Errorf("update hook %q failed: %v", name, err)) } if ar.logger.IsTrace() { @@ -162,7 +163,7 @@ func (ar *allocRunner) update(update *structs.Allocation) error { } } - return nil + return merr.ErrorOrNil() } // postrun is used to run the runners postrun hooks. From a2b696c4cf1e1256f510e8ce90f135120312a29a Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 10 Sep 2018 17:48:09 -0700 Subject: [PATCH 060/152] Use a semaphore to block until watcher exits --- client/allocrunnerv2/health_hook.go | 51 ++++++++++++++++++++--------- 1 file changed, 36 insertions(+), 15 deletions(-) diff --git a/client/allocrunnerv2/health_hook.go b/client/allocrunnerv2/health_hook.go index 748adfdb0..43c6ef374 100644 --- a/client/allocrunnerv2/health_hook.go +++ b/client/allocrunnerv2/health_hook.go @@ -14,6 +14,31 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) +// watchSema acts as a semaphore to allow one watcher at a time. +type watchSema struct { + ch chan int +} + +func newWatchSema() *watchSema { + return &watchSema{ch: make(chan int, 1)} +} + +// acquire the semaphore. +func (w *watchSema) acquire() { + w.ch <- 1 +} + +// release the semaphore. +func (w *watchSema) release() { + <-w.ch +} + +// wait blocks until the semaphore is free to be acquired. +func (w *watchSema) wait() { + w.acquire() + w.release() +} + // healthMutator is able to set/clear alloc health. type healthSetter interface { // Set health via the mutator @@ -40,17 +65,17 @@ type allocHealthWatcherHook struct { // Update and synchronous hooks. hookLock sync.Mutex - // watchLock is held by the health watching/setting goroutine to ensure - // only one health watching goroutine is running at a time. - watchLock sync.Mutex + // watchLock is acquired by the health watching/setting goroutine to + // ensure only one health watching goroutine is running at a time. + watchLock *watchSema // ranOnce is set once Prerun or Update have run at least once. This // prevents Prerun from running if an Update has already been // processed. Must hold hookLock to access. ranOnce bool - // cancelFn stops the health watching/setting goroutine. Grab the - // watchLock to block until it exits. + // cancelFn stops the health watching/setting goroutine. Wait on + // watchLock to block until the watcher exits. cancelFn context.CancelFunc // alloc set by new func or Update. Must hold hookLock to access. @@ -152,10 +177,8 @@ func (h *allocHealthWatcherHook) Update(req *interfaces.RunnerUpdateRequest) err // Cancel the old watcher and create a new one h.cancelFn() - // Acquire the watchLock to ensure the old watcher has exited before - // continuing. Kind of an ugly/easy-to-reuse done chan. - h.watchLock.Lock() - h.watchLock.Unlock() + // Wait until the watcher exits + h.watchLock.wait() // Deployment has changed, reset status if req.Alloc.DeploymentID != h.alloc.DeploymentID { @@ -175,10 +198,8 @@ func (h *allocHealthWatcherHook) Destroy() error { h.cancelFn() h.listener.Close() - // Acquire the watchLock to ensure any existing watcher has exited - // before exiting. Kind of an ugly/easy-to-reuse done chan. - h.watchLock.Lock() - h.watchLock.Unlock() + // Wait until the watcher exits + h.watchLock.wait() return nil } @@ -187,8 +208,8 @@ func (h *allocHealthWatcherHook) Destroy() error { // the context is canceled. watchHealth will be canceled and restarted on // Updates so calls are serialized with a lock. func (h *allocHealthWatcherHook) watchHealth(ctx context.Context, tracker *allochealth.Tracker) { - h.watchLock.Lock() - defer h.watchLock.Unlock() + h.watchLock.acquire() + defer h.watchLock.release() select { case <-ctx.Done(): From 64efc3d3015787441d9e8b6aab9343687006c2f8 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 10 Sep 2018 20:24:53 -0700 Subject: [PATCH 061/152] Emit events before long operations Append when there's nothing blocking between appending and sending an update to the server. --- client/allocrunnerv2/alloc_runner_hooks.go | 4 +++- client/allocrunnerv2/taskrunner/lifecycle.go | 8 ++++---- 2 files changed, 7 insertions(+), 5 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner_hooks.go b/client/allocrunnerv2/alloc_runner_hooks.go index dda27c5e5..6936fd654 100644 --- a/client/allocrunnerv2/alloc_runner_hooks.go +++ b/client/allocrunnerv2/alloc_runner_hooks.go @@ -44,7 +44,9 @@ func (a *allocHealthSetter) SetHealth(healthy, isDeploy bool, trackerTaskEvents if !healthy && isDeploy { for task, event := range trackerTaskEvents { if tr, ok := a.ar.tasks[task]; ok { - tr.EmitEvent(event) + // Append but don't emit event since the server + // will be updated below + tr.AppendEvent(event) } } } diff --git a/client/allocrunnerv2/taskrunner/lifecycle.go b/client/allocrunnerv2/taskrunner/lifecycle.go index 48c477dd0..848929636 100644 --- a/client/allocrunnerv2/taskrunner/lifecycle.go +++ b/client/allocrunnerv2/taskrunner/lifecycle.go @@ -16,8 +16,8 @@ func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, fai return ErrTaskNotRunning } - // Append the event (it will get emitted by UpdateState) - tr.AppendEvent(event) + // Emit the event since it may take a long time to kill + tr.EmitEvent(event) // Tell the restart tracker that a restart triggered the exit tr.restartTracker.SetRestartTriggered(failure) @@ -66,8 +66,8 @@ func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error return ErrTaskNotRunning } - // Append the event (it will get emitted by UpdateState) - tr.AppendEvent(event) + // Emit the event since it may take a long time to kill + tr.EmitEvent(event) // Run the hooks prior to killing the task tr.kill() From 195b8127fbdc0b19f9eec5958c8a3b14eb84999d Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 12 Sep 2018 09:36:04 -0700 Subject: [PATCH 062/152] health_hook: fix panic and add tests Still more testing to do, but I want to get this panic fixed ASAP. All new tests pass with -race --- client/allocrunnerv2/health_hook.go | 1 + client/allocrunnerv2/health_hook_test.go | 170 +++++++++++++++++++++++ 2 files changed, 171 insertions(+) create mode 100644 client/allocrunnerv2/health_hook_test.go diff --git a/client/allocrunnerv2/health_hook.go b/client/allocrunnerv2/health_hook.go index 43c6ef374..1b51fdbb3 100644 --- a/client/allocrunnerv2/health_hook.go +++ b/client/allocrunnerv2/health_hook.go @@ -103,6 +103,7 @@ func newAllocHealthWatcherHook(logger log.Logger, alloc *structs.Allocation, hs consul: consul, healthSetter: hs, listener: listener, + watchLock: newWatchSema(), } h.logger = logger.Named(h.Name()) diff --git a/client/allocrunnerv2/health_hook_test.go b/client/allocrunnerv2/health_hook_test.go new file mode 100644 index 000000000..42b2d4ac5 --- /dev/null +++ b/client/allocrunnerv2/health_hook_test.go @@ -0,0 +1,170 @@ +package allocrunnerv2 + +import ( + "context" + "sync" + "testing" + "time" + + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/consul" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/nomad/mock" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/stretchr/testify/require" +) + +// statically assert health hook implements the expected interfaces +var _ interfaces.RunnerPrerunHook = (*allocHealthWatcherHook)(nil) +var _ interfaces.RunnerUpdateHook = (*allocHealthWatcherHook)(nil) +var _ interfaces.RunnerDestroyHook = (*allocHealthWatcherHook)(nil) + +// mockHealthSetter implements healthSetter that stores health internally +type mockHealthSetter struct { + setCalls int + clearCalls int + healthy *bool + isDeploy *bool + taskEvents map[string]*structs.TaskEvent + mu sync.Mutex +} + +func (m *mockHealthSetter) SetHealth(healthy, isDeploy bool, taskEvents map[string]*structs.TaskEvent) { + m.mu.Lock() + defer m.mu.Unlock() + + m.setCalls++ + m.healthy = &healthy + m.isDeploy = &isDeploy + m.taskEvents = taskEvents +} + +func (m *mockHealthSetter) ClearHealth() { + m.mu.Lock() + defer m.mu.Unlock() + + m.clearCalls++ + m.healthy = nil + m.isDeploy = nil + m.taskEvents = nil +} + +func TestHealthHook_PrerunDestroy(t *testing.T) { + t.Parallel() + require := require.New(t) + + b := cstructs.NewAllocBroadcaster() + defer b.Close() + + consul := consul.NewMockConsulServiceClient(t) + hs := &mockHealthSetter{} + + h := newAllocHealthWatcherHook(testlog.HCLogger(t), mock.Alloc(), hs, b.Listen(), consul) + + // Assert we implemented the right interfaces + prerunh, ok := h.(interfaces.RunnerPrerunHook) + require.True(ok) + _, ok = h.(interfaces.RunnerUpdateHook) + require.True(ok) + destroyh, ok := h.(interfaces.RunnerDestroyHook) + require.True(ok) + + // Prerun + require.NoError(prerunh.Prerun(context.Background())) + + // Destroy + require.NoError(destroyh.Destroy()) +} + +// TestHealthHook_SystemNoop asserts that system jobs return the noop tracker. +func TestHealthHook_SystemNoop(t *testing.T) { + t.Parallel() + + h := newAllocHealthWatcherHook(testlog.HCLogger(t), mock.SystemAlloc(), nil, nil, nil) + + // Assert that it's the noop impl + _, ok := h.(noopAllocHealthWatcherHook) + require.True(t, ok) + + // Assert the noop impl does not implement any hooks + _, ok = h.(interfaces.RunnerPrerunHook) + require.False(t, ok) + _, ok = h.(interfaces.RunnerUpdateHook) + require.False(t, ok) + _, ok = h.(interfaces.RunnerDestroyHook) + require.False(t, ok) +} + +// TestHealthHook_BatchNoop asserts that batch jobs return the noop tracker. +func TestHealthHook_BatchNoop(t *testing.T) { + t.Parallel() + + h := newAllocHealthWatcherHook(testlog.HCLogger(t), mock.BatchAlloc(), nil, nil, nil) + + // Assert that it's the noop impl + _, ok := h.(noopAllocHealthWatcherHook) + require.True(t, ok) +} + +// TestHealthHook_WatchSema asserts only one caller can acquire the watchSema +// lock at once and all other callers are blocked. +func TestHealthHook_WatchSema(t *testing.T) { + t.Parallel() + + s := newWatchSema() + + s.acquire() + + // Second acquire should block + acq2 := make(chan struct{}) + go func() { + s.acquire() + close(acq2) + }() + + select { + case <-acq2: + t.Fatalf("should not have been able to acquire lock") + case <-time.After(33 * time.Millisecond): + // Ok! It's blocking! + } + + // Release and assert second acquire is now active + s.release() + + select { + case <-acq2: + // Ok! Second acquire unblocked! + case <-time.After(33 * time.Millisecond): + t.Fatalf("second acquire should not have blocked") + } + + // wait() should block until lock is released + waitCh := make(chan struct{}) + go func() { + s.wait() + close(waitCh) + }() + + select { + case <-waitCh: + t.Fatalf("should not have been able to acquire lock") + case <-time.After(33 * time.Millisecond): + // Ok! It's blocking! + } + + // Release and assert wait() finished + s.release() + + select { + case <-waitCh: + // Ok! wait() unblocked! + case <-time.After(33 * time.Millisecond): + t.Fatalf("wait should not have blocked") + } + + // Assert wait() did not acquire the lock + s.acquire() + s.release() +} From 66ff12e5f7d00b7b20d60abf0cb8004dd2651c97 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Thu, 13 Sep 2018 23:00:31 -0400 Subject: [PATCH 063/152] Update runc/libcontainer and friends (#4655) * vendor: bump libcontainer and docker to remove Sirupsen imports * vendor: fix bad vendoring of archive package * vendor: fix api changes to cgroups in executor * vendor: fix docker api changes * vendor: update github.com/Azure/go-ansiterm to use non capitalized logrus import --- client/driver/docker.go | 14 +- client/driver/executor/executor_linux.go | 7 +- .../azure-sdk-for-go/LICENSE} | 13 +- .../github.com/Azure/azure-sdk-for-go/NOTICE | 5 + .../2015-06-15/network/applicationgateways.go | 566 + .../network/mgmt/2015-06-15/network/client.go | 124 + .../expressroutecircuitauthorizations.go | 351 + .../network/expressroutecircuitpeerings.go | 348 + .../network/expressroutecircuits.go | 718 ++ .../network/expressrouteserviceproviders.go | 131 + .../mgmt/2015-06-15/network/interfaces.go | 702 ++ .../mgmt/2015-06-15/network/loadbalancers.go | 434 + .../network/localnetworkgateways.go | 340 + .../network/mgmt/2015-06-15/network/models.go | 9113 +++++++++++++++++ .../2015-06-15/network/publicipaddresses.go | 446 + .../network/mgmt/2015-06-15/network/routes.go | 348 + .../mgmt/2015-06-15/network/routetables.go | 434 + .../mgmt/2015-06-15/network/securitygroups.go | 434 + .../mgmt/2015-06-15/network/securityrules.go | 358 + .../mgmt/2015-06-15/network/subnets.go | 352 + .../network/mgmt/2015-06-15/network/usages.go | 140 + .../mgmt/2015-06-15/network/version.go | 30 + .../virtualnetworkgatewayconnections.go | 560 + .../network/virtualnetworkgateways.go | 484 + .../2015-06-15/network/virtualnetworks.go | 434 + .../Azure/azure-sdk-for-go/version/version.go | 21 + .../Azure/go-ansiterm/csi_entry_state.go | 4 +- .../Azure/go-ansiterm/csi_param_state.go | 4 +- .../go-ansiterm/escape_intermediate_state.go | 4 +- .../Azure/go-ansiterm/escape_state.go | 4 +- .../Azure/go-ansiterm/osc_string_state.go | 2 +- vendor/github.com/Azure/go-ansiterm/parser.go | 101 +- .../go-ansiterm/parser_action_helpers.go | 4 - .../Azure/go-ansiterm/parser_actions.go | 17 +- .../Azure/go-ansiterm/winterm/ansi.go | 2 +- .../Azure/go-ansiterm/winterm/api.go | 25 +- .../go-ansiterm/winterm/cursor_helpers.go | 2 +- .../go-ansiterm/winterm/scroll_helper.go | 4 +- .../go-ansiterm/winterm/win_event_handler.go | 117 +- vendor/github.com/Sirupsen/logrus/.gitignore | 1 - vendor/github.com/Sirupsen/logrus/.travis.yml | 9 - .../Sirupsen/logrus/json_formatter.go | 41 - vendor/github.com/Sirupsen/logrus/logger.go | 212 - .../Sirupsen/logrus/terminal_bsd.go | 9 - .../Sirupsen/logrus/terminal_notwindows.go | 21 - .../Sirupsen/logrus/terminal_solaris.go | 15 - .../Sirupsen/logrus/terminal_windows.go | 27 - .../Sirupsen/logrus/text_formatter.go | 161 - vendor/github.com/Sirupsen/logrus/writer.go | 31 - vendor/github.com/armon/circbuf/.gitignore | 22 - vendor/github.com/armon/circbuf/LICENSE | 0 .../github.com/bgentry/speakeasy/.gitignore | 2 - vendor/github.com/containerd/console/LICENSE | 201 + .../github.com/containerd/console/README.md | 17 + .../github.com/containerd/console/console.go | 62 + .../containerd/console/console_linux.go | 255 + .../containerd/console/console_unix.go | 142 + .../containerd/console/console_windows.go | 200 + .../containerd/console/tc_darwin.go | 37 + .../containerd/console/tc_freebsd.go | 29 + .../github.com/containerd/console/tc_linux.go | 33 + .../containerd/console/tc_solaris_cgo.go | 35 + .../containerd/console/tc_solaris_nocgo.go | 31 + .../github.com/containerd/console/tc_unix.go | 75 + .../github.com/containerd/continuity/LICENSE | 201 + .../continuity/pathdriver/path_driver.go | 85 + .../github.com => }/coreos/go-systemd/LICENSE | 0 .../github.com/coreos/go-systemd/dbus/dbus.go | 213 + .../coreos/go-systemd/dbus/methods.go | 565 + .../coreos/go-systemd/dbus/properties.go | 237 + .../github.com/coreos/go-systemd/dbus/set.go | 47 + .../coreos/go-systemd/dbus/subscription.go | 250 + .../go-systemd/dbus/subscription_set.go | 57 + .../github.com/coreos/go-systemd/util/util.go | 90 + .../coreos/go-systemd/util/util_cgo.go | 174 + .../coreos/go-systemd/util/util_stub.go | 23 + vendor/github.com/coreos/pkg/LICENSE | 201 + vendor/github.com/coreos/pkg/NOTICE | 5 + vendor/github.com/coreos/pkg/dlopen/dlopen.go | 82 + .../coreos/pkg/dlopen/dlopen_example.go | 56 + .../filepath-securejoin}/LICENSE | 3 +- .../cyphar/filepath-securejoin/README.md | 65 + .../cyphar/filepath-securejoin/join.go | 135 + .../cyphar/filepath-securejoin/vendor.conf | 1 + .../cyphar/filepath-securejoin/vfs.go | 41 + .../docker/go-units => docker/cli}/LICENSE | 2 +- .../docker/docker => docker/cli}/NOTICE | 2 +- .../docker/cli/cli/config/configfile/file.go | 335 + .../cli/cli/config/credentials/credentials.go | 17 + .../cli/config/credentials/default_store.go | 21 + .../credentials/default_store_darwin.go | 5 + .../config/credentials/default_store_linux.go | 13 + .../credentials/default_store_unsupported.go | 7 + .../credentials/default_store_windows.go | 5 + .../cli/cli/config/credentials/file_store.go | 64 + .../cli/config/credentials/native_store.go | 143 + vendor/github.com/docker/cli/opts/config.go | 98 + vendor/github.com/docker/cli/opts/duration.go | 64 + vendor/github.com/docker/cli/opts/env.go | 46 + vendor/github.com/docker/cli/opts/envfile.go | 22 + vendor/github.com/docker/cli/opts/file.go | 71 + vendor/github.com/docker/cli/opts/hosts.go | 165 + .../github.com/docker/cli/opts/hosts_unix.go | 8 + .../docker/cli/opts/hosts_windows.go | 6 + vendor/github.com/docker/cli/opts/ip.go | 47 + vendor/github.com/docker/cli/opts/mount.go | 174 + vendor/github.com/docker/cli/opts/network.go | 106 + vendor/github.com/docker/cli/opts/opts.go | 524 + .../github.com/docker/cli/opts/opts_unix.go | 6 + .../docker/cli/opts/opts_windows.go | 56 + vendor/github.com/docker/cli/opts/parse.go | 99 + vendor/github.com/docker/cli/opts/port.go | 167 + .../docker/cli/opts/quotedstring.go | 37 + vendor/github.com/docker/cli/opts/runtime.go | 79 + vendor/github.com/docker/cli/opts/secret.go | 98 + .../docker/cli/opts/throttledevice.go | 108 + vendor/github.com/docker/cli/opts/ulimit.go | 57 + .../docker/cli/opts/weightdevice.go | 84 + vendor/github.com/docker/distribution/AUTHORS | 182 + .../docker/distribution/BUILDING.md | 117 + .../docker/distribution/CHANGELOG.md | 106 + .../docker/distribution/CONTRIBUTING.md | 148 + .../github.com/docker/distribution/Dockerfile | 21 + .../docker/distribution/MAINTAINERS | 243 + .../github.com/docker/distribution/Makefile | 105 + .../github.com/docker/distribution/README.md | 130 + .../docker/distribution/RELEASE-CHECKLIST.md | 43 + .../github.com/docker/distribution/ROADMAP.md | 266 + .../github.com/docker/distribution/blobs.go | 265 + vendor/github.com/docker/distribution/doc.go | 7 + .../github.com/docker/distribution/errors.go | 119 + .../docker/distribution/manifests.go | 125 + .../docker/distribution/metrics/prometheus.go | 13 + .../docker/distribution/reference/helpers.go | 16 +- .../distribution/reference/normalize.go | 26 +- .../distribution/reference/reference.go | 33 +- .../docker/distribution/reference/regexp.go | 10 +- .../docker/distribution/registry.go | 118 + .../registry/api/errcode/errors.go | 267 + .../registry/api/errcode/handler.go | 40 + .../registry/api/errcode/register.go | 138 + .../registry/api/v2/descriptors.go | 1596 +++ .../distribution/registry/api/v2/doc.go | 9 + .../distribution/registry/api/v2/errors.go | 136 + .../registry/api/v2/headerparser.go | 161 + .../distribution/registry/api/v2/routes.go | 40 + .../distribution/registry/api/v2/urls.go | 266 + .../registry/client/auth/api_version.go | 58 + .../registry/client/auth/challenge/addr.go | 27 + .../client/auth/challenge/authchallenge.go | 237 + .../registry/client/auth/session.go | 532 + .../registry/client/blob_writer.go | 162 + .../distribution/registry/client/errors.go | 139 + .../registry/client/repository.go | 867 ++ .../registry/client/transport/http_reader.go | 251 + .../registry/client/transport/transport.go | 147 + .../registry/storage/cache/cache.go | 35 + .../cache/cachedblobdescriptorstore.go | 129 + .../registry/storage/cache/memory/memory.go | 179 + vendor/github.com/docker/distribution/tags.go | 27 + .../docker/distribution/vendor.conf | 51 + .../docker/docker-credential-helpers/LICENSE | 20 + .../client/client.go | 121 + .../client/command.go | 56 + .../credentials/credentials.go | 186 + .../credentials/error.go | 102 + .../credentials/helper.go | 14 + .../credentials/version.go | 4 + .../docker/docker/api/types/auth.go | 2 +- .../docker/docker/api/types/blkiodev/blkio.go | 2 +- .../docker/docker/api/types/client.go | 18 +- .../docker/docker/api/types/configs.go | 16 +- .../docker/api/types/container/config.go | 13 +- .../api/types/container/container_changes.go | 2 +- .../api/types/container/container_create.go | 2 +- .../api/types/container/container_top.go | 2 +- .../api/types/container/container_update.go | 2 +- .../api/types/container/container_wait.go | 14 +- .../docker/api/types/container/host_config.go | 113 +- .../api/types/container/hostconfig_unix.go | 42 +- .../api/types/container/hostconfig_windows.go | 55 +- .../api/types/container/waitcondition.go | 22 + .../docker/docker/api/types/filters/parse.go | 282 +- .../docker/docker/api/types/mount/mount.go | 25 +- .../docker/api/types/network/network.go | 2 +- .../docker/docker/api/types/plugin.go | 3 + .../docker/api/types/plugin_responses.go | 2 +- .../docker/docker/api/types/port.go | 2 +- .../docker/api/types/registry/authenticate.go | 2 +- .../docker/api/types/registry/registry.go | 2 +- .../docker/docker/api/types/seccomp.go | 2 +- .../docker/docker/api/types/stats.go | 2 +- .../docker/api/types/strslice/strslice.go | 2 +- .../docker/docker/api/types/swarm/common.go | 2 +- .../docker/docker/api/types/swarm/config.go | 6 +- .../docker/api/types/swarm/container.go | 3 +- .../docker/docker/api/types/swarm/network.go | 4 +- .../docker/docker/api/types/swarm/node.go | 2 +- .../docker/docker/api/types/swarm/runtime.go | 10 +- .../docker/api/types/swarm/runtime/gen.go | 2 +- .../api/types/swarm/runtime/plugin.proto | 2 + .../docker/docker/api/types/swarm/secret.go | 6 +- .../docker/docker/api/types/swarm/service.go | 2 +- .../docker/docker/api/types/swarm/swarm.go | 2 +- .../docker/docker/api/types/swarm/task.go | 35 +- .../docker/docker/api/types/types.go | 31 +- .../docker/api/types/versions/compare.go | 2 +- .../docker/cli/config/configfile/file.go | 183 - .../contrib/docker-engine-selinux/LICENSE | 340 - .../docker/docker/contrib/syntax/vim/LICENSE | 22 - .../github.com/docker/docker/errdefs/defs.go | 74 + .../github.com/docker/docker/errdefs/doc.go | 8 + .../docker/docker/errdefs/helpers.go | 240 + vendor/github.com/docker/docker/errdefs/is.go | 114 + .../docker/docker/opts/address_pools.go | 84 + vendor/github.com/docker/docker/opts/env.go | 48 + vendor/github.com/docker/docker/opts/hosts.go | 43 +- .../docker/docker/opts/hosts_unix.go | 2 +- .../docker/docker/opts/hosts_windows.go | 4 +- vendor/github.com/docker/docker/opts/ip.go | 9 +- vendor/github.com/docker/docker/opts/opts.go | 137 +- .../docker/docker/opts/opts_unix.go | 4 +- .../docker/docker/opts/opts_windows.go | 4 +- .../docker/docker/opts/quotedstring.go | 37 + .../github.com/docker/docker/opts/runtime.go | 79 + .../github.com/docker/docker/opts/ulimit.go | 81 + .../docker/docker/pkg/archive/archive.go | 667 +- .../docker/pkg/archive/archive_linux.go | 92 + .../docker/pkg/archive/archive_other.go | 7 + .../docker/docker/pkg/archive/archive_unix.go | 66 +- .../docker/pkg/archive/archive_windows.go | 27 +- .../docker/docker/pkg/archive/changes.go | 65 +- .../docker/pkg/archive/changes_linux.go | 11 +- .../docker/pkg/archive/changes_other.go | 2 +- .../docker/docker/pkg/archive/changes_unix.go | 9 +- .../docker/pkg/archive/changes_windows.go | 8 +- .../docker/docker/pkg/archive/copy.go | 80 +- .../docker/docker/pkg/archive/copy_unix.go | 2 +- .../docker/docker/pkg/archive/copy_windows.go | 2 +- .../docker/docker/pkg/archive/diff.go | 53 +- .../docker/pkg/archive/example_changes.go | 2 +- .../docker/docker/pkg/archive/time_linux.go | 4 +- .../docker/pkg/archive/time_unsupported.go | 2 +- .../docker/docker/pkg/archive/whiteouts.go | 2 +- .../docker/docker/pkg/archive/wrap.go | 8 +- .../docker/docker/pkg/fileutils/fileutils.go | 213 +- .../docker/pkg/fileutils/fileutils_darwin.go | 27 + .../docker/pkg/fileutils/fileutils_solaris.go | 7 - .../docker/pkg/fileutils/fileutils_unix.go | 4 +- .../docker/pkg/fileutils/fileutils_windows.go | 2 +- .../docker/pkg/homedir/homedir_linux.go | 21 + .../docker/pkg/homedir/homedir_others.go | 13 + .../homedir/{homedir.go => homedir_unix.go} | 13 +- .../docker/pkg/homedir/homedir_windows.go | 24 + .../docker/docker/pkg/idtools/idtools.go | 149 +- .../docker/docker/pkg/idtools/idtools_unix.go | 194 +- .../docker/pkg/idtools/idtools_windows.go | 13 +- .../docker/pkg/idtools/usergroupadd_linux.go | 26 +- .../pkg/idtools/usergroupadd_unsupported.go | 2 +- .../docker/docker/pkg/idtools/utils_unix.go | 32 + .../docker/docker/pkg/ioutils/buffer.go | 2 +- .../docker/docker/pkg/ioutils/bytespipe.go | 5 +- .../docker/docker/pkg/ioutils/fmt.go | 22 - .../docker/docker/pkg/ioutils/fswriters.go | 143 +- .../docker/docker/pkg/ioutils/multireader.go | 226 - .../docker/docker/pkg/ioutils/readers.go | 17 +- .../docker/docker/pkg/ioutils/temp_unix.go | 2 +- .../docker/docker/pkg/ioutils/temp_windows.go | 4 +- .../docker/docker/pkg/ioutils/writeflusher.go | 2 +- .../docker/docker/pkg/ioutils/writers.go | 2 +- .../docker/docker/pkg/jsonlog/jsonlog.go | 42 - .../docker/pkg/jsonlog/jsonlog_marshalling.go | 178 - .../docker/docker/pkg/jsonlog/jsonlogbytes.go | 122 - .../docker/pkg/jsonlog/time_marshalling.go | 27 - .../docker/pkg/jsonmessage/jsonmessage.go | 92 +- .../docker/docker/pkg/longpath/longpath.go | 2 +- .../docker/docker/pkg/mount/flags.go | 149 + .../docker/docker/pkg/mount/flags_freebsd.go | 49 + .../docker/docker/pkg/mount/flags_linux.go | 87 + .../docker/pkg/mount/flags_unsupported.go | 31 + .../docker/docker/pkg/mount/mount.go | 141 + .../docker/pkg/mount/mounter_freebsd.go | 60 + .../docker/docker/pkg/mount/mounter_linux.go | 57 + .../docker/pkg/mount/mounter_unsupported.go | 11 + .../docker/docker/pkg/mount/mountinfo.go | 40 + .../docker/pkg/mount/mountinfo_freebsd.go | 55 + .../docker/pkg/mount/mountinfo_linux.go | 132 + .../docker/pkg/mount/mountinfo_unsupported.go | 12 + .../docker/pkg/mount/mountinfo_windows.go | 6 + .../docker/pkg/mount/sharedsubtree_linux.go | 67 + .../docker/docker/pkg/pools/pools.go | 56 +- .../docker/docker/pkg/promise/promise.go | 11 - .../docker/docker/pkg/random/random.go | 71 - .../docker/docker/pkg/stdcopy/stdcopy.go | 31 +- .../docker/docker/pkg/stringid/stringid.go | 40 +- .../docker/docker/pkg/system/chtimes.go | 25 +- .../docker/docker/pkg/system/chtimes_unix.go | 2 +- .../docker/pkg/system/chtimes_windows.go | 23 +- .../docker/docker/pkg/system/errors.go | 5 +- .../docker/pkg/system/events_windows.go | 85 - .../docker/docker/pkg/system/exitcode.go | 16 +- .../docker/docker/pkg/system/filesys.go | 11 +- .../docker/pkg/system/filesys_windows.go | 111 +- .../docker/docker/pkg/system/init.go | 22 + .../docker/docker/pkg/system/init_unix.go | 7 + .../docker/docker/pkg/system/init_windows.go | 12 + .../docker/docker/pkg/system/lcow.go | 32 + .../docker/docker/pkg/system/lcow_unix.go | 8 + .../docker/docker/pkg/system/lcow_windows.go | 6 + .../pkg/system/{lstat.go => lstat_unix.go} | 2 +- .../docker/docker/pkg/system/lstat_windows.go | 17 +- .../docker/docker/pkg/system/meminfo.go | 2 +- .../docker/docker/pkg/system/meminfo_linux.go | 2 +- .../docker/pkg/system/meminfo_solaris.go | 129 - .../docker/pkg/system/meminfo_unsupported.go | 4 +- .../docker/pkg/system/meminfo_windows.go | 2 +- .../docker/docker/pkg/system/mknod.go | 8 +- .../docker/docker/pkg/system/mknod_windows.go | 4 +- .../docker/docker/pkg/system/path.go | 60 + .../docker/docker/pkg/system/path_unix.go | 14 - .../docker/docker/pkg/system/path_windows.go | 37 - .../docker/docker/pkg/system/process_unix.go | 12 +- .../docker/pkg/system/process_windows.go | 22 +- .../github.com/docker/docker/pkg/system/rm.go | 80 + .../docker/docker/pkg/system/stat_darwin.go | 25 +- .../docker/docker/pkg/system/stat_freebsd.go | 18 +- .../docker/docker/pkg/system/stat_linux.go | 22 +- .../docker/docker/pkg/system/stat_openbsd.go | 8 +- .../docker/docker/pkg/system/stat_solaris.go | 27 +- .../pkg/system/{stat.go => stat_unix.go} | 20 +- .../docker/pkg/system/stat_unsupported.go | 17 - .../docker/docker/pkg/system/stat_windows.go | 48 +- .../docker/docker/pkg/system/syscall_unix.go | 6 +- .../docker/pkg/system/syscall_windows.go | 42 +- .../docker/docker/pkg/system/umask.go | 6 +- .../docker/docker/pkg/system/umask_windows.go | 4 +- .../docker/pkg/system/utimes_freebsd.go | 10 +- .../docker/docker/pkg/system/utimes_linux.go | 15 +- .../docker/pkg/system/utimes_unsupported.go | 2 +- .../docker/docker/pkg/system/xattrs_linux.go | 50 +- .../docker/pkg/system/xattrs_unsupported.go | 2 +- .../docker/pkg/tarsum/builder_context.go | 21 + .../docker/docker/pkg/tarsum/fileinfosums.go | 133 + .../docker/docker/pkg/tarsum/tarsum.go | 301 + .../docker/docker/pkg/tarsum/tarsum_spec.md | 229 + .../docker/docker/pkg/tarsum/versioning.go | 158 + .../docker/docker/pkg/tarsum/writercloser.go | 22 + .../docker/docker/pkg/term/ascii.go | 4 +- .../docker/docker/pkg/term/proxy.go | 78 + .../github.com/docker/docker/pkg/term/tc.go | 20 + .../docker/docker/pkg/term/tc_linux_cgo.go | 50 - .../docker/docker/pkg/term/tc_other.go | 20 - .../docker/docker/pkg/term/tc_solaris_cgo.go | 63 - .../github.com/docker/docker/pkg/term/term.go | 7 +- .../docker/docker/pkg/term/term_solaris.go | 41 - .../docker/docker/pkg/term/term_unix.go | 29 - .../docker/docker/pkg/term/term_windows.go | 51 +- .../docker/docker/pkg/term/termios_bsd.go | 42 + .../docker/docker/pkg/term/termios_darwin.go | 69 - .../docker/docker/pkg/term/termios_freebsd.go | 69 - .../docker/docker/pkg/term/termios_linux.go | 42 +- .../docker/docker/pkg/term/termios_openbsd.go | 69 - .../docker/pkg/term/windows/ansi_reader.go | 2 +- .../docker/pkg/term/windows/ansi_writer.go | 2 +- .../docker/docker/pkg/term/windows/console.go | 2 +- .../docker/docker/pkg/term/windows/windows.go | 6 +- .../docker/docker/pkg/term/winsize.go | 20 + .../docker/docker/reference/reference.go | 216 - .../github.com/docker/docker/registry/auth.go | 250 +- .../docker/docker/registry/config.go | 173 +- .../docker/docker/registry/config_unix.go | 16 + .../docker/docker/registry/config_windows.go | 18 + .../docker/docker/registry/endpoint_v1.go | 198 + .../docker/docker/registry/errors.go | 31 + .../docker/docker/registry/registry.go | 191 + .../resumable/resumablerequestreader.go | 96 + .../docker/docker/registry/service.go | 312 +- .../docker/docker/registry/service_v1.go | 40 + .../docker/docker/registry/service_v2.go | 82 + .../docker/docker/registry/session.go | 779 ++ .../docker/docker/registry/types.go | 6 +- .../docker/go-connections/sockets/README.md | 0 .../go-connections/sockets/inmem_socket.go | 81 + .../docker/go-connections/sockets/proxy.go | 51 + .../docker/go-connections/sockets/sockets.go | 38 + .../go-connections/sockets/sockets_unix.go | 35 + .../go-connections/sockets/sockets_windows.go | 27 + .../go-connections/sockets/tcp_socket.go | 22 + .../go-connections/sockets/unix_socket.go | 32 + .../go-connections/tlsconfig/certpool_go17.go | 18 + .../tlsconfig/certpool_other.go | 13 + .../docker/go-connections/tlsconfig/config.go | 256 + .../tlsconfig/config_client_ciphers.go | 17 + .../tlsconfig/config_legacy_client_ciphers.go | 15 + .../docker/go-metrics/CONTRIBUTING.md | 55 + .../go-metrics/LICENSE.code} | 2 +- .../github.com/docker/go-metrics/LICENSE.docs | 425 + vendor/github.com/docker/go-metrics/NOTICE | 16 + vendor/github.com/docker/go-metrics/README.md | 91 + .../github.com/docker/go-metrics/counter.go | 52 + vendor/github.com/docker/go-metrics/docs.go | 3 + vendor/github.com/docker/go-metrics/gauge.go | 72 + .../github.com/docker/go-metrics/handler.go | 74 + .../github.com/docker/go-metrics/helpers.go | 10 + .../github.com/docker/go-metrics/namespace.go | 315 + .../github.com/docker/go-metrics/register.go | 15 + vendor/github.com/docker/go-metrics/timer.go | 85 + vendor/github.com/docker/go-metrics/unit.go | 12 + vendor/github.com/docker/libnetwork/LICENSE | 201 + .../docker/libnetwork/ipamutils/utils.go | 96 + .../github.com/dustin/go-humanize/.gitignore | 6 - vendor/github.com/go-ini/ini/.gitignore | 4 - .../go-ole/oleutil/connection_windows.go | 1 - .../src/github.com => }/godbus/dbus/LICENSE | 0 vendor/github.com/godbus/dbus/README.markdown | 41 + vendor/github.com/godbus/dbus/auth.go | 253 + .../github.com/godbus/dbus/auth_external.go | 26 + vendor/github.com/godbus/dbus/auth_sha1.go | 102 + vendor/github.com/godbus/dbus/call.go | 36 + vendor/github.com/godbus/dbus/conn.go | 625 ++ vendor/github.com/godbus/dbus/conn_darwin.go | 21 + vendor/github.com/godbus/dbus/conn_other.go | 27 + vendor/github.com/godbus/dbus/dbus.go | 258 + vendor/github.com/godbus/dbus/decoder.go | 228 + vendor/github.com/godbus/dbus/doc.go | 63 + vendor/github.com/godbus/dbus/encoder.go | 208 + vendor/github.com/godbus/dbus/export.go | 411 + vendor/github.com/godbus/dbus/homedir.go | 28 + .../github.com/godbus/dbus/homedir_dynamic.go | 15 + .../github.com/godbus/dbus/homedir_static.go | 45 + vendor/github.com/godbus/dbus/message.go | 346 + vendor/github.com/godbus/dbus/object.go | 126 + vendor/github.com/godbus/dbus/sig.go | 257 + .../godbus/dbus/transport_darwin.go | 6 + .../godbus/dbus/transport_generic.go | 35 + .../github.com/godbus/dbus/transport_unix.go | 196 + .../dbus/transport_unixcred_dragonfly.go | 95 + .../godbus/dbus/transport_unixcred_linux.go | 25 + vendor/github.com/godbus/dbus/variant.go | 139 + .../github.com/godbus/dbus/variant_lexer.go | 284 + .../github.com/godbus/dbus/variant_parser.go | 817 ++ .../pkg/mflag => gorilla/context}/LICENSE | 8 +- vendor/github.com/gorilla/context/README.md | 10 + vendor/github.com/gorilla/context/context.go | 143 + vendor/github.com/gorilla/context/doc.go | 88 + vendor/github.com/gorilla/mux/AUTHORS | 8 + .../github.com/gorilla/mux/ISSUE_TEMPLATE.md | 10 + .../LICENSE.BSD => gorilla/mux/LICENSE} | 8 +- vendor/github.com/gorilla/mux/README.md | 649 ++ .../github.com/gorilla/mux/context_gorilla.go | 26 + .../github.com/gorilla/mux/context_native.go | 24 + vendor/github.com/gorilla/mux/doc.go | 306 + vendor/github.com/gorilla/mux/middleware.go | 72 + vendor/github.com/gorilla/mux/mux.go | 588 ++ vendor/github.com/gorilla/mux/regexp.go | 332 + vendor/github.com/gorilla/mux/route.go | 763 ++ vendor/github.com/gorilla/mux/test_helpers.go | 19 + .../go-msgpack/codec/msgpack_test.py | 0 .../github.com/hashicorp/go-syslog/.gitignore | 22 - vendor/github.com/hashicorp/hcl/appveyor.yml | 19 + .../github.com/hashicorp/logutils/.gitignore | 22 - vendor/github.com/hashicorp/vault/main.go | 11 + .../jmespath/go-jmespath/.gitignore | 4 - .../jmespath/go-jmespath/.travis.yml | 9 - vendor/github.com/miekg/dns/.gitignore | 4 - vendor/github.com/miekg/dns/.travis.yml | 7 - vendor/github.com/miekg/dns/labels.go | 2 +- .../mitchellh/mapstructure/.travis.yml | 7 - .../fileutils/LICENSE} | 2 +- vendor/github.com/mrunalp/fileutils/README.md | 5 + .../github.com/mrunalp/fileutils/fileutils.go | 161 + .../github.com/mrunalp/fileutils/idtools.go | 49 + .../src/github.com/Sirupsen/logrus/LICENSE | 21 - .../src/github.com/codegangsta/cli/LICENSE | 21 - .../docker/docker/contrib/syntax/vim/LICENSE | 22 - .../docker/docs/project/images/red_notice.png | 3 - .../docker/docker/pkg/symlink/LICENSE.BSD | 27 - .../src/github.com/golang/protobuf/LICENSE | 31 - .../github.com/syndtr/gocapability/LICENSE | 24 - .../runc/libcontainer/README.md | 327 + .../opencontainers/runc/libcontainer/SPEC.md | 418 + .../runc/libcontainer/apparmor/apparmor.go | 54 + .../apparmor/apparmor_disabled.go | 20 + .../runc/libcontainer/capabilities_linux.go | 113 + .../runc/libcontainer/cgroups/cgroups.go | 6 +- .../runc/libcontainer/cgroups/fs/apply_raw.go | 167 +- .../runc/libcontainer/cgroups/fs/cpu.go | 55 +- .../runc/libcontainer/cgroups/fs/cpuset.go | 32 +- .../runc/libcontainer/cgroups/fs/devices.go | 26 +- .../runc/libcontainer/cgroups/fs/freezer.go | 13 +- .../runc/libcontainer/cgroups/fs/memory.go | 147 +- .../runc/libcontainer/cgroups/fs/net_cls.go | 6 +- .../runc/libcontainer/cgroups/fs/utils.go | 1 - .../runc/libcontainer/cgroups/stats.go | 3 + .../cgroups/systemd/apply_nosystemd.go | 55 + .../cgroups/systemd/apply_systemd.go | 598 ++ .../runc/libcontainer/cgroups/utils.go | 190 +- .../{cgroup_unix.go => cgroup_linux.go} | 20 +- .../configs/cgroup_unsupported.go | 6 - .../runc/libcontainer/configs/config.go | 108 +- .../runc/libcontainer/configs/config_linux.go | 61 + .../runc/libcontainer/configs/config_unix.go | 51 - .../libcontainer/configs/device_defaults.go | 20 +- .../runc/libcontainer/configs/intelrdt.go | 7 + .../runc/libcontainer/configs/mount.go | 9 + ...namespaces_unix.go => namespaces_linux.go} | 19 +- .../configs/namespaces_syscall.go | 14 +- .../configs/namespaces_syscall_unsupported.go | 2 - .../configs/namespaces_unsupported.go | 2 +- .../libcontainer/configs/validate/rootless.go | 116 + .../configs/validate/validator.go | 222 + .../runc/libcontainer/console_linux.go | 41 + .../runc/libcontainer/container.go | 166 + .../runc/libcontainer/container_linux.go | 1899 ++++ .../runc/libcontainer/criu_opts_linux.go | 40 + .../runc/libcontainer/criurpc/Makefile | 2 + .../runc/libcontainer/criurpc/criurpc.pb.go | 1178 +++ .../runc/libcontainer/criurpc/criurpc.proto | 209 + .../opencontainers/runc/libcontainer/error.go | 70 + .../runc/libcontainer/factory.go | 44 + .../runc/libcontainer/factory_linux.go | 381 + .../runc/libcontainer/generic_error.go | 92 + .../runc/libcontainer/init_linux.go | 535 + .../runc/libcontainer/intelrdt/intelrdt.go | 553 + .../runc/libcontainer/intelrdt/stats.go | 24 + .../runc/libcontainer/keys/keyctl.go | 50 + .../runc/libcontainer/message_linux.go | 89 + .../runc/libcontainer/mount/mount.go | 23 + .../runc/libcontainer/mount/mount_linux.go | 82 + .../runc/libcontainer/mount/mountinfo.go | 40 + .../runc/libcontainer/network_linux.go | 259 + .../runc/libcontainer/notify_linux.go | 90 + .../runc/libcontainer/process.go | 113 + .../runc/libcontainer/process_linux.go | 547 + .../runc/libcontainer/restored_process.go | 122 + .../runc/libcontainer/rootfs_linux.go | 838 ++ .../runc/libcontainer/seccomp/config.go | 76 + .../libcontainer/seccomp/seccomp_linux.go | 258 + .../seccomp/seccomp_unsupported.go | 24 + .../runc/libcontainer/setns_init_linux.go | 76 + .../runc/libcontainer/stacktrace/capture.go | 27 + .../runc/libcontainer/stacktrace/frame.go | 38 + .../libcontainer/stacktrace/stacktrace.go | 5 + .../runc/libcontainer/standard_init_linux.go | 194 + .../runc/libcontainer/state_linux.go | 255 + .../opencontainers/runc/libcontainer/stats.go | 15 + .../runc/libcontainer/stats_linux.go | 10 + .../opencontainers/runc/libcontainer/sync.go | 107 + .../runc/libcontainer/system/linux.go | 87 +- .../runc/libcontainer/system/proc.go | 112 +- .../runc/libcontainer/system/setns_linux.go | 40 - ...scall_linux_arm.go => syscall_linux_32.go} | 9 +- .../libcontainer/system/syscall_linux_386.go | 25 - .../libcontainer/system/syscall_linux_64.go | 9 +- .../runc/libcontainer/system/sysconfig.go | 2 +- .../runc/libcontainer/system/unsupported.go | 18 + .../runc/libcontainer/system/xattrs_linux.go | 88 +- .../runc/libcontainer/user/lookup.go | 77 +- .../runc/libcontainer/user/lookup_unix.go | 112 + .../libcontainer/user/lookup_unsupported.go | 21 - .../runc/libcontainer/user/lookup_windows.go | 40 + .../runc/libcontainer/user/user.go | 201 +- .../runc/libcontainer/utils/cmsg.go | 93 + .../runc/libcontainer/utils/utils.go | 45 +- .../runc/libcontainer/utils/utils_unix.go | 17 +- .../specs => runtime-spec}/LICENSE | 0 .../runtime-spec/specs-go/config.go | 570 ++ .../runtime-spec/specs-go/state.go | 17 + .../runtime-spec/specs-go/version.go | 18 + .../github.com/opencontainers/selinux/LICENSE | 201 + .../selinux/go-selinux/label/label.go | 84 + .../selinux/go-selinux/label/label_selinux.go | 204 + .../selinux/go-selinux/selinux.go | 593 ++ .../selinux/go-selinux/xattrs.go | 78 + vendor/github.com/ryanuber/columnize/LICENSE | 20 + .../seccomp/libseccomp-golang/LICENSE | 0 .../seccomp/libseccomp-golang/README | 50 + .../seccomp/libseccomp-golang/seccomp.go | 864 ++ .../libseccomp-golang/seccomp_internal.go | 508 + .../logrus/CHANGELOG.md | 57 + .../{Sirupsen => sirupsen}/logrus/LICENSE | 0 .../{Sirupsen => sirupsen}/logrus/README.md | 191 +- vendor/github.com/sirupsen/logrus/alt_exit.go | 64 + .../github.com/sirupsen/logrus/appveyor.yml | 14 + .../{Sirupsen => sirupsen}/logrus/doc.go | 4 +- .../{Sirupsen => sirupsen}/logrus/entry.go | 138 +- .../{Sirupsen => sirupsen}/logrus/exported.go | 41 +- .../logrus/formatter.go | 25 +- .../{Sirupsen => sirupsen}/logrus/hooks.go | 0 .../sirupsen/logrus/json_formatter.go | 96 + vendor/github.com/sirupsen/logrus/logger.go | 367 + .../{Sirupsen => sirupsen}/logrus/logrus.go | 9 +- .../sirupsen/logrus/terminal_bsd.go | 10 + .../logrus/terminal_check_appengine.go | 11 + .../sirupsen/logrus/terminal_check_js.go | 11 + .../logrus/terminal_check_notappengine.go | 19 + .../logrus/terminal_linux.go | 8 +- .../sirupsen/logrus/text_formatter.go | 217 + vendor/github.com/sirupsen/logrus/writer.go | 62 + vendor/github.com/ulikunitz/xz/example.go | 40 - .../vishvananda/netlink/LICENSE | 0 .../github.com/vishvananda/netlink/README.md | 89 + vendor/github.com/vishvananda/netlink/addr.go | 43 + .../vishvananda/netlink/addr_linux.go | 128 + .../github.com/vishvananda/netlink/filter.go | 55 + .../vishvananda/netlink/filter_linux.go | 191 + vendor/github.com/vishvananda/netlink/link.go | 223 + .../vishvananda/netlink/link_linux.go | 750 ++ .../github.com/vishvananda/netlink/neigh.go | 22 + .../vishvananda/netlink/neigh_linux.go | 189 + .../github.com/vishvananda/netlink/netlink.go | 39 + .../netlink/netlink_unspecified.go | 143 + .../vishvananda/netlink/nl/addr_linux.go | 47 + .../vishvananda/netlink/nl/link_linux.go | 104 + .../vishvananda/netlink/nl/nl_linux.go | 418 + .../vishvananda/netlink/nl/route_linux.go | 42 + .../vishvananda/netlink/nl/tc_linux.go | 359 + .../vishvananda/netlink/nl/xfrm_linux.go | 258 + .../netlink/nl/xfrm_policy_linux.go | 119 + .../netlink/nl/xfrm_state_linux.go | 221 + .../vishvananda/netlink/protinfo.go | 53 + .../vishvananda/netlink/protinfo_linux.go | 60 + .../github.com/vishvananda/netlink/qdisc.go | 138 + .../vishvananda/netlink/qdisc_linux.go | 263 + .../github.com/vishvananda/netlink/route.go | 35 + .../vishvananda/netlink/route_linux.go | 225 + vendor/github.com/vishvananda/netlink/xfrm.go | 64 + .../vishvananda/netlink/xfrm_policy.go | 59 + .../vishvananda/netlink/xfrm_policy_linux.go | 127 + .../vishvananda/netlink/xfrm_state.go | 53 + .../vishvananda/netlink/xfrm_state_linux.go | 181 + .../go-cty/cty/convert/mismatch_msg.go | 2 +- vendor/golang.org/x/net/proxy/direct.go | 18 + vendor/golang.org/x/net/proxy/per_host.go | 140 + vendor/golang.org/x/net/proxy/proxy.go | 134 + vendor/golang.org/x/net/proxy/socks5.go | 214 + vendor/golang.org/x/sys/unix/creds_test.go | 134 - .../golang.org/x/sys/unix/dev_linux_test.go | 56 - vendor/golang.org/x/sys/unix/example_test.go | 19 - vendor/golang.org/x/sys/unix/export_test.go | 9 - vendor/golang.org/x/sys/unix/mkpost.go | 98 - .../golang.org/x/sys/unix/mmap_unix_test.go | 35 - vendor/golang.org/x/sys/unix/openbsd_test.go | 113 - .../golang.org/x/sys/unix/syscall_bsd_test.go | 93 - .../x/sys/unix/syscall_darwin_test.go | 19 - .../x/sys/unix/syscall_freebsd_test.go | 312 - .../x/sys/unix/syscall_linux_test.go | 421 - .../x/sys/unix/syscall_solaris_test.go | 55 - vendor/golang.org/x/sys/unix/syscall_test.go | 60 - .../x/sys/unix/syscall_unix_test.go | 657 -- .../golang.org/x/sys/unix/timestruct_test.go | 54 - .../golang.org/x/sys/unix/types_dragonfly.go | 280 - vendor/golang.org/x/sys/unix/types_freebsd.go | 402 - vendor/golang.org/x/sys/unix/types_netbsd.go | 281 - vendor/golang.org/x/sys/unix/types_openbsd.go | 282 - vendor/golang.org/x/sys/unix/xattr_test.go | 119 - vendor/vendor.json | 221 +- 657 files changed, 70636 insertions(+), 9075 deletions(-) rename vendor/github.com/{docker/docker/pkg/symlink/LICENSE.APACHE => Azure/azure-sdk-for-go/LICENSE} (94%) create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/NOTICE create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/applicationgateways.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/client.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuitauthorizations.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuitpeerings.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuits.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressrouteserviceproviders.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/interfaces.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/loadbalancers.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/localnetworkgateways.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/models.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/publicipaddresses.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/routes.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/routetables.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/securitygroups.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/securityrules.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/subnets.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/usages.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/version.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworkgatewayconnections.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworkgateways.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworks.go create mode 100644 vendor/github.com/Azure/azure-sdk-for-go/version/version.go delete mode 100644 vendor/github.com/Sirupsen/logrus/.gitignore delete mode 100644 vendor/github.com/Sirupsen/logrus/.travis.yml delete mode 100644 vendor/github.com/Sirupsen/logrus/json_formatter.go delete mode 100644 vendor/github.com/Sirupsen/logrus/logger.go delete mode 100644 vendor/github.com/Sirupsen/logrus/terminal_bsd.go delete mode 100644 vendor/github.com/Sirupsen/logrus/terminal_notwindows.go delete mode 100644 vendor/github.com/Sirupsen/logrus/terminal_solaris.go delete mode 100644 vendor/github.com/Sirupsen/logrus/terminal_windows.go delete mode 100644 vendor/github.com/Sirupsen/logrus/text_formatter.go delete mode 100644 vendor/github.com/Sirupsen/logrus/writer.go delete mode 100644 vendor/github.com/armon/circbuf/.gitignore mode change 100644 => 100755 vendor/github.com/armon/circbuf/LICENSE delete mode 100644 vendor/github.com/bgentry/speakeasy/.gitignore create mode 100644 vendor/github.com/containerd/console/LICENSE create mode 100644 vendor/github.com/containerd/console/README.md create mode 100644 vendor/github.com/containerd/console/console.go create mode 100644 vendor/github.com/containerd/console/console_linux.go create mode 100644 vendor/github.com/containerd/console/console_unix.go create mode 100644 vendor/github.com/containerd/console/console_windows.go create mode 100644 vendor/github.com/containerd/console/tc_darwin.go create mode 100644 vendor/github.com/containerd/console/tc_freebsd.go create mode 100644 vendor/github.com/containerd/console/tc_linux.go create mode 100644 vendor/github.com/containerd/console/tc_solaris_cgo.go create mode 100644 vendor/github.com/containerd/console/tc_solaris_nocgo.go create mode 100644 vendor/github.com/containerd/console/tc_unix.go create mode 100644 vendor/github.com/containerd/continuity/LICENSE create mode 100644 vendor/github.com/containerd/continuity/pathdriver/path_driver.go rename vendor/github.com/{opencontainers/runc/Godeps/_workspace/src/github.com => }/coreos/go-systemd/LICENSE (100%) create mode 100644 vendor/github.com/coreos/go-systemd/dbus/dbus.go create mode 100644 vendor/github.com/coreos/go-systemd/dbus/methods.go create mode 100644 vendor/github.com/coreos/go-systemd/dbus/properties.go create mode 100644 vendor/github.com/coreos/go-systemd/dbus/set.go create mode 100644 vendor/github.com/coreos/go-systemd/dbus/subscription.go create mode 100644 vendor/github.com/coreos/go-systemd/dbus/subscription_set.go create mode 100644 vendor/github.com/coreos/go-systemd/util/util.go create mode 100644 vendor/github.com/coreos/go-systemd/util/util_cgo.go create mode 100644 vendor/github.com/coreos/go-systemd/util/util_stub.go create mode 100644 vendor/github.com/coreos/pkg/LICENSE create mode 100644 vendor/github.com/coreos/pkg/NOTICE create mode 100644 vendor/github.com/coreos/pkg/dlopen/dlopen.go create mode 100644 vendor/github.com/coreos/pkg/dlopen/dlopen_example.go rename vendor/github.com/{docker/docker/pkg/mflag => cyphar/filepath-securejoin}/LICENSE (92%) create mode 100644 vendor/github.com/cyphar/filepath-securejoin/README.md create mode 100644 vendor/github.com/cyphar/filepath-securejoin/join.go create mode 100644 vendor/github.com/cyphar/filepath-securejoin/vendor.conf create mode 100644 vendor/github.com/cyphar/filepath-securejoin/vfs.go rename vendor/github.com/{opencontainers/runc/Godeps/_workspace/src/github.com/docker/go-units => docker/cli}/LICENSE (99%) rename vendor/github.com/{opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker => docker/cli}/NOTICE (94%) create mode 100644 vendor/github.com/docker/cli/cli/config/configfile/file.go create mode 100644 vendor/github.com/docker/cli/cli/config/credentials/credentials.go create mode 100644 vendor/github.com/docker/cli/cli/config/credentials/default_store.go create mode 100644 vendor/github.com/docker/cli/cli/config/credentials/default_store_darwin.go create mode 100644 vendor/github.com/docker/cli/cli/config/credentials/default_store_linux.go create mode 100644 vendor/github.com/docker/cli/cli/config/credentials/default_store_unsupported.go create mode 100644 vendor/github.com/docker/cli/cli/config/credentials/default_store_windows.go create mode 100644 vendor/github.com/docker/cli/cli/config/credentials/file_store.go create mode 100644 vendor/github.com/docker/cli/cli/config/credentials/native_store.go create mode 100644 vendor/github.com/docker/cli/opts/config.go create mode 100644 vendor/github.com/docker/cli/opts/duration.go create mode 100644 vendor/github.com/docker/cli/opts/env.go create mode 100644 vendor/github.com/docker/cli/opts/envfile.go create mode 100644 vendor/github.com/docker/cli/opts/file.go create mode 100644 vendor/github.com/docker/cli/opts/hosts.go create mode 100644 vendor/github.com/docker/cli/opts/hosts_unix.go create mode 100644 vendor/github.com/docker/cli/opts/hosts_windows.go create mode 100644 vendor/github.com/docker/cli/opts/ip.go create mode 100644 vendor/github.com/docker/cli/opts/mount.go create mode 100644 vendor/github.com/docker/cli/opts/network.go create mode 100644 vendor/github.com/docker/cli/opts/opts.go create mode 100644 vendor/github.com/docker/cli/opts/opts_unix.go create mode 100644 vendor/github.com/docker/cli/opts/opts_windows.go create mode 100644 vendor/github.com/docker/cli/opts/parse.go create mode 100644 vendor/github.com/docker/cli/opts/port.go create mode 100644 vendor/github.com/docker/cli/opts/quotedstring.go create mode 100644 vendor/github.com/docker/cli/opts/runtime.go create mode 100644 vendor/github.com/docker/cli/opts/secret.go create mode 100644 vendor/github.com/docker/cli/opts/throttledevice.go create mode 100644 vendor/github.com/docker/cli/opts/ulimit.go create mode 100644 vendor/github.com/docker/cli/opts/weightdevice.go create mode 100644 vendor/github.com/docker/distribution/AUTHORS create mode 100644 vendor/github.com/docker/distribution/BUILDING.md create mode 100644 vendor/github.com/docker/distribution/CHANGELOG.md create mode 100644 vendor/github.com/docker/distribution/CONTRIBUTING.md create mode 100644 vendor/github.com/docker/distribution/Dockerfile create mode 100644 vendor/github.com/docker/distribution/MAINTAINERS create mode 100644 vendor/github.com/docker/distribution/Makefile create mode 100644 vendor/github.com/docker/distribution/README.md create mode 100644 vendor/github.com/docker/distribution/RELEASE-CHECKLIST.md create mode 100644 vendor/github.com/docker/distribution/ROADMAP.md create mode 100644 vendor/github.com/docker/distribution/blobs.go create mode 100644 vendor/github.com/docker/distribution/doc.go create mode 100644 vendor/github.com/docker/distribution/errors.go create mode 100644 vendor/github.com/docker/distribution/manifests.go create mode 100644 vendor/github.com/docker/distribution/metrics/prometheus.go create mode 100644 vendor/github.com/docker/distribution/registry.go create mode 100644 vendor/github.com/docker/distribution/registry/api/errcode/errors.go create mode 100644 vendor/github.com/docker/distribution/registry/api/errcode/handler.go create mode 100644 vendor/github.com/docker/distribution/registry/api/errcode/register.go create mode 100644 vendor/github.com/docker/distribution/registry/api/v2/descriptors.go create mode 100644 vendor/github.com/docker/distribution/registry/api/v2/doc.go create mode 100644 vendor/github.com/docker/distribution/registry/api/v2/errors.go create mode 100644 vendor/github.com/docker/distribution/registry/api/v2/headerparser.go create mode 100644 vendor/github.com/docker/distribution/registry/api/v2/routes.go create mode 100644 vendor/github.com/docker/distribution/registry/api/v2/urls.go create mode 100644 vendor/github.com/docker/distribution/registry/client/auth/api_version.go create mode 100644 vendor/github.com/docker/distribution/registry/client/auth/challenge/addr.go create mode 100644 vendor/github.com/docker/distribution/registry/client/auth/challenge/authchallenge.go create mode 100644 vendor/github.com/docker/distribution/registry/client/auth/session.go create mode 100644 vendor/github.com/docker/distribution/registry/client/blob_writer.go create mode 100644 vendor/github.com/docker/distribution/registry/client/errors.go create mode 100644 vendor/github.com/docker/distribution/registry/client/repository.go create mode 100644 vendor/github.com/docker/distribution/registry/client/transport/http_reader.go create mode 100644 vendor/github.com/docker/distribution/registry/client/transport/transport.go create mode 100644 vendor/github.com/docker/distribution/registry/storage/cache/cache.go create mode 100644 vendor/github.com/docker/distribution/registry/storage/cache/cachedblobdescriptorstore.go create mode 100644 vendor/github.com/docker/distribution/registry/storage/cache/memory/memory.go create mode 100644 vendor/github.com/docker/distribution/tags.go create mode 100644 vendor/github.com/docker/distribution/vendor.conf create mode 100644 vendor/github.com/docker/docker-credential-helpers/LICENSE create mode 100644 vendor/github.com/docker/docker-credential-helpers/client/client.go create mode 100644 vendor/github.com/docker/docker-credential-helpers/client/command.go create mode 100644 vendor/github.com/docker/docker-credential-helpers/credentials/credentials.go create mode 100644 vendor/github.com/docker/docker-credential-helpers/credentials/error.go create mode 100644 vendor/github.com/docker/docker-credential-helpers/credentials/helper.go create mode 100644 vendor/github.com/docker/docker-credential-helpers/credentials/version.go create mode 100644 vendor/github.com/docker/docker/api/types/container/waitcondition.go delete mode 100644 vendor/github.com/docker/docker/cli/config/configfile/file.go delete mode 100644 vendor/github.com/docker/docker/contrib/docker-engine-selinux/LICENSE delete mode 100644 vendor/github.com/docker/docker/contrib/syntax/vim/LICENSE create mode 100644 vendor/github.com/docker/docker/errdefs/defs.go create mode 100644 vendor/github.com/docker/docker/errdefs/doc.go create mode 100644 vendor/github.com/docker/docker/errdefs/helpers.go create mode 100644 vendor/github.com/docker/docker/errdefs/is.go create mode 100644 vendor/github.com/docker/docker/opts/address_pools.go create mode 100644 vendor/github.com/docker/docker/opts/env.go create mode 100644 vendor/github.com/docker/docker/opts/quotedstring.go create mode 100644 vendor/github.com/docker/docker/opts/runtime.go create mode 100644 vendor/github.com/docker/docker/opts/ulimit.go create mode 100644 vendor/github.com/docker/docker/pkg/archive/archive_linux.go create mode 100644 vendor/github.com/docker/docker/pkg/archive/archive_other.go create mode 100644 vendor/github.com/docker/docker/pkg/fileutils/fileutils_darwin.go delete mode 100644 vendor/github.com/docker/docker/pkg/fileutils/fileutils_solaris.go create mode 100644 vendor/github.com/docker/docker/pkg/homedir/homedir_linux.go create mode 100644 vendor/github.com/docker/docker/pkg/homedir/homedir_others.go rename vendor/github.com/docker/docker/pkg/homedir/{homedir.go => homedir_unix.go} (75%) create mode 100644 vendor/github.com/docker/docker/pkg/homedir/homedir_windows.go create mode 100644 vendor/github.com/docker/docker/pkg/idtools/utils_unix.go delete mode 100644 vendor/github.com/docker/docker/pkg/ioutils/fmt.go delete mode 100644 vendor/github.com/docker/docker/pkg/ioutils/multireader.go delete mode 100644 vendor/github.com/docker/docker/pkg/jsonlog/jsonlog.go delete mode 100644 vendor/github.com/docker/docker/pkg/jsonlog/jsonlog_marshalling.go delete mode 100644 vendor/github.com/docker/docker/pkg/jsonlog/jsonlogbytes.go delete mode 100644 vendor/github.com/docker/docker/pkg/jsonlog/time_marshalling.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/flags.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/flags_freebsd.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/flags_linux.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/flags_unsupported.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/mount.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/mounter_freebsd.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/mounter_linux.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/mounter_unsupported.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/mountinfo.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/mountinfo_freebsd.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/mountinfo_linux.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/mountinfo_unsupported.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/mountinfo_windows.go create mode 100644 vendor/github.com/docker/docker/pkg/mount/sharedsubtree_linux.go delete mode 100644 vendor/github.com/docker/docker/pkg/promise/promise.go delete mode 100644 vendor/github.com/docker/docker/pkg/random/random.go delete mode 100644 vendor/github.com/docker/docker/pkg/system/events_windows.go create mode 100644 vendor/github.com/docker/docker/pkg/system/init.go create mode 100644 vendor/github.com/docker/docker/pkg/system/init_unix.go create mode 100644 vendor/github.com/docker/docker/pkg/system/init_windows.go create mode 100644 vendor/github.com/docker/docker/pkg/system/lcow.go create mode 100644 vendor/github.com/docker/docker/pkg/system/lcow_unix.go create mode 100644 vendor/github.com/docker/docker/pkg/system/lcow_windows.go rename vendor/github.com/docker/docker/pkg/system/{lstat.go => lstat_unix.go} (84%) delete mode 100644 vendor/github.com/docker/docker/pkg/system/meminfo_solaris.go create mode 100644 vendor/github.com/docker/docker/pkg/system/path.go delete mode 100644 vendor/github.com/docker/docker/pkg/system/path_unix.go delete mode 100644 vendor/github.com/docker/docker/pkg/system/path_windows.go create mode 100644 vendor/github.com/docker/docker/pkg/system/rm.go rename vendor/github.com/docker/docker/pkg/system/{stat.go => stat_unix.go} (63%) delete mode 100644 vendor/github.com/docker/docker/pkg/system/stat_unsupported.go create mode 100644 vendor/github.com/docker/docker/pkg/tarsum/builder_context.go create mode 100644 vendor/github.com/docker/docker/pkg/tarsum/fileinfosums.go create mode 100644 vendor/github.com/docker/docker/pkg/tarsum/tarsum.go create mode 100644 vendor/github.com/docker/docker/pkg/tarsum/tarsum_spec.md create mode 100644 vendor/github.com/docker/docker/pkg/tarsum/versioning.go create mode 100644 vendor/github.com/docker/docker/pkg/tarsum/writercloser.go create mode 100644 vendor/github.com/docker/docker/pkg/term/proxy.go create mode 100644 vendor/github.com/docker/docker/pkg/term/tc.go delete mode 100644 vendor/github.com/docker/docker/pkg/term/tc_linux_cgo.go delete mode 100644 vendor/github.com/docker/docker/pkg/term/tc_other.go delete mode 100644 vendor/github.com/docker/docker/pkg/term/tc_solaris_cgo.go delete mode 100644 vendor/github.com/docker/docker/pkg/term/term_solaris.go delete mode 100644 vendor/github.com/docker/docker/pkg/term/term_unix.go create mode 100644 vendor/github.com/docker/docker/pkg/term/termios_bsd.go delete mode 100644 vendor/github.com/docker/docker/pkg/term/termios_darwin.go delete mode 100644 vendor/github.com/docker/docker/pkg/term/termios_freebsd.go delete mode 100644 vendor/github.com/docker/docker/pkg/term/termios_openbsd.go create mode 100644 vendor/github.com/docker/docker/pkg/term/winsize.go delete mode 100644 vendor/github.com/docker/docker/reference/reference.go create mode 100644 vendor/github.com/docker/docker/registry/config_unix.go create mode 100644 vendor/github.com/docker/docker/registry/config_windows.go create mode 100644 vendor/github.com/docker/docker/registry/endpoint_v1.go create mode 100644 vendor/github.com/docker/docker/registry/errors.go create mode 100644 vendor/github.com/docker/docker/registry/registry.go create mode 100644 vendor/github.com/docker/docker/registry/resumable/resumablerequestreader.go create mode 100644 vendor/github.com/docker/docker/registry/service_v1.go create mode 100644 vendor/github.com/docker/docker/registry/service_v2.go create mode 100644 vendor/github.com/docker/docker/registry/session.go create mode 100644 vendor/github.com/docker/go-connections/sockets/README.md create mode 100644 vendor/github.com/docker/go-connections/sockets/inmem_socket.go create mode 100644 vendor/github.com/docker/go-connections/sockets/proxy.go create mode 100644 vendor/github.com/docker/go-connections/sockets/sockets.go create mode 100644 vendor/github.com/docker/go-connections/sockets/sockets_unix.go create mode 100644 vendor/github.com/docker/go-connections/sockets/sockets_windows.go create mode 100644 vendor/github.com/docker/go-connections/sockets/tcp_socket.go create mode 100644 vendor/github.com/docker/go-connections/sockets/unix_socket.go create mode 100644 vendor/github.com/docker/go-connections/tlsconfig/certpool_go17.go create mode 100644 vendor/github.com/docker/go-connections/tlsconfig/certpool_other.go create mode 100644 vendor/github.com/docker/go-connections/tlsconfig/config.go create mode 100644 vendor/github.com/docker/go-connections/tlsconfig/config_client_ciphers.go create mode 100644 vendor/github.com/docker/go-connections/tlsconfig/config_legacy_client_ciphers.go create mode 100644 vendor/github.com/docker/go-metrics/CONTRIBUTING.md rename vendor/github.com/{opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/LICENSE => docker/go-metrics/LICENSE.code} (99%) create mode 100644 vendor/github.com/docker/go-metrics/LICENSE.docs create mode 100644 vendor/github.com/docker/go-metrics/NOTICE create mode 100644 vendor/github.com/docker/go-metrics/README.md create mode 100644 vendor/github.com/docker/go-metrics/counter.go create mode 100644 vendor/github.com/docker/go-metrics/docs.go create mode 100644 vendor/github.com/docker/go-metrics/gauge.go create mode 100644 vendor/github.com/docker/go-metrics/handler.go create mode 100644 vendor/github.com/docker/go-metrics/helpers.go create mode 100644 vendor/github.com/docker/go-metrics/namespace.go create mode 100644 vendor/github.com/docker/go-metrics/register.go create mode 100644 vendor/github.com/docker/go-metrics/timer.go create mode 100644 vendor/github.com/docker/go-metrics/unit.go create mode 100644 vendor/github.com/docker/libnetwork/LICENSE create mode 100644 vendor/github.com/docker/libnetwork/ipamutils/utils.go delete mode 100644 vendor/github.com/dustin/go-humanize/.gitignore delete mode 100644 vendor/github.com/go-ini/ini/.gitignore rename vendor/github.com/{opencontainers/runc/Godeps/_workspace/src/github.com => }/godbus/dbus/LICENSE (100%) create mode 100644 vendor/github.com/godbus/dbus/README.markdown create mode 100644 vendor/github.com/godbus/dbus/auth.go create mode 100644 vendor/github.com/godbus/dbus/auth_external.go create mode 100644 vendor/github.com/godbus/dbus/auth_sha1.go create mode 100644 vendor/github.com/godbus/dbus/call.go create mode 100644 vendor/github.com/godbus/dbus/conn.go create mode 100644 vendor/github.com/godbus/dbus/conn_darwin.go create mode 100644 vendor/github.com/godbus/dbus/conn_other.go create mode 100644 vendor/github.com/godbus/dbus/dbus.go create mode 100644 vendor/github.com/godbus/dbus/decoder.go create mode 100644 vendor/github.com/godbus/dbus/doc.go create mode 100644 vendor/github.com/godbus/dbus/encoder.go create mode 100644 vendor/github.com/godbus/dbus/export.go create mode 100644 vendor/github.com/godbus/dbus/homedir.go create mode 100644 vendor/github.com/godbus/dbus/homedir_dynamic.go create mode 100644 vendor/github.com/godbus/dbus/homedir_static.go create mode 100644 vendor/github.com/godbus/dbus/message.go create mode 100644 vendor/github.com/godbus/dbus/object.go create mode 100644 vendor/github.com/godbus/dbus/sig.go create mode 100644 vendor/github.com/godbus/dbus/transport_darwin.go create mode 100644 vendor/github.com/godbus/dbus/transport_generic.go create mode 100644 vendor/github.com/godbus/dbus/transport_unix.go create mode 100644 vendor/github.com/godbus/dbus/transport_unixcred_dragonfly.go create mode 100644 vendor/github.com/godbus/dbus/transport_unixcred_linux.go create mode 100644 vendor/github.com/godbus/dbus/variant.go create mode 100644 vendor/github.com/godbus/dbus/variant_lexer.go create mode 100644 vendor/github.com/godbus/dbus/variant_parser.go rename vendor/github.com/{opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/mflag => gorilla/context}/LICENSE (82%) create mode 100644 vendor/github.com/gorilla/context/README.md create mode 100644 vendor/github.com/gorilla/context/context.go create mode 100644 vendor/github.com/gorilla/context/doc.go create mode 100644 vendor/github.com/gorilla/mux/AUTHORS create mode 100644 vendor/github.com/gorilla/mux/ISSUE_TEMPLATE.md rename vendor/github.com/{docker/docker/pkg/symlink/LICENSE.BSD => gorilla/mux/LICENSE} (82%) create mode 100644 vendor/github.com/gorilla/mux/README.md create mode 100644 vendor/github.com/gorilla/mux/context_gorilla.go create mode 100644 vendor/github.com/gorilla/mux/context_native.go create mode 100644 vendor/github.com/gorilla/mux/doc.go create mode 100644 vendor/github.com/gorilla/mux/middleware.go create mode 100644 vendor/github.com/gorilla/mux/mux.go create mode 100644 vendor/github.com/gorilla/mux/regexp.go create mode 100644 vendor/github.com/gorilla/mux/route.go create mode 100644 vendor/github.com/gorilla/mux/test_helpers.go mode change 100644 => 100755 vendor/github.com/hashicorp/go-msgpack/codec/msgpack_test.py delete mode 100644 vendor/github.com/hashicorp/go-syslog/.gitignore create mode 100644 vendor/github.com/hashicorp/hcl/appveyor.yml delete mode 100644 vendor/github.com/hashicorp/logutils/.gitignore create mode 100644 vendor/github.com/hashicorp/vault/main.go delete mode 100644 vendor/github.com/jmespath/go-jmespath/.gitignore delete mode 100644 vendor/github.com/jmespath/go-jmespath/.travis.yml delete mode 100644 vendor/github.com/miekg/dns/.gitignore delete mode 100644 vendor/github.com/miekg/dns/.travis.yml delete mode 100644 vendor/github.com/mitchellh/mapstructure/.travis.yml rename vendor/github.com/{opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/symlink/LICENSE.APACHE => mrunalp/fileutils/LICENSE} (99%) create mode 100644 vendor/github.com/mrunalp/fileutils/README.md create mode 100644 vendor/github.com/mrunalp/fileutils/fileutils.go create mode 100644 vendor/github.com/mrunalp/fileutils/idtools.go delete mode 100644 vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/Sirupsen/logrus/LICENSE delete mode 100644 vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/codegangsta/cli/LICENSE delete mode 100644 vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/contrib/syntax/vim/LICENSE delete mode 100644 vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/docs/project/images/red_notice.png delete mode 100644 vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/symlink/LICENSE.BSD delete mode 100644 vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/golang/protobuf/LICENSE delete mode 100644 vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/syndtr/gocapability/LICENSE create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/README.md create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/SPEC.md create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/apparmor/apparmor.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/apparmor/apparmor_disabled.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/capabilities_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/cgroups/systemd/apply_nosystemd.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/cgroups/systemd/apply_systemd.go rename vendor/github.com/opencontainers/runc/libcontainer/configs/{cgroup_unix.go => cgroup_linux.go} (88%) delete mode 100644 vendor/github.com/opencontainers/runc/libcontainer/configs/cgroup_unsupported.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/configs/config_linux.go delete mode 100644 vendor/github.com/opencontainers/runc/libcontainer/configs/config_unix.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/configs/intelrdt.go rename vendor/github.com/opencontainers/runc/libcontainer/configs/{namespaces_unix.go => namespaces_linux.go} (89%) create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/configs/validate/rootless.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/configs/validate/validator.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/console_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/container.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/container_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/criu_opts_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/criurpc/Makefile create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/criurpc/criurpc.pb.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/criurpc/criurpc.proto create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/error.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/factory.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/factory_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/generic_error.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/init_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/intelrdt/intelrdt.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/intelrdt/stats.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/keys/keyctl.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/message_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/mount/mount.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/mount/mount_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/mount/mountinfo.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/network_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/notify_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/process.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/process_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/restored_process.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/rootfs_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/seccomp/config.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/seccomp/seccomp_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/seccomp/seccomp_unsupported.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/setns_init_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/stacktrace/capture.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/stacktrace/frame.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/stacktrace/stacktrace.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/standard_init_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/state_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/stats.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/stats_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/sync.go delete mode 100644 vendor/github.com/opencontainers/runc/libcontainer/system/setns_linux.go rename vendor/github.com/opencontainers/runc/libcontainer/system/{syscall_linux_arm.go => syscall_linux_32.go} (61%) delete mode 100644 vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_386.go delete mode 100644 vendor/github.com/opencontainers/runc/libcontainer/user/lookup_unsupported.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/user/lookup_windows.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/utils/cmsg.go rename vendor/github.com/opencontainers/{runc/Godeps/_workspace/src/github.com/opencontainers/specs => runtime-spec}/LICENSE (100%) create mode 100644 vendor/github.com/opencontainers/runtime-spec/specs-go/config.go create mode 100644 vendor/github.com/opencontainers/runtime-spec/specs-go/state.go create mode 100644 vendor/github.com/opencontainers/runtime-spec/specs-go/version.go create mode 100644 vendor/github.com/opencontainers/selinux/LICENSE create mode 100644 vendor/github.com/opencontainers/selinux/go-selinux/label/label.go create mode 100644 vendor/github.com/opencontainers/selinux/go-selinux/label/label_selinux.go create mode 100644 vendor/github.com/opencontainers/selinux/go-selinux/selinux.go create mode 100644 vendor/github.com/opencontainers/selinux/go-selinux/xattrs.go create mode 100644 vendor/github.com/ryanuber/columnize/LICENSE rename vendor/github.com/{opencontainers/runc/Godeps/_workspace/src/github.com => }/seccomp/libseccomp-golang/LICENSE (100%) create mode 100644 vendor/github.com/seccomp/libseccomp-golang/README create mode 100644 vendor/github.com/seccomp/libseccomp-golang/seccomp.go create mode 100644 vendor/github.com/seccomp/libseccomp-golang/seccomp_internal.go rename vendor/github.com/{Sirupsen => sirupsen}/logrus/CHANGELOG.md (59%) rename vendor/github.com/{Sirupsen => sirupsen}/logrus/LICENSE (100%) rename vendor/github.com/{Sirupsen => sirupsen}/logrus/README.md (63%) create mode 100644 vendor/github.com/sirupsen/logrus/alt_exit.go create mode 100644 vendor/github.com/sirupsen/logrus/appveyor.yml rename vendor/github.com/{Sirupsen => sirupsen}/logrus/doc.go (83%) rename vendor/github.com/{Sirupsen => sirupsen}/logrus/entry.go (68%) rename vendor/github.com/{Sirupsen => sirupsen}/logrus/exported.go (82%) rename vendor/github.com/{Sirupsen => sirupsen}/logrus/formatter.go (69%) rename vendor/github.com/{Sirupsen => sirupsen}/logrus/hooks.go (100%) create mode 100644 vendor/github.com/sirupsen/logrus/json_formatter.go create mode 100644 vendor/github.com/sirupsen/logrus/logger.go rename vendor/github.com/{Sirupsen => sirupsen}/logrus/logrus.go (95%) create mode 100644 vendor/github.com/sirupsen/logrus/terminal_bsd.go create mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_appengine.go create mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_js.go create mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_notappengine.go rename vendor/github.com/{Sirupsen => sirupsen}/logrus/terminal_linux.go (62%) create mode 100644 vendor/github.com/sirupsen/logrus/text_formatter.go create mode 100644 vendor/github.com/sirupsen/logrus/writer.go delete mode 100644 vendor/github.com/ulikunitz/xz/example.go rename vendor/github.com/{opencontainers/runc/Godeps/_workspace/src/github.com => }/vishvananda/netlink/LICENSE (100%) create mode 100644 vendor/github.com/vishvananda/netlink/README.md create mode 100644 vendor/github.com/vishvananda/netlink/addr.go create mode 100644 vendor/github.com/vishvananda/netlink/addr_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/filter.go create mode 100644 vendor/github.com/vishvananda/netlink/filter_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/link.go create mode 100644 vendor/github.com/vishvananda/netlink/link_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/neigh.go create mode 100644 vendor/github.com/vishvananda/netlink/neigh_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/netlink.go create mode 100644 vendor/github.com/vishvananda/netlink/netlink_unspecified.go create mode 100644 vendor/github.com/vishvananda/netlink/nl/addr_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/nl/link_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/nl/nl_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/nl/route_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/nl/tc_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/nl/xfrm_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/nl/xfrm_policy_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/nl/xfrm_state_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/protinfo.go create mode 100644 vendor/github.com/vishvananda/netlink/protinfo_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/qdisc.go create mode 100644 vendor/github.com/vishvananda/netlink/qdisc_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/route.go create mode 100644 vendor/github.com/vishvananda/netlink/route_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/xfrm.go create mode 100644 vendor/github.com/vishvananda/netlink/xfrm_policy.go create mode 100644 vendor/github.com/vishvananda/netlink/xfrm_policy_linux.go create mode 100644 vendor/github.com/vishvananda/netlink/xfrm_state.go create mode 100644 vendor/github.com/vishvananda/netlink/xfrm_state_linux.go create mode 100644 vendor/golang.org/x/net/proxy/direct.go create mode 100644 vendor/golang.org/x/net/proxy/per_host.go create mode 100644 vendor/golang.org/x/net/proxy/proxy.go create mode 100644 vendor/golang.org/x/net/proxy/socks5.go delete mode 100644 vendor/golang.org/x/sys/unix/creds_test.go delete mode 100644 vendor/golang.org/x/sys/unix/dev_linux_test.go delete mode 100644 vendor/golang.org/x/sys/unix/example_test.go delete mode 100644 vendor/golang.org/x/sys/unix/export_test.go delete mode 100644 vendor/golang.org/x/sys/unix/mkpost.go delete mode 100644 vendor/golang.org/x/sys/unix/mmap_unix_test.go delete mode 100644 vendor/golang.org/x/sys/unix/openbsd_test.go delete mode 100644 vendor/golang.org/x/sys/unix/syscall_bsd_test.go delete mode 100644 vendor/golang.org/x/sys/unix/syscall_darwin_test.go delete mode 100644 vendor/golang.org/x/sys/unix/syscall_freebsd_test.go delete mode 100644 vendor/golang.org/x/sys/unix/syscall_linux_test.go delete mode 100644 vendor/golang.org/x/sys/unix/syscall_solaris_test.go delete mode 100644 vendor/golang.org/x/sys/unix/syscall_test.go delete mode 100644 vendor/golang.org/x/sys/unix/syscall_unix_test.go delete mode 100644 vendor/golang.org/x/sys/unix/timestruct_test.go delete mode 100644 vendor/golang.org/x/sys/unix/types_dragonfly.go delete mode 100644 vendor/golang.org/x/sys/unix/types_freebsd.go delete mode 100644 vendor/golang.org/x/sys/unix/types_netbsd.go delete mode 100644 vendor/golang.org/x/sys/unix/types_openbsd.go delete mode 100644 vendor/golang.org/x/sys/unix/xattr_test.go diff --git a/client/driver/docker.go b/client/driver/docker.go index a35f3f648..4680dbdc7 100644 --- a/client/driver/docker.go +++ b/client/driver/docker.go @@ -17,15 +17,15 @@ import ( "time" "github.com/armon/circbuf" - "github.com/fsouza/go-dockerclient" + metrics "github.com/armon/go-metrics" + docker "github.com/fsouza/go-dockerclient" - "github.com/docker/docker/cli/config/configfile" - "github.com/docker/docker/reference" + "github.com/docker/cli/cli/config/configfile" + "github.com/docker/distribution/reference" "github.com/docker/docker/registry" - "github.com/armon/go-metrics" "github.com/hashicorp/go-multierror" - "github.com/hashicorp/go-plugin" + plugin "github.com/hashicorp/go-plugin" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/driver/env" "github.com/hashicorp/nomad/client/driver/executor" @@ -2190,13 +2190,13 @@ func authFromHelper(helperName string) authBackend { helper := dockerAuthHelperPrefix + helperName cmd := exec.Command(helper, "get") - repoParsed, err := reference.ParseNamed(repo) + repoInfo, err := parseRepositoryInfo(repo) if err != nil { return nil, err } // Ensure that the HTTPs prefix exists - repoAddr := fmt.Sprintf("https://%s", repoParsed.Hostname()) + repoAddr := fmt.Sprintf("https://%s", repoInfo.Index.Name) cmd.Stdin = strings.NewReader(repoAddr) output, err := cmd.Output() diff --git a/client/driver/executor/executor_linux.go b/client/driver/executor/executor_linux.go index da7e7a239..7a94d4b69 100644 --- a/client/driver/executor/executor_linux.go +++ b/client/driver/executor/executor_linux.go @@ -129,7 +129,8 @@ func (e *UniversalExecutor) configureCgroups(resources *structs.Resources) error e.resConCtx.groups.Path = filepath.Join("/nomad", cgroupName) // Allow access to /dev/ - e.resConCtx.groups.Resources.AllowAllDevices = true + allowAllDevs := true + e.resConCtx.groups.Resources.AllowAllDevices = &allowAllDevs // Use a cgroup but don't apply limits if !e.command.ResourceLimits { @@ -140,7 +141,7 @@ func (e *UniversalExecutor) configureCgroups(resources *structs.Resources) error // Total amount of memory allowed to consume e.resConCtx.groups.Resources.Memory = int64(resources.MemoryMB * 1024 * 1024) // Disable swap to avoid issues on the machine - var memSwappiness int64 = 0 + var memSwappiness uint64 = 0 e.resConCtx.groups.Resources.MemorySwappiness = &memSwappiness } @@ -149,7 +150,7 @@ func (e *UniversalExecutor) configureCgroups(resources *structs.Resources) error } // Set the relative CPU shares for this cgroup. - e.resConCtx.groups.Resources.CpuShares = int64(resources.CPU) + e.resConCtx.groups.Resources.CpuShares = uint64(resources.CPU) if resources.IOPS != 0 { // Validate it is in an acceptable range. diff --git a/vendor/github.com/docker/docker/pkg/symlink/LICENSE.APACHE b/vendor/github.com/Azure/azure-sdk-for-go/LICENSE similarity index 94% rename from vendor/github.com/docker/docker/pkg/symlink/LICENSE.APACHE rename to vendor/github.com/Azure/azure-sdk-for-go/LICENSE index 34c4ea7c5..af39a91e7 100644 --- a/vendor/github.com/docker/docker/pkg/symlink/LICENSE.APACHE +++ b/vendor/github.com/Azure/azure-sdk-for-go/LICENSE @@ -176,7 +176,18 @@ END OF TERMS AND CONDITIONS - Copyright 2014-2016 Docker, Inc. + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "[]" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright 2016 Microsoft Corporation Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. diff --git a/vendor/github.com/Azure/azure-sdk-for-go/NOTICE b/vendor/github.com/Azure/azure-sdk-for-go/NOTICE new file mode 100644 index 000000000..2d1d72608 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/NOTICE @@ -0,0 +1,5 @@ +Microsoft Azure-SDK-for-Go +Copyright 2014-2017 Microsoft + +This product includes software developed at +the Microsoft Corporation (https://www.microsoft.com). diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/applicationgateways.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/applicationgateways.go new file mode 100644 index 000000000..dd4757ae8 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/applicationgateways.go @@ -0,0 +1,566 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// ApplicationGatewaysClient is the network Client +type ApplicationGatewaysClient struct { + BaseClient +} + +// NewApplicationGatewaysClient creates an instance of the ApplicationGatewaysClient client. +func NewApplicationGatewaysClient(subscriptionID string) ApplicationGatewaysClient { + return NewApplicationGatewaysClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewApplicationGatewaysClientWithBaseURI creates an instance of the ApplicationGatewaysClient client. +func NewApplicationGatewaysClientWithBaseURI(baseURI string, subscriptionID string) ApplicationGatewaysClient { + return ApplicationGatewaysClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates the specified application gateway. +// Parameters: +// resourceGroupName - the name of the resource group. +// applicationGatewayName - the name of the application gateway. +// parameters - parameters supplied to the create or update application gateway operation. +func (client ApplicationGatewaysClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, applicationGatewayName string, parameters ApplicationGateway) (result ApplicationGatewaysCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, applicationGatewayName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client ApplicationGatewaysClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, applicationGatewayName string, parameters ApplicationGateway) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "applicationGatewayName": autorest.Encode("path", applicationGatewayName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/applicationGateways/{applicationGatewayName}", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client ApplicationGatewaysClient) CreateOrUpdateSender(req *http.Request) (future ApplicationGatewaysCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client ApplicationGatewaysClient) CreateOrUpdateResponder(resp *http.Response) (result ApplicationGateway, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified application gateway. +// Parameters: +// resourceGroupName - the name of the resource group. +// applicationGatewayName - the name of the application gateway. +func (client ApplicationGatewaysClient) Delete(ctx context.Context, resourceGroupName string, applicationGatewayName string) (result ApplicationGatewaysDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, applicationGatewayName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client ApplicationGatewaysClient) DeletePreparer(ctx context.Context, resourceGroupName string, applicationGatewayName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "applicationGatewayName": autorest.Encode("path", applicationGatewayName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/applicationGateways/{applicationGatewayName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client ApplicationGatewaysClient) DeleteSender(req *http.Request) (future ApplicationGatewaysDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client ApplicationGatewaysClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified application gateway. +// Parameters: +// resourceGroupName - the name of the resource group. +// applicationGatewayName - the name of the application gateway. +func (client ApplicationGatewaysClient) Get(ctx context.Context, resourceGroupName string, applicationGatewayName string) (result ApplicationGateway, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, applicationGatewayName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client ApplicationGatewaysClient) GetPreparer(ctx context.Context, resourceGroupName string, applicationGatewayName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "applicationGatewayName": autorest.Encode("path", applicationGatewayName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/applicationGateways/{applicationGatewayName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client ApplicationGatewaysClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client ApplicationGatewaysClient) GetResponder(resp *http.Response) (result ApplicationGateway, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List lists all application gateways in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +func (client ApplicationGatewaysClient) List(ctx context.Context, resourceGroupName string) (result ApplicationGatewayListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.aglr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "List", resp, "Failure sending request") + return + } + + result.aglr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client ApplicationGatewaysClient) ListPreparer(ctx context.Context, resourceGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/applicationGateways", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client ApplicationGatewaysClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client ApplicationGatewaysClient) ListResponder(resp *http.Response) (result ApplicationGatewayListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client ApplicationGatewaysClient) listNextResults(lastResults ApplicationGatewayListResult) (result ApplicationGatewayListResult, err error) { + req, err := lastResults.applicationGatewayListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client ApplicationGatewaysClient) ListComplete(ctx context.Context, resourceGroupName string) (result ApplicationGatewayListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName) + return +} + +// ListAll gets all the application gateways in a subscription. +func (client ApplicationGatewaysClient) ListAll(ctx context.Context) (result ApplicationGatewayListResultPage, err error) { + result.fn = client.listAllNextResults + req, err := client.ListAllPreparer(ctx) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "ListAll", nil, "Failure preparing request") + return + } + + resp, err := client.ListAllSender(req) + if err != nil { + result.aglr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "ListAll", resp, "Failure sending request") + return + } + + result.aglr, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "ListAll", resp, "Failure responding to request") + } + + return +} + +// ListAllPreparer prepares the ListAll request. +func (client ApplicationGatewaysClient) ListAllPreparer(ctx context.Context) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/Microsoft.Network/applicationGateways", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListAllSender sends the ListAll request. The method will close the +// http.Response Body if it receives an error. +func (client ApplicationGatewaysClient) ListAllSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListAllResponder handles the response to the ListAll request. The method always +// closes the http.Response Body. +func (client ApplicationGatewaysClient) ListAllResponder(resp *http.Response) (result ApplicationGatewayListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listAllNextResults retrieves the next set of results, if any. +func (client ApplicationGatewaysClient) listAllNextResults(lastResults ApplicationGatewayListResult) (result ApplicationGatewayListResult, err error) { + req, err := lastResults.applicationGatewayListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "listAllNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListAllSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "listAllNextResults", resp, "Failure sending next results request") + } + result, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "listAllNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListAllComplete enumerates all values, automatically crossing page boundaries as required. +func (client ApplicationGatewaysClient) ListAllComplete(ctx context.Context) (result ApplicationGatewayListResultIterator, err error) { + result.page, err = client.ListAll(ctx) + return +} + +// Start starts the specified application gateway. +// Parameters: +// resourceGroupName - the name of the resource group. +// applicationGatewayName - the name of the application gateway. +func (client ApplicationGatewaysClient) Start(ctx context.Context, resourceGroupName string, applicationGatewayName string) (result ApplicationGatewaysStartFuture, err error) { + req, err := client.StartPreparer(ctx, resourceGroupName, applicationGatewayName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "Start", nil, "Failure preparing request") + return + } + + result, err = client.StartSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "Start", result.Response(), "Failure sending request") + return + } + + return +} + +// StartPreparer prepares the Start request. +func (client ApplicationGatewaysClient) StartPreparer(ctx context.Context, resourceGroupName string, applicationGatewayName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "applicationGatewayName": autorest.Encode("path", applicationGatewayName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/applicationGateways/{applicationGatewayName}/start", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// StartSender sends the Start request. The method will close the +// http.Response Body if it receives an error. +func (client ApplicationGatewaysClient) StartSender(req *http.Request) (future ApplicationGatewaysStartFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted)) + return +} + +// StartResponder handles the response to the Start request. The method always +// closes the http.Response Body. +func (client ApplicationGatewaysClient) StartResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// Stop stops the specified application gateway in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// applicationGatewayName - the name of the application gateway. +func (client ApplicationGatewaysClient) Stop(ctx context.Context, resourceGroupName string, applicationGatewayName string) (result ApplicationGatewaysStopFuture, err error) { + req, err := client.StopPreparer(ctx, resourceGroupName, applicationGatewayName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "Stop", nil, "Failure preparing request") + return + } + + result, err = client.StopSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysClient", "Stop", result.Response(), "Failure sending request") + return + } + + return +} + +// StopPreparer prepares the Stop request. +func (client ApplicationGatewaysClient) StopPreparer(ctx context.Context, resourceGroupName string, applicationGatewayName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "applicationGatewayName": autorest.Encode("path", applicationGatewayName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/applicationGateways/{applicationGatewayName}/stop", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// StopSender sends the Stop request. The method will close the +// http.Response Body if it receives an error. +func (client ApplicationGatewaysClient) StopSender(req *http.Request) (future ApplicationGatewaysStopFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted)) + return +} + +// StopResponder handles the response to the Stop request. The method always +// closes the http.Response Body. +func (client ApplicationGatewaysClient) StopResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/client.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/client.go new file mode 100644 index 000000000..f487b26a8 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/client.go @@ -0,0 +1,124 @@ +// Package network implements the Azure ARM Network service API version 2015-06-15. +// +// Network Client +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +const ( + // DefaultBaseURI is the default URI used for the service Network + DefaultBaseURI = "https://management.azure.com" +) + +// BaseClient is the base client for Network. +type BaseClient struct { + autorest.Client + BaseURI string + SubscriptionID string +} + +// New creates an instance of the BaseClient client. +func New(subscriptionID string) BaseClient { + return NewWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewWithBaseURI creates an instance of the BaseClient client. +func NewWithBaseURI(baseURI string, subscriptionID string) BaseClient { + return BaseClient{ + Client: autorest.NewClientWithUserAgent(UserAgent()), + BaseURI: baseURI, + SubscriptionID: subscriptionID, + } +} + +// CheckDNSNameAvailability checks whether a domain name in the cloudapp.net zone is available for use. +// Parameters: +// location - the location of the domain name. +// domainNameLabel - the domain name to be verified. It must conform to the following regular expression: +// ^[a-z][a-z0-9-]{1,61}[a-z0-9]$. +func (client BaseClient) CheckDNSNameAvailability(ctx context.Context, location string, domainNameLabel string) (result DNSNameAvailabilityResult, err error) { + req, err := client.CheckDNSNameAvailabilityPreparer(ctx, location, domainNameLabel) + if err != nil { + err = autorest.NewErrorWithError(err, "network.BaseClient", "CheckDNSNameAvailability", nil, "Failure preparing request") + return + } + + resp, err := client.CheckDNSNameAvailabilitySender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.BaseClient", "CheckDNSNameAvailability", resp, "Failure sending request") + return + } + + result, err = client.CheckDNSNameAvailabilityResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.BaseClient", "CheckDNSNameAvailability", resp, "Failure responding to request") + } + + return +} + +// CheckDNSNameAvailabilityPreparer prepares the CheckDNSNameAvailability request. +func (client BaseClient) CheckDNSNameAvailabilityPreparer(ctx context.Context, location string, domainNameLabel string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "location": autorest.Encode("path", location), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(domainNameLabel) > 0 { + queryParameters["domainNameLabel"] = autorest.Encode("query", domainNameLabel) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/Microsoft.Network/locations/{location}/CheckDnsNameAvailability", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CheckDNSNameAvailabilitySender sends the CheckDNSNameAvailability request. The method will close the +// http.Response Body if it receives an error. +func (client BaseClient) CheckDNSNameAvailabilitySender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// CheckDNSNameAvailabilityResponder handles the response to the CheckDNSNameAvailability request. The method always +// closes the http.Response Body. +func (client BaseClient) CheckDNSNameAvailabilityResponder(resp *http.Response) (result DNSNameAvailabilityResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuitauthorizations.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuitauthorizations.go new file mode 100644 index 000000000..e9b3f1027 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuitauthorizations.go @@ -0,0 +1,351 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// ExpressRouteCircuitAuthorizationsClient is the network Client +type ExpressRouteCircuitAuthorizationsClient struct { + BaseClient +} + +// NewExpressRouteCircuitAuthorizationsClient creates an instance of the ExpressRouteCircuitAuthorizationsClient +// client. +func NewExpressRouteCircuitAuthorizationsClient(subscriptionID string) ExpressRouteCircuitAuthorizationsClient { + return NewExpressRouteCircuitAuthorizationsClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewExpressRouteCircuitAuthorizationsClientWithBaseURI creates an instance of the +// ExpressRouteCircuitAuthorizationsClient client. +func NewExpressRouteCircuitAuthorizationsClientWithBaseURI(baseURI string, subscriptionID string) ExpressRouteCircuitAuthorizationsClient { + return ExpressRouteCircuitAuthorizationsClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates an authorization in the specified express route circuit. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the express route circuit. +// authorizationName - the name of the authorization. +// authorizationParameters - parameters supplied to the create or update express route circuit authorization +// operation. +func (client ExpressRouteCircuitAuthorizationsClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, circuitName string, authorizationName string, authorizationParameters ExpressRouteCircuitAuthorization) (result ExpressRouteCircuitAuthorizationsCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, circuitName, authorizationName, authorizationParameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client ExpressRouteCircuitAuthorizationsClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, circuitName string, authorizationName string, authorizationParameters ExpressRouteCircuitAuthorization) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "authorizationName": autorest.Encode("path", authorizationName), + "circuitName": autorest.Encode("path", circuitName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}/authorizations/{authorizationName}", pathParameters), + autorest.WithJSON(authorizationParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitAuthorizationsClient) CreateOrUpdateSender(req *http.Request) (future ExpressRouteCircuitAuthorizationsCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitAuthorizationsClient) CreateOrUpdateResponder(resp *http.Response) (result ExpressRouteCircuitAuthorization, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified authorization from the specified express route circuit. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the express route circuit. +// authorizationName - the name of the authorization. +func (client ExpressRouteCircuitAuthorizationsClient) Delete(ctx context.Context, resourceGroupName string, circuitName string, authorizationName string) (result ExpressRouteCircuitAuthorizationsDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, circuitName, authorizationName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client ExpressRouteCircuitAuthorizationsClient) DeletePreparer(ctx context.Context, resourceGroupName string, circuitName string, authorizationName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "authorizationName": autorest.Encode("path", authorizationName), + "circuitName": autorest.Encode("path", circuitName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}/authorizations/{authorizationName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitAuthorizationsClient) DeleteSender(req *http.Request) (future ExpressRouteCircuitAuthorizationsDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitAuthorizationsClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified authorization from the specified express route circuit. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the express route circuit. +// authorizationName - the name of the authorization. +func (client ExpressRouteCircuitAuthorizationsClient) Get(ctx context.Context, resourceGroupName string, circuitName string, authorizationName string) (result ExpressRouteCircuitAuthorization, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, circuitName, authorizationName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client ExpressRouteCircuitAuthorizationsClient) GetPreparer(ctx context.Context, resourceGroupName string, circuitName string, authorizationName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "authorizationName": autorest.Encode("path", authorizationName), + "circuitName": autorest.Encode("path", circuitName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}/authorizations/{authorizationName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitAuthorizationsClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitAuthorizationsClient) GetResponder(resp *http.Response) (result ExpressRouteCircuitAuthorization, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all authorizations in an express route circuit. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the circuit. +func (client ExpressRouteCircuitAuthorizationsClient) List(ctx context.Context, resourceGroupName string, circuitName string) (result AuthorizationListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName, circuitName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.alr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "List", resp, "Failure sending request") + return + } + + result.alr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client ExpressRouteCircuitAuthorizationsClient) ListPreparer(ctx context.Context, resourceGroupName string, circuitName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "circuitName": autorest.Encode("path", circuitName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}/authorizations", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitAuthorizationsClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitAuthorizationsClient) ListResponder(resp *http.Response) (result AuthorizationListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client ExpressRouteCircuitAuthorizationsClient) listNextResults(lastResults AuthorizationListResult) (result AuthorizationListResult, err error) { + req, err := lastResults.authorizationListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client ExpressRouteCircuitAuthorizationsClient) ListComplete(ctx context.Context, resourceGroupName string, circuitName string) (result AuthorizationListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName, circuitName) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuitpeerings.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuitpeerings.go new file mode 100644 index 000000000..84e7d0b2c --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuitpeerings.go @@ -0,0 +1,348 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// ExpressRouteCircuitPeeringsClient is the network Client +type ExpressRouteCircuitPeeringsClient struct { + BaseClient +} + +// NewExpressRouteCircuitPeeringsClient creates an instance of the ExpressRouteCircuitPeeringsClient client. +func NewExpressRouteCircuitPeeringsClient(subscriptionID string) ExpressRouteCircuitPeeringsClient { + return NewExpressRouteCircuitPeeringsClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewExpressRouteCircuitPeeringsClientWithBaseURI creates an instance of the ExpressRouteCircuitPeeringsClient client. +func NewExpressRouteCircuitPeeringsClientWithBaseURI(baseURI string, subscriptionID string) ExpressRouteCircuitPeeringsClient { + return ExpressRouteCircuitPeeringsClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a peering in the specified express route circuits. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the express route circuit. +// peeringName - the name of the peering. +// peeringParameters - parameters supplied to the create or update express route circuit peering operation. +func (client ExpressRouteCircuitPeeringsClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, circuitName string, peeringName string, peeringParameters ExpressRouteCircuitPeering) (result ExpressRouteCircuitPeeringsCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, circuitName, peeringName, peeringParameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client ExpressRouteCircuitPeeringsClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, circuitName string, peeringName string, peeringParameters ExpressRouteCircuitPeering) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "circuitName": autorest.Encode("path", circuitName), + "peeringName": autorest.Encode("path", peeringName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}/peerings/{peeringName}", pathParameters), + autorest.WithJSON(peeringParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitPeeringsClient) CreateOrUpdateSender(req *http.Request) (future ExpressRouteCircuitPeeringsCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitPeeringsClient) CreateOrUpdateResponder(resp *http.Response) (result ExpressRouteCircuitPeering, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified peering from the specified express route circuit. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the express route circuit. +// peeringName - the name of the peering. +func (client ExpressRouteCircuitPeeringsClient) Delete(ctx context.Context, resourceGroupName string, circuitName string, peeringName string) (result ExpressRouteCircuitPeeringsDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, circuitName, peeringName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client ExpressRouteCircuitPeeringsClient) DeletePreparer(ctx context.Context, resourceGroupName string, circuitName string, peeringName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "circuitName": autorest.Encode("path", circuitName), + "peeringName": autorest.Encode("path", peeringName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}/peerings/{peeringName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitPeeringsClient) DeleteSender(req *http.Request) (future ExpressRouteCircuitPeeringsDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitPeeringsClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified authorization from the specified express route circuit. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the express route circuit. +// peeringName - the name of the peering. +func (client ExpressRouteCircuitPeeringsClient) Get(ctx context.Context, resourceGroupName string, circuitName string, peeringName string) (result ExpressRouteCircuitPeering, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, circuitName, peeringName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client ExpressRouteCircuitPeeringsClient) GetPreparer(ctx context.Context, resourceGroupName string, circuitName string, peeringName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "circuitName": autorest.Encode("path", circuitName), + "peeringName": autorest.Encode("path", peeringName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}/peerings/{peeringName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitPeeringsClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitPeeringsClient) GetResponder(resp *http.Response) (result ExpressRouteCircuitPeering, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all peerings in a specified express route circuit. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the express route circuit. +func (client ExpressRouteCircuitPeeringsClient) List(ctx context.Context, resourceGroupName string, circuitName string) (result ExpressRouteCircuitPeeringListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName, circuitName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.ercplr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "List", resp, "Failure sending request") + return + } + + result.ercplr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client ExpressRouteCircuitPeeringsClient) ListPreparer(ctx context.Context, resourceGroupName string, circuitName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "circuitName": autorest.Encode("path", circuitName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}/peerings", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitPeeringsClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitPeeringsClient) ListResponder(resp *http.Response) (result ExpressRouteCircuitPeeringListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client ExpressRouteCircuitPeeringsClient) listNextResults(lastResults ExpressRouteCircuitPeeringListResult) (result ExpressRouteCircuitPeeringListResult, err error) { + req, err := lastResults.expressRouteCircuitPeeringListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client ExpressRouteCircuitPeeringsClient) ListComplete(ctx context.Context, resourceGroupName string, circuitName string) (result ExpressRouteCircuitPeeringListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName, circuitName) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuits.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuits.go new file mode 100644 index 000000000..1b6246cad --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressroutecircuits.go @@ -0,0 +1,718 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// ExpressRouteCircuitsClient is the network Client +type ExpressRouteCircuitsClient struct { + BaseClient +} + +// NewExpressRouteCircuitsClient creates an instance of the ExpressRouteCircuitsClient client. +func NewExpressRouteCircuitsClient(subscriptionID string) ExpressRouteCircuitsClient { + return NewExpressRouteCircuitsClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewExpressRouteCircuitsClientWithBaseURI creates an instance of the ExpressRouteCircuitsClient client. +func NewExpressRouteCircuitsClientWithBaseURI(baseURI string, subscriptionID string) ExpressRouteCircuitsClient { + return ExpressRouteCircuitsClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates an express route circuit. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the circuit. +// parameters - parameters supplied to the create or update express route circuit operation. +func (client ExpressRouteCircuitsClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, circuitName string, parameters ExpressRouteCircuit) (result ExpressRouteCircuitsCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, circuitName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client ExpressRouteCircuitsClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, circuitName string, parameters ExpressRouteCircuit) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "circuitName": autorest.Encode("path", circuitName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitsClient) CreateOrUpdateSender(req *http.Request) (future ExpressRouteCircuitsCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitsClient) CreateOrUpdateResponder(resp *http.Response) (result ExpressRouteCircuit, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified express route circuit. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the express route circuit. +func (client ExpressRouteCircuitsClient) Delete(ctx context.Context, resourceGroupName string, circuitName string) (result ExpressRouteCircuitsDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, circuitName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client ExpressRouteCircuitsClient) DeletePreparer(ctx context.Context, resourceGroupName string, circuitName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "circuitName": autorest.Encode("path", circuitName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitsClient) DeleteSender(req *http.Request) (future ExpressRouteCircuitsDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitsClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets information about the specified express route circuit. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of express route circuit. +func (client ExpressRouteCircuitsClient) Get(ctx context.Context, resourceGroupName string, circuitName string) (result ExpressRouteCircuit, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, circuitName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client ExpressRouteCircuitsClient) GetPreparer(ctx context.Context, resourceGroupName string, circuitName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "circuitName": autorest.Encode("path", circuitName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitsClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitsClient) GetResponder(resp *http.Response) (result ExpressRouteCircuit, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all the express route circuits in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +func (client ExpressRouteCircuitsClient) List(ctx context.Context, resourceGroupName string) (result ExpressRouteCircuitListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.erclr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "List", resp, "Failure sending request") + return + } + + result.erclr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client ExpressRouteCircuitsClient) ListPreparer(ctx context.Context, resourceGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitsClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitsClient) ListResponder(resp *http.Response) (result ExpressRouteCircuitListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client ExpressRouteCircuitsClient) listNextResults(lastResults ExpressRouteCircuitListResult) (result ExpressRouteCircuitListResult, err error) { + req, err := lastResults.expressRouteCircuitListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client ExpressRouteCircuitsClient) ListComplete(ctx context.Context, resourceGroupName string) (result ExpressRouteCircuitListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName) + return +} + +// ListAll gets all the express route circuits in a subscription. +func (client ExpressRouteCircuitsClient) ListAll(ctx context.Context) (result ExpressRouteCircuitListResultPage, err error) { + result.fn = client.listAllNextResults + req, err := client.ListAllPreparer(ctx) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListAll", nil, "Failure preparing request") + return + } + + resp, err := client.ListAllSender(req) + if err != nil { + result.erclr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListAll", resp, "Failure sending request") + return + } + + result.erclr, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListAll", resp, "Failure responding to request") + } + + return +} + +// ListAllPreparer prepares the ListAll request. +func (client ExpressRouteCircuitsClient) ListAllPreparer(ctx context.Context) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/Microsoft.Network/expressRouteCircuits", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListAllSender sends the ListAll request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitsClient) ListAllSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListAllResponder handles the response to the ListAll request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitsClient) ListAllResponder(resp *http.Response) (result ExpressRouteCircuitListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listAllNextResults retrieves the next set of results, if any. +func (client ExpressRouteCircuitsClient) listAllNextResults(lastResults ExpressRouteCircuitListResult) (result ExpressRouteCircuitListResult, err error) { + req, err := lastResults.expressRouteCircuitListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listAllNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListAllSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listAllNextResults", resp, "Failure sending next results request") + } + result, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listAllNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListAllComplete enumerates all values, automatically crossing page boundaries as required. +func (client ExpressRouteCircuitsClient) ListAllComplete(ctx context.Context) (result ExpressRouteCircuitListResultIterator, err error) { + result.page, err = client.ListAll(ctx) + return +} + +// ListArpTable the ListArpTable from ExpressRouteCircuit opertion retrieves the currently advertised arp table +// associated with the ExpressRouteCircuits in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the circuit. +func (client ExpressRouteCircuitsClient) ListArpTable(ctx context.Context, resourceGroupName string, circuitName string) (result ExpressRouteCircuitsArpTableListResultPage, err error) { + result.fn = client.listArpTableNextResults + req, err := client.ListArpTablePreparer(ctx, resourceGroupName, circuitName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListArpTable", nil, "Failure preparing request") + return + } + + resp, err := client.ListArpTableSender(req) + if err != nil { + result.ercatlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListArpTable", resp, "Failure sending request") + return + } + + result.ercatlr, err = client.ListArpTableResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListArpTable", resp, "Failure responding to request") + } + + return +} + +// ListArpTablePreparer prepares the ListArpTable request. +func (client ExpressRouteCircuitsClient) ListArpTablePreparer(ctx context.Context, resourceGroupName string, circuitName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "circuitName": autorest.Encode("path", circuitName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}/arpTable", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListArpTableSender sends the ListArpTable request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitsClient) ListArpTableSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListArpTableResponder handles the response to the ListArpTable request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitsClient) ListArpTableResponder(resp *http.Response) (result ExpressRouteCircuitsArpTableListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listArpTableNextResults retrieves the next set of results, if any. +func (client ExpressRouteCircuitsClient) listArpTableNextResults(lastResults ExpressRouteCircuitsArpTableListResult) (result ExpressRouteCircuitsArpTableListResult, err error) { + req, err := lastResults.expressRouteCircuitsArpTableListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listArpTableNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListArpTableSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listArpTableNextResults", resp, "Failure sending next results request") + } + result, err = client.ListArpTableResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listArpTableNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListArpTableComplete enumerates all values, automatically crossing page boundaries as required. +func (client ExpressRouteCircuitsClient) ListArpTableComplete(ctx context.Context, resourceGroupName string, circuitName string) (result ExpressRouteCircuitsArpTableListResultIterator, err error) { + result.page, err = client.ListArpTable(ctx, resourceGroupName, circuitName) + return +} + +// ListRoutesTable the ListRoutesTable from ExpressRouteCircuit opertion retrieves the currently advertised routes +// table associated with the ExpressRouteCircuits in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the circuit. +func (client ExpressRouteCircuitsClient) ListRoutesTable(ctx context.Context, resourceGroupName string, circuitName string) (result ExpressRouteCircuitsRoutesTableListResultPage, err error) { + result.fn = client.listRoutesTableNextResults + req, err := client.ListRoutesTablePreparer(ctx, resourceGroupName, circuitName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListRoutesTable", nil, "Failure preparing request") + return + } + + resp, err := client.ListRoutesTableSender(req) + if err != nil { + result.ercrtlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListRoutesTable", resp, "Failure sending request") + return + } + + result.ercrtlr, err = client.ListRoutesTableResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListRoutesTable", resp, "Failure responding to request") + } + + return +} + +// ListRoutesTablePreparer prepares the ListRoutesTable request. +func (client ExpressRouteCircuitsClient) ListRoutesTablePreparer(ctx context.Context, resourceGroupName string, circuitName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "circuitName": autorest.Encode("path", circuitName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}/routesTable", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListRoutesTableSender sends the ListRoutesTable request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitsClient) ListRoutesTableSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListRoutesTableResponder handles the response to the ListRoutesTable request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitsClient) ListRoutesTableResponder(resp *http.Response) (result ExpressRouteCircuitsRoutesTableListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listRoutesTableNextResults retrieves the next set of results, if any. +func (client ExpressRouteCircuitsClient) listRoutesTableNextResults(lastResults ExpressRouteCircuitsRoutesTableListResult) (result ExpressRouteCircuitsRoutesTableListResult, err error) { + req, err := lastResults.expressRouteCircuitsRoutesTableListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listRoutesTableNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListRoutesTableSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listRoutesTableNextResults", resp, "Failure sending next results request") + } + result, err = client.ListRoutesTableResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listRoutesTableNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListRoutesTableComplete enumerates all values, automatically crossing page boundaries as required. +func (client ExpressRouteCircuitsClient) ListRoutesTableComplete(ctx context.Context, resourceGroupName string, circuitName string) (result ExpressRouteCircuitsRoutesTableListResultIterator, err error) { + result.page, err = client.ListRoutesTable(ctx, resourceGroupName, circuitName) + return +} + +// ListStats the Liststats ExpressRouteCircuit opertion retrieves all the stats from a ExpressRouteCircuits in a +// resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// circuitName - the name of the loadBalancer. +func (client ExpressRouteCircuitsClient) ListStats(ctx context.Context, resourceGroupName string, circuitName string) (result ExpressRouteCircuitsStatsListResultPage, err error) { + result.fn = client.listStatsNextResults + req, err := client.ListStatsPreparer(ctx, resourceGroupName, circuitName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListStats", nil, "Failure preparing request") + return + } + + resp, err := client.ListStatsSender(req) + if err != nil { + result.ercslr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListStats", resp, "Failure sending request") + return + } + + result.ercslr, err = client.ListStatsResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "ListStats", resp, "Failure responding to request") + } + + return +} + +// ListStatsPreparer prepares the ListStats request. +func (client ExpressRouteCircuitsClient) ListStatsPreparer(ctx context.Context, resourceGroupName string, circuitName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "circuitName": autorest.Encode("path", circuitName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/expressRouteCircuits/{circuitName}/stats", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListStatsSender sends the ListStats request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteCircuitsClient) ListStatsSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListStatsResponder handles the response to the ListStats request. The method always +// closes the http.Response Body. +func (client ExpressRouteCircuitsClient) ListStatsResponder(resp *http.Response) (result ExpressRouteCircuitsStatsListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listStatsNextResults retrieves the next set of results, if any. +func (client ExpressRouteCircuitsClient) listStatsNextResults(lastResults ExpressRouteCircuitsStatsListResult) (result ExpressRouteCircuitsStatsListResult, err error) { + req, err := lastResults.expressRouteCircuitsStatsListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listStatsNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListStatsSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listStatsNextResults", resp, "Failure sending next results request") + } + result, err = client.ListStatsResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsClient", "listStatsNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListStatsComplete enumerates all values, automatically crossing page boundaries as required. +func (client ExpressRouteCircuitsClient) ListStatsComplete(ctx context.Context, resourceGroupName string, circuitName string) (result ExpressRouteCircuitsStatsListResultIterator, err error) { + result.page, err = client.ListStats(ctx, resourceGroupName, circuitName) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressrouteserviceproviders.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressrouteserviceproviders.go new file mode 100644 index 000000000..612452ba3 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/expressrouteserviceproviders.go @@ -0,0 +1,131 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// ExpressRouteServiceProvidersClient is the network Client +type ExpressRouteServiceProvidersClient struct { + BaseClient +} + +// NewExpressRouteServiceProvidersClient creates an instance of the ExpressRouteServiceProvidersClient client. +func NewExpressRouteServiceProvidersClient(subscriptionID string) ExpressRouteServiceProvidersClient { + return NewExpressRouteServiceProvidersClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewExpressRouteServiceProvidersClientWithBaseURI creates an instance of the ExpressRouteServiceProvidersClient +// client. +func NewExpressRouteServiceProvidersClientWithBaseURI(baseURI string, subscriptionID string) ExpressRouteServiceProvidersClient { + return ExpressRouteServiceProvidersClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// List gets all the available express route service providers. +func (client ExpressRouteServiceProvidersClient) List(ctx context.Context) (result ExpressRouteServiceProviderListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteServiceProvidersClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.ersplr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.ExpressRouteServiceProvidersClient", "List", resp, "Failure sending request") + return + } + + result.ersplr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteServiceProvidersClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client ExpressRouteServiceProvidersClient) ListPreparer(ctx context.Context) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/Microsoft.Network/expressRouteServiceProviders", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client ExpressRouteServiceProvidersClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client ExpressRouteServiceProvidersClient) ListResponder(resp *http.Response) (result ExpressRouteServiceProviderListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client ExpressRouteServiceProvidersClient) listNextResults(lastResults ExpressRouteServiceProviderListResult) (result ExpressRouteServiceProviderListResult, err error) { + req, err := lastResults.expressRouteServiceProviderListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.ExpressRouteServiceProvidersClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.ExpressRouteServiceProvidersClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteServiceProvidersClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client ExpressRouteServiceProvidersClient) ListComplete(ctx context.Context) (result ExpressRouteServiceProviderListResultIterator, err error) { + result.page, err = client.List(ctx) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/interfaces.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/interfaces.go new file mode 100644 index 000000000..f989d7bc7 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/interfaces.go @@ -0,0 +1,702 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// InterfacesClient is the network Client +type InterfacesClient struct { + BaseClient +} + +// NewInterfacesClient creates an instance of the InterfacesClient client. +func NewInterfacesClient(subscriptionID string) InterfacesClient { + return NewInterfacesClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewInterfacesClientWithBaseURI creates an instance of the InterfacesClient client. +func NewInterfacesClientWithBaseURI(baseURI string, subscriptionID string) InterfacesClient { + return InterfacesClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a network interface. +// Parameters: +// resourceGroupName - the name of the resource group. +// networkInterfaceName - the name of the network interface. +// parameters - parameters supplied to the create or update network interface operation. +func (client InterfacesClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, networkInterfaceName string, parameters Interface) (result InterfacesCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, networkInterfaceName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client InterfacesClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, networkInterfaceName string, parameters Interface) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "networkInterfaceName": autorest.Encode("path", networkInterfaceName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkInterfaces/{networkInterfaceName}", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client InterfacesClient) CreateOrUpdateSender(req *http.Request) (future InterfacesCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client InterfacesClient) CreateOrUpdateResponder(resp *http.Response) (result Interface, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified network interface. +// Parameters: +// resourceGroupName - the name of the resource group. +// networkInterfaceName - the name of the network interface. +func (client InterfacesClient) Delete(ctx context.Context, resourceGroupName string, networkInterfaceName string) (result InterfacesDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, networkInterfaceName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client InterfacesClient) DeletePreparer(ctx context.Context, resourceGroupName string, networkInterfaceName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "networkInterfaceName": autorest.Encode("path", networkInterfaceName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkInterfaces/{networkInterfaceName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client InterfacesClient) DeleteSender(req *http.Request) (future InterfacesDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client InterfacesClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets information about the specified network interface. +// Parameters: +// resourceGroupName - the name of the resource group. +// networkInterfaceName - the name of the network interface. +// expand - expands referenced resources. +func (client InterfacesClient) Get(ctx context.Context, resourceGroupName string, networkInterfaceName string, expand string) (result Interface, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, networkInterfaceName, expand) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client InterfacesClient) GetPreparer(ctx context.Context, resourceGroupName string, networkInterfaceName string, expand string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "networkInterfaceName": autorest.Encode("path", networkInterfaceName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkInterfaces/{networkInterfaceName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client InterfacesClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client InterfacesClient) GetResponder(resp *http.Response) (result Interface, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// GetVirtualMachineScaleSetNetworkInterface get the specified network interface in a virtual machine scale set. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualMachineScaleSetName - the name of the virtual machine scale set. +// virtualmachineIndex - the virtual machine index. +// networkInterfaceName - the name of the network interface. +// expand - expands referenced resources. +func (client InterfacesClient) GetVirtualMachineScaleSetNetworkInterface(ctx context.Context, resourceGroupName string, virtualMachineScaleSetName string, virtualmachineIndex string, networkInterfaceName string, expand string) (result Interface, err error) { + req, err := client.GetVirtualMachineScaleSetNetworkInterfacePreparer(ctx, resourceGroupName, virtualMachineScaleSetName, virtualmachineIndex, networkInterfaceName, expand) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "GetVirtualMachineScaleSetNetworkInterface", nil, "Failure preparing request") + return + } + + resp, err := client.GetVirtualMachineScaleSetNetworkInterfaceSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "GetVirtualMachineScaleSetNetworkInterface", resp, "Failure sending request") + return + } + + result, err = client.GetVirtualMachineScaleSetNetworkInterfaceResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "GetVirtualMachineScaleSetNetworkInterface", resp, "Failure responding to request") + } + + return +} + +// GetVirtualMachineScaleSetNetworkInterfacePreparer prepares the GetVirtualMachineScaleSetNetworkInterface request. +func (client InterfacesClient) GetVirtualMachineScaleSetNetworkInterfacePreparer(ctx context.Context, resourceGroupName string, virtualMachineScaleSetName string, virtualmachineIndex string, networkInterfaceName string, expand string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "networkInterfaceName": autorest.Encode("path", networkInterfaceName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualmachineIndex": autorest.Encode("path", virtualmachineIndex), + "virtualMachineScaleSetName": autorest.Encode("path", virtualMachineScaleSetName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/microsoft.Compute/virtualMachineScaleSets/{virtualMachineScaleSetName}/virtualMachines/{virtualmachineIndex}/networkInterfaces/{networkInterfaceName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetVirtualMachineScaleSetNetworkInterfaceSender sends the GetVirtualMachineScaleSetNetworkInterface request. The method will close the +// http.Response Body if it receives an error. +func (client InterfacesClient) GetVirtualMachineScaleSetNetworkInterfaceSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetVirtualMachineScaleSetNetworkInterfaceResponder handles the response to the GetVirtualMachineScaleSetNetworkInterface request. The method always +// closes the http.Response Body. +func (client InterfacesClient) GetVirtualMachineScaleSetNetworkInterfaceResponder(resp *http.Response) (result Interface, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all network interfaces in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +func (client InterfacesClient) List(ctx context.Context, resourceGroupName string) (result InterfaceListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.ilr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "List", resp, "Failure sending request") + return + } + + result.ilr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client InterfacesClient) ListPreparer(ctx context.Context, resourceGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkInterfaces", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client InterfacesClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client InterfacesClient) ListResponder(resp *http.Response) (result InterfaceListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client InterfacesClient) listNextResults(lastResults InterfaceListResult) (result InterfaceListResult, err error) { + req, err := lastResults.interfaceListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.InterfacesClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.InterfacesClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client InterfacesClient) ListComplete(ctx context.Context, resourceGroupName string) (result InterfaceListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName) + return +} + +// ListAll gets all network interfaces in a subscription. +func (client InterfacesClient) ListAll(ctx context.Context) (result InterfaceListResultPage, err error) { + result.fn = client.listAllNextResults + req, err := client.ListAllPreparer(ctx) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "ListAll", nil, "Failure preparing request") + return + } + + resp, err := client.ListAllSender(req) + if err != nil { + result.ilr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "ListAll", resp, "Failure sending request") + return + } + + result.ilr, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "ListAll", resp, "Failure responding to request") + } + + return +} + +// ListAllPreparer prepares the ListAll request. +func (client InterfacesClient) ListAllPreparer(ctx context.Context) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/Microsoft.Network/networkInterfaces", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListAllSender sends the ListAll request. The method will close the +// http.Response Body if it receives an error. +func (client InterfacesClient) ListAllSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListAllResponder handles the response to the ListAll request. The method always +// closes the http.Response Body. +func (client InterfacesClient) ListAllResponder(resp *http.Response) (result InterfaceListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listAllNextResults retrieves the next set of results, if any. +func (client InterfacesClient) listAllNextResults(lastResults InterfaceListResult) (result InterfaceListResult, err error) { + req, err := lastResults.interfaceListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.InterfacesClient", "listAllNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListAllSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.InterfacesClient", "listAllNextResults", resp, "Failure sending next results request") + } + result, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "listAllNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListAllComplete enumerates all values, automatically crossing page boundaries as required. +func (client InterfacesClient) ListAllComplete(ctx context.Context) (result InterfaceListResultIterator, err error) { + result.page, err = client.ListAll(ctx) + return +} + +// ListVirtualMachineScaleSetNetworkInterfaces gets all network interfaces in a virtual machine scale set. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualMachineScaleSetName - the name of the virtual machine scale set. +func (client InterfacesClient) ListVirtualMachineScaleSetNetworkInterfaces(ctx context.Context, resourceGroupName string, virtualMachineScaleSetName string) (result InterfaceListResultPage, err error) { + result.fn = client.listVirtualMachineScaleSetNetworkInterfacesNextResults + req, err := client.ListVirtualMachineScaleSetNetworkInterfacesPreparer(ctx, resourceGroupName, virtualMachineScaleSetName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "ListVirtualMachineScaleSetNetworkInterfaces", nil, "Failure preparing request") + return + } + + resp, err := client.ListVirtualMachineScaleSetNetworkInterfacesSender(req) + if err != nil { + result.ilr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "ListVirtualMachineScaleSetNetworkInterfaces", resp, "Failure sending request") + return + } + + result.ilr, err = client.ListVirtualMachineScaleSetNetworkInterfacesResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "ListVirtualMachineScaleSetNetworkInterfaces", resp, "Failure responding to request") + } + + return +} + +// ListVirtualMachineScaleSetNetworkInterfacesPreparer prepares the ListVirtualMachineScaleSetNetworkInterfaces request. +func (client InterfacesClient) ListVirtualMachineScaleSetNetworkInterfacesPreparer(ctx context.Context, resourceGroupName string, virtualMachineScaleSetName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualMachineScaleSetName": autorest.Encode("path", virtualMachineScaleSetName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/microsoft.Compute/virtualMachineScaleSets/{virtualMachineScaleSetName}/networkInterfaces", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListVirtualMachineScaleSetNetworkInterfacesSender sends the ListVirtualMachineScaleSetNetworkInterfaces request. The method will close the +// http.Response Body if it receives an error. +func (client InterfacesClient) ListVirtualMachineScaleSetNetworkInterfacesSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListVirtualMachineScaleSetNetworkInterfacesResponder handles the response to the ListVirtualMachineScaleSetNetworkInterfaces request. The method always +// closes the http.Response Body. +func (client InterfacesClient) ListVirtualMachineScaleSetNetworkInterfacesResponder(resp *http.Response) (result InterfaceListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listVirtualMachineScaleSetNetworkInterfacesNextResults retrieves the next set of results, if any. +func (client InterfacesClient) listVirtualMachineScaleSetNetworkInterfacesNextResults(lastResults InterfaceListResult) (result InterfaceListResult, err error) { + req, err := lastResults.interfaceListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.InterfacesClient", "listVirtualMachineScaleSetNetworkInterfacesNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListVirtualMachineScaleSetNetworkInterfacesSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.InterfacesClient", "listVirtualMachineScaleSetNetworkInterfacesNextResults", resp, "Failure sending next results request") + } + result, err = client.ListVirtualMachineScaleSetNetworkInterfacesResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "listVirtualMachineScaleSetNetworkInterfacesNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListVirtualMachineScaleSetNetworkInterfacesComplete enumerates all values, automatically crossing page boundaries as required. +func (client InterfacesClient) ListVirtualMachineScaleSetNetworkInterfacesComplete(ctx context.Context, resourceGroupName string, virtualMachineScaleSetName string) (result InterfaceListResultIterator, err error) { + result.page, err = client.ListVirtualMachineScaleSetNetworkInterfaces(ctx, resourceGroupName, virtualMachineScaleSetName) + return +} + +// ListVirtualMachineScaleSetVMNetworkInterfaces gets information about all network interfaces in a virtual machine in +// a virtual machine scale set. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualMachineScaleSetName - the name of the virtual machine scale set. +// virtualmachineIndex - the virtual machine index. +func (client InterfacesClient) ListVirtualMachineScaleSetVMNetworkInterfaces(ctx context.Context, resourceGroupName string, virtualMachineScaleSetName string, virtualmachineIndex string) (result InterfaceListResultPage, err error) { + result.fn = client.listVirtualMachineScaleSetVMNetworkInterfacesNextResults + req, err := client.ListVirtualMachineScaleSetVMNetworkInterfacesPreparer(ctx, resourceGroupName, virtualMachineScaleSetName, virtualmachineIndex) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "ListVirtualMachineScaleSetVMNetworkInterfaces", nil, "Failure preparing request") + return + } + + resp, err := client.ListVirtualMachineScaleSetVMNetworkInterfacesSender(req) + if err != nil { + result.ilr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "ListVirtualMachineScaleSetVMNetworkInterfaces", resp, "Failure sending request") + return + } + + result.ilr, err = client.ListVirtualMachineScaleSetVMNetworkInterfacesResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "ListVirtualMachineScaleSetVMNetworkInterfaces", resp, "Failure responding to request") + } + + return +} + +// ListVirtualMachineScaleSetVMNetworkInterfacesPreparer prepares the ListVirtualMachineScaleSetVMNetworkInterfaces request. +func (client InterfacesClient) ListVirtualMachineScaleSetVMNetworkInterfacesPreparer(ctx context.Context, resourceGroupName string, virtualMachineScaleSetName string, virtualmachineIndex string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualmachineIndex": autorest.Encode("path", virtualmachineIndex), + "virtualMachineScaleSetName": autorest.Encode("path", virtualMachineScaleSetName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/microsoft.Compute/virtualMachineScaleSets/{virtualMachineScaleSetName}/virtualMachines/{virtualmachineIndex}/networkInterfaces", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListVirtualMachineScaleSetVMNetworkInterfacesSender sends the ListVirtualMachineScaleSetVMNetworkInterfaces request. The method will close the +// http.Response Body if it receives an error. +func (client InterfacesClient) ListVirtualMachineScaleSetVMNetworkInterfacesSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListVirtualMachineScaleSetVMNetworkInterfacesResponder handles the response to the ListVirtualMachineScaleSetVMNetworkInterfaces request. The method always +// closes the http.Response Body. +func (client InterfacesClient) ListVirtualMachineScaleSetVMNetworkInterfacesResponder(resp *http.Response) (result InterfaceListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listVirtualMachineScaleSetVMNetworkInterfacesNextResults retrieves the next set of results, if any. +func (client InterfacesClient) listVirtualMachineScaleSetVMNetworkInterfacesNextResults(lastResults InterfaceListResult) (result InterfaceListResult, err error) { + req, err := lastResults.interfaceListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.InterfacesClient", "listVirtualMachineScaleSetVMNetworkInterfacesNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListVirtualMachineScaleSetVMNetworkInterfacesSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.InterfacesClient", "listVirtualMachineScaleSetVMNetworkInterfacesNextResults", resp, "Failure sending next results request") + } + result, err = client.ListVirtualMachineScaleSetVMNetworkInterfacesResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesClient", "listVirtualMachineScaleSetVMNetworkInterfacesNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListVirtualMachineScaleSetVMNetworkInterfacesComplete enumerates all values, automatically crossing page boundaries as required. +func (client InterfacesClient) ListVirtualMachineScaleSetVMNetworkInterfacesComplete(ctx context.Context, resourceGroupName string, virtualMachineScaleSetName string, virtualmachineIndex string) (result InterfaceListResultIterator, err error) { + result.page, err = client.ListVirtualMachineScaleSetVMNetworkInterfaces(ctx, resourceGroupName, virtualMachineScaleSetName, virtualmachineIndex) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/loadbalancers.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/loadbalancers.go new file mode 100644 index 000000000..c2f0b9768 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/loadbalancers.go @@ -0,0 +1,434 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// LoadBalancersClient is the network Client +type LoadBalancersClient struct { + BaseClient +} + +// NewLoadBalancersClient creates an instance of the LoadBalancersClient client. +func NewLoadBalancersClient(subscriptionID string) LoadBalancersClient { + return NewLoadBalancersClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewLoadBalancersClientWithBaseURI creates an instance of the LoadBalancersClient client. +func NewLoadBalancersClientWithBaseURI(baseURI string, subscriptionID string) LoadBalancersClient { + return LoadBalancersClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a load balancer. +// Parameters: +// resourceGroupName - the name of the resource group. +// loadBalancerName - the name of the load balancer. +// parameters - parameters supplied to the create or update load balancer operation. +func (client LoadBalancersClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, loadBalancerName string, parameters LoadBalancer) (result LoadBalancersCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, loadBalancerName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client LoadBalancersClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, loadBalancerName string, parameters LoadBalancer) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "loadBalancerName": autorest.Encode("path", loadBalancerName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/loadBalancers/{loadBalancerName}", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client LoadBalancersClient) CreateOrUpdateSender(req *http.Request) (future LoadBalancersCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client LoadBalancersClient) CreateOrUpdateResponder(resp *http.Response) (result LoadBalancer, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified load balancer. +// Parameters: +// resourceGroupName - the name of the resource group. +// loadBalancerName - the name of the load balancer. +func (client LoadBalancersClient) Delete(ctx context.Context, resourceGroupName string, loadBalancerName string) (result LoadBalancersDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, loadBalancerName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client LoadBalancersClient) DeletePreparer(ctx context.Context, resourceGroupName string, loadBalancerName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "loadBalancerName": autorest.Encode("path", loadBalancerName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/loadBalancers/{loadBalancerName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client LoadBalancersClient) DeleteSender(req *http.Request) (future LoadBalancersDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client LoadBalancersClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified load balancer. +// Parameters: +// resourceGroupName - the name of the resource group. +// loadBalancerName - the name of the load balancer. +// expand - expands referenced resources. +func (client LoadBalancersClient) Get(ctx context.Context, resourceGroupName string, loadBalancerName string, expand string) (result LoadBalancer, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, loadBalancerName, expand) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client LoadBalancersClient) GetPreparer(ctx context.Context, resourceGroupName string, loadBalancerName string, expand string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "loadBalancerName": autorest.Encode("path", loadBalancerName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/loadBalancers/{loadBalancerName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client LoadBalancersClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client LoadBalancersClient) GetResponder(resp *http.Response) (result LoadBalancer, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all the load balancers in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +func (client LoadBalancersClient) List(ctx context.Context, resourceGroupName string) (result LoadBalancerListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.lblr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "List", resp, "Failure sending request") + return + } + + result.lblr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client LoadBalancersClient) ListPreparer(ctx context.Context, resourceGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/loadBalancers", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client LoadBalancersClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client LoadBalancersClient) ListResponder(resp *http.Response) (result LoadBalancerListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client LoadBalancersClient) listNextResults(lastResults LoadBalancerListResult) (result LoadBalancerListResult, err error) { + req, err := lastResults.loadBalancerListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.LoadBalancersClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.LoadBalancersClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client LoadBalancersClient) ListComplete(ctx context.Context, resourceGroupName string) (result LoadBalancerListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName) + return +} + +// ListAll gets all the load balancers in a subscription. +func (client LoadBalancersClient) ListAll(ctx context.Context) (result LoadBalancerListResultPage, err error) { + result.fn = client.listAllNextResults + req, err := client.ListAllPreparer(ctx) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "ListAll", nil, "Failure preparing request") + return + } + + resp, err := client.ListAllSender(req) + if err != nil { + result.lblr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "ListAll", resp, "Failure sending request") + return + } + + result.lblr, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "ListAll", resp, "Failure responding to request") + } + + return +} + +// ListAllPreparer prepares the ListAll request. +func (client LoadBalancersClient) ListAllPreparer(ctx context.Context) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/Microsoft.Network/loadBalancers", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListAllSender sends the ListAll request. The method will close the +// http.Response Body if it receives an error. +func (client LoadBalancersClient) ListAllSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListAllResponder handles the response to the ListAll request. The method always +// closes the http.Response Body. +func (client LoadBalancersClient) ListAllResponder(resp *http.Response) (result LoadBalancerListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listAllNextResults retrieves the next set of results, if any. +func (client LoadBalancersClient) listAllNextResults(lastResults LoadBalancerListResult) (result LoadBalancerListResult, err error) { + req, err := lastResults.loadBalancerListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.LoadBalancersClient", "listAllNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListAllSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.LoadBalancersClient", "listAllNextResults", resp, "Failure sending next results request") + } + result, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersClient", "listAllNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListAllComplete enumerates all values, automatically crossing page boundaries as required. +func (client LoadBalancersClient) ListAllComplete(ctx context.Context) (result LoadBalancerListResultIterator, err error) { + result.page, err = client.ListAll(ctx) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/localnetworkgateways.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/localnetworkgateways.go new file mode 100644 index 000000000..4e7d86fcc --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/localnetworkgateways.go @@ -0,0 +1,340 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// LocalNetworkGatewaysClient is the network Client +type LocalNetworkGatewaysClient struct { + BaseClient +} + +// NewLocalNetworkGatewaysClient creates an instance of the LocalNetworkGatewaysClient client. +func NewLocalNetworkGatewaysClient(subscriptionID string) LocalNetworkGatewaysClient { + return NewLocalNetworkGatewaysClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewLocalNetworkGatewaysClientWithBaseURI creates an instance of the LocalNetworkGatewaysClient client. +func NewLocalNetworkGatewaysClientWithBaseURI(baseURI string, subscriptionID string) LocalNetworkGatewaysClient { + return LocalNetworkGatewaysClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a local network gateway in the specified resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// localNetworkGatewayName - the name of the local network gateway. +// parameters - parameters supplied to the create or update local network gateway operation. +func (client LocalNetworkGatewaysClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, localNetworkGatewayName string, parameters LocalNetworkGateway) (result LocalNetworkGatewaysCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, localNetworkGatewayName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client LocalNetworkGatewaysClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, localNetworkGatewayName string, parameters LocalNetworkGateway) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "localNetworkGatewayName": autorest.Encode("path", localNetworkGatewayName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/localNetworkGateways/{localNetworkGatewayName}", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client LocalNetworkGatewaysClient) CreateOrUpdateSender(req *http.Request) (future LocalNetworkGatewaysCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client LocalNetworkGatewaysClient) CreateOrUpdateResponder(resp *http.Response) (result LocalNetworkGateway, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified local network gateway. +// Parameters: +// resourceGroupName - the name of the resource group. +// localNetworkGatewayName - the name of the local network gateway. +func (client LocalNetworkGatewaysClient) Delete(ctx context.Context, resourceGroupName string, localNetworkGatewayName string) (result LocalNetworkGatewaysDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, localNetworkGatewayName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client LocalNetworkGatewaysClient) DeletePreparer(ctx context.Context, resourceGroupName string, localNetworkGatewayName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "localNetworkGatewayName": autorest.Encode("path", localNetworkGatewayName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/localNetworkGateways/{localNetworkGatewayName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client LocalNetworkGatewaysClient) DeleteSender(req *http.Request) (future LocalNetworkGatewaysDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client LocalNetworkGatewaysClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified local network gateway in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// localNetworkGatewayName - the name of the local network gateway. +func (client LocalNetworkGatewaysClient) Get(ctx context.Context, resourceGroupName string, localNetworkGatewayName string) (result LocalNetworkGateway, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, localNetworkGatewayName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client LocalNetworkGatewaysClient) GetPreparer(ctx context.Context, resourceGroupName string, localNetworkGatewayName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "localNetworkGatewayName": autorest.Encode("path", localNetworkGatewayName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/localNetworkGateways/{localNetworkGatewayName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client LocalNetworkGatewaysClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client LocalNetworkGatewaysClient) GetResponder(resp *http.Response) (result LocalNetworkGateway, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all the local network gateways in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +func (client LocalNetworkGatewaysClient) List(ctx context.Context, resourceGroupName string) (result LocalNetworkGatewayListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.lnglr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "List", resp, "Failure sending request") + return + } + + result.lnglr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client LocalNetworkGatewaysClient) ListPreparer(ctx context.Context, resourceGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/localNetworkGateways", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client LocalNetworkGatewaysClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client LocalNetworkGatewaysClient) ListResponder(resp *http.Response) (result LocalNetworkGatewayListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client LocalNetworkGatewaysClient) listNextResults(lastResults LocalNetworkGatewayListResult) (result LocalNetworkGatewayListResult, err error) { + req, err := lastResults.localNetworkGatewayListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client LocalNetworkGatewaysClient) ListComplete(ctx context.Context, resourceGroupName string) (result LocalNetworkGatewayListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/models.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/models.go new file mode 100644 index 000000000..4efbedb85 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/models.go @@ -0,0 +1,9113 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "encoding/json" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/to" + "net/http" +) + +// ApplicationGatewayCookieBasedAffinity enumerates the values for application gateway cookie based affinity. +type ApplicationGatewayCookieBasedAffinity string + +const ( + // Disabled ... + Disabled ApplicationGatewayCookieBasedAffinity = "Disabled" + // Enabled ... + Enabled ApplicationGatewayCookieBasedAffinity = "Enabled" +) + +// PossibleApplicationGatewayCookieBasedAffinityValues returns an array of possible values for the ApplicationGatewayCookieBasedAffinity const type. +func PossibleApplicationGatewayCookieBasedAffinityValues() []ApplicationGatewayCookieBasedAffinity { + return []ApplicationGatewayCookieBasedAffinity{Disabled, Enabled} +} + +// ApplicationGatewayOperationalState enumerates the values for application gateway operational state. +type ApplicationGatewayOperationalState string + +const ( + // Running ... + Running ApplicationGatewayOperationalState = "Running" + // Starting ... + Starting ApplicationGatewayOperationalState = "Starting" + // Stopped ... + Stopped ApplicationGatewayOperationalState = "Stopped" + // Stopping ... + Stopping ApplicationGatewayOperationalState = "Stopping" +) + +// PossibleApplicationGatewayOperationalStateValues returns an array of possible values for the ApplicationGatewayOperationalState const type. +func PossibleApplicationGatewayOperationalStateValues() []ApplicationGatewayOperationalState { + return []ApplicationGatewayOperationalState{Running, Starting, Stopped, Stopping} +} + +// ApplicationGatewayProtocol enumerates the values for application gateway protocol. +type ApplicationGatewayProtocol string + +const ( + // HTTP ... + HTTP ApplicationGatewayProtocol = "Http" + // HTTPS ... + HTTPS ApplicationGatewayProtocol = "Https" +) + +// PossibleApplicationGatewayProtocolValues returns an array of possible values for the ApplicationGatewayProtocol const type. +func PossibleApplicationGatewayProtocolValues() []ApplicationGatewayProtocol { + return []ApplicationGatewayProtocol{HTTP, HTTPS} +} + +// ApplicationGatewayRequestRoutingRuleType enumerates the values for application gateway request routing rule +// type. +type ApplicationGatewayRequestRoutingRuleType string + +const ( + // Basic ... + Basic ApplicationGatewayRequestRoutingRuleType = "Basic" + // PathBasedRouting ... + PathBasedRouting ApplicationGatewayRequestRoutingRuleType = "PathBasedRouting" +) + +// PossibleApplicationGatewayRequestRoutingRuleTypeValues returns an array of possible values for the ApplicationGatewayRequestRoutingRuleType const type. +func PossibleApplicationGatewayRequestRoutingRuleTypeValues() []ApplicationGatewayRequestRoutingRuleType { + return []ApplicationGatewayRequestRoutingRuleType{Basic, PathBasedRouting} +} + +// ApplicationGatewaySkuName enumerates the values for application gateway sku name. +type ApplicationGatewaySkuName string + +const ( + // StandardLarge ... + StandardLarge ApplicationGatewaySkuName = "Standard_Large" + // StandardMedium ... + StandardMedium ApplicationGatewaySkuName = "Standard_Medium" + // StandardSmall ... + StandardSmall ApplicationGatewaySkuName = "Standard_Small" +) + +// PossibleApplicationGatewaySkuNameValues returns an array of possible values for the ApplicationGatewaySkuName const type. +func PossibleApplicationGatewaySkuNameValues() []ApplicationGatewaySkuName { + return []ApplicationGatewaySkuName{StandardLarge, StandardMedium, StandardSmall} +} + +// ApplicationGatewayTier enumerates the values for application gateway tier. +type ApplicationGatewayTier string + +const ( + // Standard ... + Standard ApplicationGatewayTier = "Standard" +) + +// PossibleApplicationGatewayTierValues returns an array of possible values for the ApplicationGatewayTier const type. +func PossibleApplicationGatewayTierValues() []ApplicationGatewayTier { + return []ApplicationGatewayTier{Standard} +} + +// AuthorizationUseStatus enumerates the values for authorization use status. +type AuthorizationUseStatus string + +const ( + // Available ... + Available AuthorizationUseStatus = "Available" + // InUse ... + InUse AuthorizationUseStatus = "InUse" +) + +// PossibleAuthorizationUseStatusValues returns an array of possible values for the AuthorizationUseStatus const type. +func PossibleAuthorizationUseStatusValues() []AuthorizationUseStatus { + return []AuthorizationUseStatus{Available, InUse} +} + +// ExpressRouteCircuitPeeringAdvertisedPublicPrefixState enumerates the values for express route circuit +// peering advertised public prefix state. +type ExpressRouteCircuitPeeringAdvertisedPublicPrefixState string + +const ( + // Configured ... + Configured ExpressRouteCircuitPeeringAdvertisedPublicPrefixState = "Configured" + // Configuring ... + Configuring ExpressRouteCircuitPeeringAdvertisedPublicPrefixState = "Configuring" + // NotConfigured ... + NotConfigured ExpressRouteCircuitPeeringAdvertisedPublicPrefixState = "NotConfigured" + // ValidationNeeded ... + ValidationNeeded ExpressRouteCircuitPeeringAdvertisedPublicPrefixState = "ValidationNeeded" +) + +// PossibleExpressRouteCircuitPeeringAdvertisedPublicPrefixStateValues returns an array of possible values for the ExpressRouteCircuitPeeringAdvertisedPublicPrefixState const type. +func PossibleExpressRouteCircuitPeeringAdvertisedPublicPrefixStateValues() []ExpressRouteCircuitPeeringAdvertisedPublicPrefixState { + return []ExpressRouteCircuitPeeringAdvertisedPublicPrefixState{Configured, Configuring, NotConfigured, ValidationNeeded} +} + +// ExpressRouteCircuitPeeringState enumerates the values for express route circuit peering state. +type ExpressRouteCircuitPeeringState string + +const ( + // ExpressRouteCircuitPeeringStateDisabled ... + ExpressRouteCircuitPeeringStateDisabled ExpressRouteCircuitPeeringState = "Disabled" + // ExpressRouteCircuitPeeringStateEnabled ... + ExpressRouteCircuitPeeringStateEnabled ExpressRouteCircuitPeeringState = "Enabled" +) + +// PossibleExpressRouteCircuitPeeringStateValues returns an array of possible values for the ExpressRouteCircuitPeeringState const type. +func PossibleExpressRouteCircuitPeeringStateValues() []ExpressRouteCircuitPeeringState { + return []ExpressRouteCircuitPeeringState{ExpressRouteCircuitPeeringStateDisabled, ExpressRouteCircuitPeeringStateEnabled} +} + +// ExpressRouteCircuitPeeringType enumerates the values for express route circuit peering type. +type ExpressRouteCircuitPeeringType string + +const ( + // AzurePrivatePeering ... + AzurePrivatePeering ExpressRouteCircuitPeeringType = "AzurePrivatePeering" + // AzurePublicPeering ... + AzurePublicPeering ExpressRouteCircuitPeeringType = "AzurePublicPeering" + // MicrosoftPeering ... + MicrosoftPeering ExpressRouteCircuitPeeringType = "MicrosoftPeering" +) + +// PossibleExpressRouteCircuitPeeringTypeValues returns an array of possible values for the ExpressRouteCircuitPeeringType const type. +func PossibleExpressRouteCircuitPeeringTypeValues() []ExpressRouteCircuitPeeringType { + return []ExpressRouteCircuitPeeringType{AzurePrivatePeering, AzurePublicPeering, MicrosoftPeering} +} + +// ExpressRouteCircuitSkuFamily enumerates the values for express route circuit sku family. +type ExpressRouteCircuitSkuFamily string + +const ( + // MeteredData ... + MeteredData ExpressRouteCircuitSkuFamily = "MeteredData" + // UnlimitedData ... + UnlimitedData ExpressRouteCircuitSkuFamily = "UnlimitedData" +) + +// PossibleExpressRouteCircuitSkuFamilyValues returns an array of possible values for the ExpressRouteCircuitSkuFamily const type. +func PossibleExpressRouteCircuitSkuFamilyValues() []ExpressRouteCircuitSkuFamily { + return []ExpressRouteCircuitSkuFamily{MeteredData, UnlimitedData} +} + +// ExpressRouteCircuitSkuTier enumerates the values for express route circuit sku tier. +type ExpressRouteCircuitSkuTier string + +const ( + // ExpressRouteCircuitSkuTierPremium ... + ExpressRouteCircuitSkuTierPremium ExpressRouteCircuitSkuTier = "Premium" + // ExpressRouteCircuitSkuTierStandard ... + ExpressRouteCircuitSkuTierStandard ExpressRouteCircuitSkuTier = "Standard" +) + +// PossibleExpressRouteCircuitSkuTierValues returns an array of possible values for the ExpressRouteCircuitSkuTier const type. +func PossibleExpressRouteCircuitSkuTierValues() []ExpressRouteCircuitSkuTier { + return []ExpressRouteCircuitSkuTier{ExpressRouteCircuitSkuTierPremium, ExpressRouteCircuitSkuTierStandard} +} + +// IPAllocationMethod enumerates the values for ip allocation method. +type IPAllocationMethod string + +const ( + // Dynamic ... + Dynamic IPAllocationMethod = "Dynamic" + // Static ... + Static IPAllocationMethod = "Static" +) + +// PossibleIPAllocationMethodValues returns an array of possible values for the IPAllocationMethod const type. +func PossibleIPAllocationMethodValues() []IPAllocationMethod { + return []IPAllocationMethod{Dynamic, Static} +} + +// LoadDistribution enumerates the values for load distribution. +type LoadDistribution string + +const ( + // Default ... + Default LoadDistribution = "Default" + // SourceIP ... + SourceIP LoadDistribution = "SourceIP" + // SourceIPProtocol ... + SourceIPProtocol LoadDistribution = "SourceIPProtocol" +) + +// PossibleLoadDistributionValues returns an array of possible values for the LoadDistribution const type. +func PossibleLoadDistributionValues() []LoadDistribution { + return []LoadDistribution{Default, SourceIP, SourceIPProtocol} +} + +// OperationStatus enumerates the values for operation status. +type OperationStatus string + +const ( + // Failed ... + Failed OperationStatus = "Failed" + // InProgress ... + InProgress OperationStatus = "InProgress" + // Succeeded ... + Succeeded OperationStatus = "Succeeded" +) + +// PossibleOperationStatusValues returns an array of possible values for the OperationStatus const type. +func PossibleOperationStatusValues() []OperationStatus { + return []OperationStatus{Failed, InProgress, Succeeded} +} + +// ProbeProtocol enumerates the values for probe protocol. +type ProbeProtocol string + +const ( + // ProbeProtocolHTTP ... + ProbeProtocolHTTP ProbeProtocol = "Http" + // ProbeProtocolTCP ... + ProbeProtocolTCP ProbeProtocol = "Tcp" +) + +// PossibleProbeProtocolValues returns an array of possible values for the ProbeProtocol const type. +func PossibleProbeProtocolValues() []ProbeProtocol { + return []ProbeProtocol{ProbeProtocolHTTP, ProbeProtocolTCP} +} + +// ProcessorArchitecture enumerates the values for processor architecture. +type ProcessorArchitecture string + +const ( + // Amd64 ... + Amd64 ProcessorArchitecture = "Amd64" + // X86 ... + X86 ProcessorArchitecture = "X86" +) + +// PossibleProcessorArchitectureValues returns an array of possible values for the ProcessorArchitecture const type. +func PossibleProcessorArchitectureValues() []ProcessorArchitecture { + return []ProcessorArchitecture{Amd64, X86} +} + +// RouteNextHopType enumerates the values for route next hop type. +type RouteNextHopType string + +const ( + // RouteNextHopTypeInternet ... + RouteNextHopTypeInternet RouteNextHopType = "Internet" + // RouteNextHopTypeNone ... + RouteNextHopTypeNone RouteNextHopType = "None" + // RouteNextHopTypeVirtualAppliance ... + RouteNextHopTypeVirtualAppliance RouteNextHopType = "VirtualAppliance" + // RouteNextHopTypeVirtualNetworkGateway ... + RouteNextHopTypeVirtualNetworkGateway RouteNextHopType = "VirtualNetworkGateway" + // RouteNextHopTypeVnetLocal ... + RouteNextHopTypeVnetLocal RouteNextHopType = "VnetLocal" +) + +// PossibleRouteNextHopTypeValues returns an array of possible values for the RouteNextHopType const type. +func PossibleRouteNextHopTypeValues() []RouteNextHopType { + return []RouteNextHopType{RouteNextHopTypeInternet, RouteNextHopTypeNone, RouteNextHopTypeVirtualAppliance, RouteNextHopTypeVirtualNetworkGateway, RouteNextHopTypeVnetLocal} +} + +// SecurityRuleAccess enumerates the values for security rule access. +type SecurityRuleAccess string + +const ( + // Allow ... + Allow SecurityRuleAccess = "Allow" + // Deny ... + Deny SecurityRuleAccess = "Deny" +) + +// PossibleSecurityRuleAccessValues returns an array of possible values for the SecurityRuleAccess const type. +func PossibleSecurityRuleAccessValues() []SecurityRuleAccess { + return []SecurityRuleAccess{Allow, Deny} +} + +// SecurityRuleDirection enumerates the values for security rule direction. +type SecurityRuleDirection string + +const ( + // Inbound ... + Inbound SecurityRuleDirection = "Inbound" + // Outbound ... + Outbound SecurityRuleDirection = "Outbound" +) + +// PossibleSecurityRuleDirectionValues returns an array of possible values for the SecurityRuleDirection const type. +func PossibleSecurityRuleDirectionValues() []SecurityRuleDirection { + return []SecurityRuleDirection{Inbound, Outbound} +} + +// SecurityRuleProtocol enumerates the values for security rule protocol. +type SecurityRuleProtocol string + +const ( + // Asterisk ... + Asterisk SecurityRuleProtocol = "*" + // TCP ... + TCP SecurityRuleProtocol = "Tcp" + // UDP ... + UDP SecurityRuleProtocol = "Udp" +) + +// PossibleSecurityRuleProtocolValues returns an array of possible values for the SecurityRuleProtocol const type. +func PossibleSecurityRuleProtocolValues() []SecurityRuleProtocol { + return []SecurityRuleProtocol{Asterisk, TCP, UDP} +} + +// ServiceProviderProvisioningState enumerates the values for service provider provisioning state. +type ServiceProviderProvisioningState string + +const ( + // Deprovisioning ... + Deprovisioning ServiceProviderProvisioningState = "Deprovisioning" + // NotProvisioned ... + NotProvisioned ServiceProviderProvisioningState = "NotProvisioned" + // Provisioned ... + Provisioned ServiceProviderProvisioningState = "Provisioned" + // Provisioning ... + Provisioning ServiceProviderProvisioningState = "Provisioning" +) + +// PossibleServiceProviderProvisioningStateValues returns an array of possible values for the ServiceProviderProvisioningState const type. +func PossibleServiceProviderProvisioningStateValues() []ServiceProviderProvisioningState { + return []ServiceProviderProvisioningState{Deprovisioning, NotProvisioned, Provisioned, Provisioning} +} + +// TransportProtocol enumerates the values for transport protocol. +type TransportProtocol string + +const ( + // TransportProtocolTCP ... + TransportProtocolTCP TransportProtocol = "Tcp" + // TransportProtocolUDP ... + TransportProtocolUDP TransportProtocol = "Udp" +) + +// PossibleTransportProtocolValues returns an array of possible values for the TransportProtocol const type. +func PossibleTransportProtocolValues() []TransportProtocol { + return []TransportProtocol{TransportProtocolTCP, TransportProtocolUDP} +} + +// VirtualNetworkGatewayConnectionStatus enumerates the values for virtual network gateway connection status. +type VirtualNetworkGatewayConnectionStatus string + +const ( + // Connected ... + Connected VirtualNetworkGatewayConnectionStatus = "Connected" + // Connecting ... + Connecting VirtualNetworkGatewayConnectionStatus = "Connecting" + // NotConnected ... + NotConnected VirtualNetworkGatewayConnectionStatus = "NotConnected" + // Unknown ... + Unknown VirtualNetworkGatewayConnectionStatus = "Unknown" +) + +// PossibleVirtualNetworkGatewayConnectionStatusValues returns an array of possible values for the VirtualNetworkGatewayConnectionStatus const type. +func PossibleVirtualNetworkGatewayConnectionStatusValues() []VirtualNetworkGatewayConnectionStatus { + return []VirtualNetworkGatewayConnectionStatus{Connected, Connecting, NotConnected, Unknown} +} + +// VirtualNetworkGatewayConnectionType enumerates the values for virtual network gateway connection type. +type VirtualNetworkGatewayConnectionType string + +const ( + // ExpressRoute ... + ExpressRoute VirtualNetworkGatewayConnectionType = "ExpressRoute" + // IPsec ... + IPsec VirtualNetworkGatewayConnectionType = "IPsec" + // Vnet2Vnet ... + Vnet2Vnet VirtualNetworkGatewayConnectionType = "Vnet2Vnet" + // VPNClient ... + VPNClient VirtualNetworkGatewayConnectionType = "VPNClient" +) + +// PossibleVirtualNetworkGatewayConnectionTypeValues returns an array of possible values for the VirtualNetworkGatewayConnectionType const type. +func PossibleVirtualNetworkGatewayConnectionTypeValues() []VirtualNetworkGatewayConnectionType { + return []VirtualNetworkGatewayConnectionType{ExpressRoute, IPsec, Vnet2Vnet, VPNClient} +} + +// VirtualNetworkGatewaySkuName enumerates the values for virtual network gateway sku name. +type VirtualNetworkGatewaySkuName string + +const ( + // VirtualNetworkGatewaySkuNameBasic ... + VirtualNetworkGatewaySkuNameBasic VirtualNetworkGatewaySkuName = "Basic" + // VirtualNetworkGatewaySkuNameHighPerformance ... + VirtualNetworkGatewaySkuNameHighPerformance VirtualNetworkGatewaySkuName = "HighPerformance" + // VirtualNetworkGatewaySkuNameStandard ... + VirtualNetworkGatewaySkuNameStandard VirtualNetworkGatewaySkuName = "Standard" +) + +// PossibleVirtualNetworkGatewaySkuNameValues returns an array of possible values for the VirtualNetworkGatewaySkuName const type. +func PossibleVirtualNetworkGatewaySkuNameValues() []VirtualNetworkGatewaySkuName { + return []VirtualNetworkGatewaySkuName{VirtualNetworkGatewaySkuNameBasic, VirtualNetworkGatewaySkuNameHighPerformance, VirtualNetworkGatewaySkuNameStandard} +} + +// VirtualNetworkGatewaySkuTier enumerates the values for virtual network gateway sku tier. +type VirtualNetworkGatewaySkuTier string + +const ( + // VirtualNetworkGatewaySkuTierBasic ... + VirtualNetworkGatewaySkuTierBasic VirtualNetworkGatewaySkuTier = "Basic" + // VirtualNetworkGatewaySkuTierHighPerformance ... + VirtualNetworkGatewaySkuTierHighPerformance VirtualNetworkGatewaySkuTier = "HighPerformance" + // VirtualNetworkGatewaySkuTierStandard ... + VirtualNetworkGatewaySkuTierStandard VirtualNetworkGatewaySkuTier = "Standard" +) + +// PossibleVirtualNetworkGatewaySkuTierValues returns an array of possible values for the VirtualNetworkGatewaySkuTier const type. +func PossibleVirtualNetworkGatewaySkuTierValues() []VirtualNetworkGatewaySkuTier { + return []VirtualNetworkGatewaySkuTier{VirtualNetworkGatewaySkuTierBasic, VirtualNetworkGatewaySkuTierHighPerformance, VirtualNetworkGatewaySkuTierStandard} +} + +// VirtualNetworkGatewayType enumerates the values for virtual network gateway type. +type VirtualNetworkGatewayType string + +const ( + // VirtualNetworkGatewayTypeExpressRoute ... + VirtualNetworkGatewayTypeExpressRoute VirtualNetworkGatewayType = "ExpressRoute" + // VirtualNetworkGatewayTypeVpn ... + VirtualNetworkGatewayTypeVpn VirtualNetworkGatewayType = "Vpn" +) + +// PossibleVirtualNetworkGatewayTypeValues returns an array of possible values for the VirtualNetworkGatewayType const type. +func PossibleVirtualNetworkGatewayTypeValues() []VirtualNetworkGatewayType { + return []VirtualNetworkGatewayType{VirtualNetworkGatewayTypeExpressRoute, VirtualNetworkGatewayTypeVpn} +} + +// VpnType enumerates the values for vpn type. +type VpnType string + +const ( + // PolicyBased ... + PolicyBased VpnType = "PolicyBased" + // RouteBased ... + RouteBased VpnType = "RouteBased" +) + +// PossibleVpnTypeValues returns an array of possible values for the VpnType const type. +func PossibleVpnTypeValues() []VpnType { + return []VpnType{PolicyBased, RouteBased} +} + +// AddressSpace addressSpace contains an array of IP address ranges that can be used by subnets of the virtual +// network. +type AddressSpace struct { + // AddressPrefixes - A list of address blocks reserved for this virtual network in CIDR notation. + AddressPrefixes *[]string `json:"addressPrefixes,omitempty"` +} + +// ApplicationGateway application gateway resource +type ApplicationGateway struct { + autorest.Response `json:"-"` + *ApplicationGatewayPropertiesFormat `json:"properties,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for ApplicationGateway. +func (ag ApplicationGateway) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if ag.ApplicationGatewayPropertiesFormat != nil { + objectMap["properties"] = ag.ApplicationGatewayPropertiesFormat + } + if ag.Etag != nil { + objectMap["etag"] = ag.Etag + } + if ag.ID != nil { + objectMap["id"] = ag.ID + } + if ag.Name != nil { + objectMap["name"] = ag.Name + } + if ag.Type != nil { + objectMap["type"] = ag.Type + } + if ag.Location != nil { + objectMap["location"] = ag.Location + } + if ag.Tags != nil { + objectMap["tags"] = ag.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGateway struct. +func (ag *ApplicationGateway) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewayPropertiesFormat ApplicationGatewayPropertiesFormat + err = json.Unmarshal(*v, &applicationGatewayPropertiesFormat) + if err != nil { + return err + } + ag.ApplicationGatewayPropertiesFormat = &applicationGatewayPropertiesFormat + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + ag.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + ag.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + ag.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + ag.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + ag.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + ag.Tags = tags + } + } + } + + return nil +} + +// ApplicationGatewayBackendAddress backend address of an application gateway. +type ApplicationGatewayBackendAddress struct { + // Fqdn - Fully qualified domain name (FQDN). + Fqdn *string `json:"fqdn,omitempty"` + // IPAddress - IP address + IPAddress *string `json:"ipAddress,omitempty"` +} + +// ApplicationGatewayBackendAddressPool backend Address Pool of an application gateway. +type ApplicationGatewayBackendAddressPool struct { + *ApplicationGatewayBackendAddressPoolPropertiesFormat `json:"properties,omitempty"` + // Name - Resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ApplicationGatewayBackendAddressPool. +func (agbap ApplicationGatewayBackendAddressPool) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if agbap.ApplicationGatewayBackendAddressPoolPropertiesFormat != nil { + objectMap["properties"] = agbap.ApplicationGatewayBackendAddressPoolPropertiesFormat + } + if agbap.Name != nil { + objectMap["name"] = agbap.Name + } + if agbap.Etag != nil { + objectMap["etag"] = agbap.Etag + } + if agbap.ID != nil { + objectMap["id"] = agbap.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGatewayBackendAddressPool struct. +func (agbap *ApplicationGatewayBackendAddressPool) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewayBackendAddressPoolPropertiesFormat ApplicationGatewayBackendAddressPoolPropertiesFormat + err = json.Unmarshal(*v, &applicationGatewayBackendAddressPoolPropertiesFormat) + if err != nil { + return err + } + agbap.ApplicationGatewayBackendAddressPoolPropertiesFormat = &applicationGatewayBackendAddressPoolPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + agbap.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + agbap.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + agbap.ID = &ID + } + } + } + + return nil +} + +// ApplicationGatewayBackendAddressPoolPropertiesFormat properties of Backend Address Pool of an application +// gateway. +type ApplicationGatewayBackendAddressPoolPropertiesFormat struct { + // BackendIPConfigurations - Collection of references to IPs defined in network interfaces. + BackendIPConfigurations *[]InterfaceIPConfiguration `json:"backendIPConfigurations,omitempty"` + // BackendAddresses - Backend addresses + BackendAddresses *[]ApplicationGatewayBackendAddress `json:"backendAddresses,omitempty"` + // ProvisioningState - Provisioning state of the backend address pool resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ApplicationGatewayBackendHTTPSettings backend address pool settings of an application gateway. +type ApplicationGatewayBackendHTTPSettings struct { + *ApplicationGatewayBackendHTTPSettingsPropertiesFormat `json:"properties,omitempty"` + // Name - Name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ApplicationGatewayBackendHTTPSettings. +func (agbhs ApplicationGatewayBackendHTTPSettings) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if agbhs.ApplicationGatewayBackendHTTPSettingsPropertiesFormat != nil { + objectMap["properties"] = agbhs.ApplicationGatewayBackendHTTPSettingsPropertiesFormat + } + if agbhs.Name != nil { + objectMap["name"] = agbhs.Name + } + if agbhs.Etag != nil { + objectMap["etag"] = agbhs.Etag + } + if agbhs.ID != nil { + objectMap["id"] = agbhs.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGatewayBackendHTTPSettings struct. +func (agbhs *ApplicationGatewayBackendHTTPSettings) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewayBackendHTTPSettingsPropertiesFormat ApplicationGatewayBackendHTTPSettingsPropertiesFormat + err = json.Unmarshal(*v, &applicationGatewayBackendHTTPSettingsPropertiesFormat) + if err != nil { + return err + } + agbhs.ApplicationGatewayBackendHTTPSettingsPropertiesFormat = &applicationGatewayBackendHTTPSettingsPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + agbhs.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + agbhs.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + agbhs.ID = &ID + } + } + } + + return nil +} + +// ApplicationGatewayBackendHTTPSettingsPropertiesFormat properties of Backend address pool settings of an +// application gateway. +type ApplicationGatewayBackendHTTPSettingsPropertiesFormat struct { + // Port - Port + Port *int32 `json:"port,omitempty"` + // Protocol - Protocol. Possible values are: 'Http' and 'Https'. Possible values include: 'HTTP', 'HTTPS' + Protocol ApplicationGatewayProtocol `json:"protocol,omitempty"` + // CookieBasedAffinity - Cookie based affinity. Possible values are: 'Enabled' and 'Disabled'. Possible values include: 'Enabled', 'Disabled' + CookieBasedAffinity ApplicationGatewayCookieBasedAffinity `json:"cookieBasedAffinity,omitempty"` + // RequestTimeout - Request timeout in seconds. Application Gateway will fail the request if response is not received within RequestTimeout. Acceptable values are from 1 second to 86400 seconds. + RequestTimeout *int32 `json:"requestTimeout,omitempty"` + // Probe - Probe resource of an application gateway. + Probe *SubResource `json:"probe,omitempty"` + // ProvisioningState - Gets or sets Provisioning state of the backend http settings resource Updating/Deleting/Failed + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ApplicationGatewayFrontendIPConfiguration frontend IP configuration of an application gateway. +type ApplicationGatewayFrontendIPConfiguration struct { + *ApplicationGatewayFrontendIPConfigurationPropertiesFormat `json:"properties,omitempty"` + // Name - Name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ApplicationGatewayFrontendIPConfiguration. +func (agfic ApplicationGatewayFrontendIPConfiguration) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if agfic.ApplicationGatewayFrontendIPConfigurationPropertiesFormat != nil { + objectMap["properties"] = agfic.ApplicationGatewayFrontendIPConfigurationPropertiesFormat + } + if agfic.Name != nil { + objectMap["name"] = agfic.Name + } + if agfic.Etag != nil { + objectMap["etag"] = agfic.Etag + } + if agfic.ID != nil { + objectMap["id"] = agfic.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGatewayFrontendIPConfiguration struct. +func (agfic *ApplicationGatewayFrontendIPConfiguration) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewayFrontendIPConfigurationPropertiesFormat ApplicationGatewayFrontendIPConfigurationPropertiesFormat + err = json.Unmarshal(*v, &applicationGatewayFrontendIPConfigurationPropertiesFormat) + if err != nil { + return err + } + agfic.ApplicationGatewayFrontendIPConfigurationPropertiesFormat = &applicationGatewayFrontendIPConfigurationPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + agfic.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + agfic.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + agfic.ID = &ID + } + } + } + + return nil +} + +// ApplicationGatewayFrontendIPConfigurationPropertiesFormat properties of Frontend IP configuration of an +// application gateway. +type ApplicationGatewayFrontendIPConfigurationPropertiesFormat struct { + // PrivateIPAddress - PrivateIPAddress of the network interface IP Configuration. + PrivateIPAddress *string `json:"privateIPAddress,omitempty"` + // PrivateIPAllocationMethod - PrivateIP allocation method. Possible values are: 'Static' and 'Dynamic'. Possible values include: 'Static', 'Dynamic' + PrivateIPAllocationMethod IPAllocationMethod `json:"privateIPAllocationMethod,omitempty"` + // Subnet - Reference of the subnet resource. + Subnet *SubResource `json:"subnet,omitempty"` + // PublicIPAddress - Reference of the PublicIP resource. + PublicIPAddress *SubResource `json:"publicIPAddress,omitempty"` + // ProvisioningState - Provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ApplicationGatewayFrontendPort frontend port of an application gateway. +type ApplicationGatewayFrontendPort struct { + *ApplicationGatewayFrontendPortPropertiesFormat `json:"properties,omitempty"` + // Name - Name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ApplicationGatewayFrontendPort. +func (agfp ApplicationGatewayFrontendPort) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if agfp.ApplicationGatewayFrontendPortPropertiesFormat != nil { + objectMap["properties"] = agfp.ApplicationGatewayFrontendPortPropertiesFormat + } + if agfp.Name != nil { + objectMap["name"] = agfp.Name + } + if agfp.Etag != nil { + objectMap["etag"] = agfp.Etag + } + if agfp.ID != nil { + objectMap["id"] = agfp.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGatewayFrontendPort struct. +func (agfp *ApplicationGatewayFrontendPort) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewayFrontendPortPropertiesFormat ApplicationGatewayFrontendPortPropertiesFormat + err = json.Unmarshal(*v, &applicationGatewayFrontendPortPropertiesFormat) + if err != nil { + return err + } + agfp.ApplicationGatewayFrontendPortPropertiesFormat = &applicationGatewayFrontendPortPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + agfp.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + agfp.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + agfp.ID = &ID + } + } + } + + return nil +} + +// ApplicationGatewayFrontendPortPropertiesFormat properties of Frontend port of an application gateway. +type ApplicationGatewayFrontendPortPropertiesFormat struct { + // Port - Frontend port + Port *int32 `json:"port,omitempty"` + // ProvisioningState - Provisioning state of the frontend port resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ApplicationGatewayHTTPListener http listener of an application gateway. +type ApplicationGatewayHTTPListener struct { + *ApplicationGatewayHTTPListenerPropertiesFormat `json:"properties,omitempty"` + // Name - Name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ApplicationGatewayHTTPListener. +func (aghl ApplicationGatewayHTTPListener) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if aghl.ApplicationGatewayHTTPListenerPropertiesFormat != nil { + objectMap["properties"] = aghl.ApplicationGatewayHTTPListenerPropertiesFormat + } + if aghl.Name != nil { + objectMap["name"] = aghl.Name + } + if aghl.Etag != nil { + objectMap["etag"] = aghl.Etag + } + if aghl.ID != nil { + objectMap["id"] = aghl.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGatewayHTTPListener struct. +func (aghl *ApplicationGatewayHTTPListener) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewayHTTPListenerPropertiesFormat ApplicationGatewayHTTPListenerPropertiesFormat + err = json.Unmarshal(*v, &applicationGatewayHTTPListenerPropertiesFormat) + if err != nil { + return err + } + aghl.ApplicationGatewayHTTPListenerPropertiesFormat = &applicationGatewayHTTPListenerPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + aghl.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + aghl.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + aghl.ID = &ID + } + } + } + + return nil +} + +// ApplicationGatewayHTTPListenerPropertiesFormat properties of HTTP listener of an application gateway. +type ApplicationGatewayHTTPListenerPropertiesFormat struct { + // FrontendIPConfiguration - Frontend IP configuration resource of an application gateway. + FrontendIPConfiguration *SubResource `json:"frontendIPConfiguration,omitempty"` + // FrontendPort - Frontend port resource of an application gateway. + FrontendPort *SubResource `json:"frontendPort,omitempty"` + // Protocol - Protocol. Possible values are: 'Http' and 'Https'. Possible values include: 'HTTP', 'HTTPS' + Protocol ApplicationGatewayProtocol `json:"protocol,omitempty"` + // HostName - Host name of HTTP listener. + HostName *string `json:"hostName,omitempty"` + // SslCertificate - SSL certificate resource of an application gateway. + SslCertificate *SubResource `json:"sslCertificate,omitempty"` + // RequireServerNameIndication - Applicable only if protocol is https. Enables SNI for multi-hosting. + RequireServerNameIndication *bool `json:"requireServerNameIndication,omitempty"` + // ProvisioningState - Provisioning state of the HTTP listener resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ApplicationGatewayIPConfiguration IP configuration of an application gateway. Currently 1 public and 1 private +// IP configuration is allowed. +type ApplicationGatewayIPConfiguration struct { + *ApplicationGatewayIPConfigurationPropertiesFormat `json:"properties,omitempty"` + // Name - Name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ApplicationGatewayIPConfiguration. +func (agic ApplicationGatewayIPConfiguration) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if agic.ApplicationGatewayIPConfigurationPropertiesFormat != nil { + objectMap["properties"] = agic.ApplicationGatewayIPConfigurationPropertiesFormat + } + if agic.Name != nil { + objectMap["name"] = agic.Name + } + if agic.Etag != nil { + objectMap["etag"] = agic.Etag + } + if agic.ID != nil { + objectMap["id"] = agic.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGatewayIPConfiguration struct. +func (agic *ApplicationGatewayIPConfiguration) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewayIPConfigurationPropertiesFormat ApplicationGatewayIPConfigurationPropertiesFormat + err = json.Unmarshal(*v, &applicationGatewayIPConfigurationPropertiesFormat) + if err != nil { + return err + } + agic.ApplicationGatewayIPConfigurationPropertiesFormat = &applicationGatewayIPConfigurationPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + agic.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + agic.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + agic.ID = &ID + } + } + } + + return nil +} + +// ApplicationGatewayIPConfigurationPropertiesFormat properties of IP configuration of an application gateway. +type ApplicationGatewayIPConfigurationPropertiesFormat struct { + // Subnet - Reference of the subnet resource. A subnet from where application gateway gets its private address. + Subnet *SubResource `json:"subnet,omitempty"` + // ProvisioningState - Provisioning state of the application gateway subnet resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ApplicationGatewayListResult response for ListApplicationGateways API service call. +type ApplicationGatewayListResult struct { + autorest.Response `json:"-"` + // Value - List of an application gateways in a resource group. + Value *[]ApplicationGateway `json:"value,omitempty"` + // NextLink - URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// ApplicationGatewayListResultIterator provides access to a complete listing of ApplicationGateway values. +type ApplicationGatewayListResultIterator struct { + i int + page ApplicationGatewayListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *ApplicationGatewayListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter ApplicationGatewayListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter ApplicationGatewayListResultIterator) Response() ApplicationGatewayListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter ApplicationGatewayListResultIterator) Value() ApplicationGateway { + if !iter.page.NotDone() { + return ApplicationGateway{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (aglr ApplicationGatewayListResult) IsEmpty() bool { + return aglr.Value == nil || len(*aglr.Value) == 0 +} + +// applicationGatewayListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (aglr ApplicationGatewayListResult) applicationGatewayListResultPreparer() (*http.Request, error) { + if aglr.NextLink == nil || len(to.String(aglr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(aglr.NextLink))) +} + +// ApplicationGatewayListResultPage contains a page of ApplicationGateway values. +type ApplicationGatewayListResultPage struct { + fn func(ApplicationGatewayListResult) (ApplicationGatewayListResult, error) + aglr ApplicationGatewayListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *ApplicationGatewayListResultPage) Next() error { + next, err := page.fn(page.aglr) + if err != nil { + return err + } + page.aglr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page ApplicationGatewayListResultPage) NotDone() bool { + return !page.aglr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page ApplicationGatewayListResultPage) Response() ApplicationGatewayListResult { + return page.aglr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page ApplicationGatewayListResultPage) Values() []ApplicationGateway { + if page.aglr.IsEmpty() { + return nil + } + return *page.aglr.Value +} + +// ApplicationGatewayPathRule path rule of URL path map of an application gateway. +type ApplicationGatewayPathRule struct { + *ApplicationGatewayPathRulePropertiesFormat `json:"properties,omitempty"` + // Name - Name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ApplicationGatewayPathRule. +func (agpr ApplicationGatewayPathRule) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if agpr.ApplicationGatewayPathRulePropertiesFormat != nil { + objectMap["properties"] = agpr.ApplicationGatewayPathRulePropertiesFormat + } + if agpr.Name != nil { + objectMap["name"] = agpr.Name + } + if agpr.Etag != nil { + objectMap["etag"] = agpr.Etag + } + if agpr.ID != nil { + objectMap["id"] = agpr.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGatewayPathRule struct. +func (agpr *ApplicationGatewayPathRule) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewayPathRulePropertiesFormat ApplicationGatewayPathRulePropertiesFormat + err = json.Unmarshal(*v, &applicationGatewayPathRulePropertiesFormat) + if err != nil { + return err + } + agpr.ApplicationGatewayPathRulePropertiesFormat = &applicationGatewayPathRulePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + agpr.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + agpr.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + agpr.ID = &ID + } + } + } + + return nil +} + +// ApplicationGatewayPathRulePropertiesFormat properties of probe of an application gateway. +type ApplicationGatewayPathRulePropertiesFormat struct { + // Paths - Path rules of URL path map. + Paths *[]string `json:"paths,omitempty"` + // BackendAddressPool - Backend address pool resource of URL path map. + BackendAddressPool *SubResource `json:"backendAddressPool,omitempty"` + // BackendHTTPSettings - Backend http settings resource of URL path map. + BackendHTTPSettings *SubResource `json:"backendHttpSettings,omitempty"` + // ProvisioningState - Path rule of URL path map resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ApplicationGatewayProbe probe of the application gateway. +type ApplicationGatewayProbe struct { + *ApplicationGatewayProbePropertiesFormat `json:"properties,omitempty"` + // Name - Name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ApplicationGatewayProbe. +func (agp ApplicationGatewayProbe) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if agp.ApplicationGatewayProbePropertiesFormat != nil { + objectMap["properties"] = agp.ApplicationGatewayProbePropertiesFormat + } + if agp.Name != nil { + objectMap["name"] = agp.Name + } + if agp.Etag != nil { + objectMap["etag"] = agp.Etag + } + if agp.ID != nil { + objectMap["id"] = agp.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGatewayProbe struct. +func (agp *ApplicationGatewayProbe) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewayProbePropertiesFormat ApplicationGatewayProbePropertiesFormat + err = json.Unmarshal(*v, &applicationGatewayProbePropertiesFormat) + if err != nil { + return err + } + agp.ApplicationGatewayProbePropertiesFormat = &applicationGatewayProbePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + agp.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + agp.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + agp.ID = &ID + } + } + } + + return nil +} + +// ApplicationGatewayProbePropertiesFormat properties of probe of an application gateway. +type ApplicationGatewayProbePropertiesFormat struct { + // Protocol - Protocol. Possible values are: 'Http' and 'Https'. Possible values include: 'HTTP', 'HTTPS' + Protocol ApplicationGatewayProtocol `json:"protocol,omitempty"` + // Host - Host name to send the probe to. + Host *string `json:"host,omitempty"` + // Path - Relative path of probe. Valid path starts from '/'. Probe is sent to ://: + Path *string `json:"path,omitempty"` + // Interval - The probing interval in seconds. This is the time interval between two consecutive probes. Acceptable values are from 1 second to 86400 seconds. + Interval *int32 `json:"interval,omitempty"` + // Timeout - the probe timeout in seconds. Probe marked as failed if valid response is not received with this timeout period. Acceptable values are from 1 second to 86400 seconds. + Timeout *int32 `json:"timeout,omitempty"` + // UnhealthyThreshold - The probe retry count. Backend server is marked down after consecutive probe failure count reaches UnhealthyThreshold. Acceptable values are from 1 second to 20. + UnhealthyThreshold *int32 `json:"unhealthyThreshold,omitempty"` + // ProvisioningState - Provisioning state of the backend http settings resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ApplicationGatewayPropertiesFormat properties of the application gateway. +type ApplicationGatewayPropertiesFormat struct { + // Sku - SKU of the application gateway resource. + Sku *ApplicationGatewaySku `json:"sku,omitempty"` + // OperationalState - Operational state of the application gateway resource. Possible values are: 'Stopped', 'Started', 'Running', and 'Stopping'. Possible values include: 'Stopped', 'Starting', 'Running', 'Stopping' + OperationalState ApplicationGatewayOperationalState `json:"operationalState,omitempty"` + // GatewayIPConfigurations - Gets or sets subnets of application gateway resource + GatewayIPConfigurations *[]ApplicationGatewayIPConfiguration `json:"gatewayIPConfigurations,omitempty"` + // SslCertificates - SSL certificates of the application gateway resource. + SslCertificates *[]ApplicationGatewaySslCertificate `json:"sslCertificates,omitempty"` + // FrontendIPConfigurations - Frontend IP addresses of the application gateway resource. + FrontendIPConfigurations *[]ApplicationGatewayFrontendIPConfiguration `json:"frontendIPConfigurations,omitempty"` + // FrontendPorts - Frontend ports of the application gateway resource. + FrontendPorts *[]ApplicationGatewayFrontendPort `json:"frontendPorts,omitempty"` + // Probes - Probes of the application gateway resource. + Probes *[]ApplicationGatewayProbe `json:"probes,omitempty"` + // BackendAddressPools - Backend address pool of the application gateway resource. + BackendAddressPools *[]ApplicationGatewayBackendAddressPool `json:"backendAddressPools,omitempty"` + // BackendHTTPSettingsCollection - Backend http settings of the application gateway resource. + BackendHTTPSettingsCollection *[]ApplicationGatewayBackendHTTPSettings `json:"backendHttpSettingsCollection,omitempty"` + // HTTPListeners - Http listeners of the application gateway resource. + HTTPListeners *[]ApplicationGatewayHTTPListener `json:"httpListeners,omitempty"` + // URLPathMaps - URL path map of the application gateway resource. + URLPathMaps *[]ApplicationGatewayURLPathMap `json:"urlPathMaps,omitempty"` + // RequestRoutingRules - Request routing rules of the application gateway resource. + RequestRoutingRules *[]ApplicationGatewayRequestRoutingRule `json:"requestRoutingRules,omitempty"` + // ResourceGUID - Resource GUID property of the application gateway resource. + ResourceGUID *string `json:"resourceGuid,omitempty"` + // ProvisioningState - Provisioning state of the application gateway resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ApplicationGatewayRequestRoutingRule request routing rule of an application gateway. +type ApplicationGatewayRequestRoutingRule struct { + *ApplicationGatewayRequestRoutingRulePropertiesFormat `json:"properties,omitempty"` + // Name - Name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ApplicationGatewayRequestRoutingRule. +func (agrrr ApplicationGatewayRequestRoutingRule) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if agrrr.ApplicationGatewayRequestRoutingRulePropertiesFormat != nil { + objectMap["properties"] = agrrr.ApplicationGatewayRequestRoutingRulePropertiesFormat + } + if agrrr.Name != nil { + objectMap["name"] = agrrr.Name + } + if agrrr.Etag != nil { + objectMap["etag"] = agrrr.Etag + } + if agrrr.ID != nil { + objectMap["id"] = agrrr.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGatewayRequestRoutingRule struct. +func (agrrr *ApplicationGatewayRequestRoutingRule) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewayRequestRoutingRulePropertiesFormat ApplicationGatewayRequestRoutingRulePropertiesFormat + err = json.Unmarshal(*v, &applicationGatewayRequestRoutingRulePropertiesFormat) + if err != nil { + return err + } + agrrr.ApplicationGatewayRequestRoutingRulePropertiesFormat = &applicationGatewayRequestRoutingRulePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + agrrr.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + agrrr.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + agrrr.ID = &ID + } + } + } + + return nil +} + +// ApplicationGatewayRequestRoutingRulePropertiesFormat properties of request routing rule of the application +// gateway. +type ApplicationGatewayRequestRoutingRulePropertiesFormat struct { + // RuleType - Rule type. Possible values are: 'Basic' and 'PathBasedRouting'. Possible values include: 'Basic', 'PathBasedRouting' + RuleType ApplicationGatewayRequestRoutingRuleType `json:"ruleType,omitempty"` + // BackendAddressPool - Backend address pool resource of the application gateway. + BackendAddressPool *SubResource `json:"backendAddressPool,omitempty"` + // BackendHTTPSettings - Frontend port resource of the application gateway. + BackendHTTPSettings *SubResource `json:"backendHttpSettings,omitempty"` + // HTTPListener - Http listener resource of the application gateway. + HTTPListener *SubResource `json:"httpListener,omitempty"` + // URLPathMap - URL path map resource of the application gateway. + URLPathMap *SubResource `json:"urlPathMap,omitempty"` + // ProvisioningState - Provisioning state of the request routing rule resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ApplicationGatewaysCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a +// long-running operation. +type ApplicationGatewaysCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future ApplicationGatewaysCreateOrUpdateFuture) Result(client ApplicationGatewaysClient) (ag ApplicationGateway, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ag, azure.NewAsyncOpIncompleteError("network.ApplicationGatewaysCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ag, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + ag, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// ApplicationGatewaysDeleteFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type ApplicationGatewaysDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future ApplicationGatewaysDeleteFuture) Result(client ApplicationGatewaysClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.ApplicationGatewaysDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// ApplicationGatewaySku SKU of application gateway +type ApplicationGatewaySku struct { + // Name - Name of an application gateway SKU. Possible values are: 'Standard_Small', 'Standard_Medium', 'Standard_Large', 'WAF_Medium', and 'WAF_Large'. Possible values include: 'StandardSmall', 'StandardMedium', 'StandardLarge' + Name ApplicationGatewaySkuName `json:"name,omitempty"` + // Tier - Tier of an application gateway. Possible values include: 'Standard' + Tier ApplicationGatewayTier `json:"tier,omitempty"` + // Capacity - Capacity (instance count) of an application gateway. + Capacity *int32 `json:"capacity,omitempty"` +} + +// ApplicationGatewaySslCertificate SSL certificates of an application gateway. +type ApplicationGatewaySslCertificate struct { + *ApplicationGatewaySslCertificatePropertiesFormat `json:"properties,omitempty"` + // Name - Name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ApplicationGatewaySslCertificate. +func (agsc ApplicationGatewaySslCertificate) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if agsc.ApplicationGatewaySslCertificatePropertiesFormat != nil { + objectMap["properties"] = agsc.ApplicationGatewaySslCertificatePropertiesFormat + } + if agsc.Name != nil { + objectMap["name"] = agsc.Name + } + if agsc.Etag != nil { + objectMap["etag"] = agsc.Etag + } + if agsc.ID != nil { + objectMap["id"] = agsc.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGatewaySslCertificate struct. +func (agsc *ApplicationGatewaySslCertificate) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewaySslCertificatePropertiesFormat ApplicationGatewaySslCertificatePropertiesFormat + err = json.Unmarshal(*v, &applicationGatewaySslCertificatePropertiesFormat) + if err != nil { + return err + } + agsc.ApplicationGatewaySslCertificatePropertiesFormat = &applicationGatewaySslCertificatePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + agsc.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + agsc.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + agsc.ID = &ID + } + } + } + + return nil +} + +// ApplicationGatewaySslCertificatePropertiesFormat properties of SSL certificates of an application gateway. +type ApplicationGatewaySslCertificatePropertiesFormat struct { + // Data - Base-64 encoded pfx certificate. Only applicable in PUT Request. + Data *string `json:"data,omitempty"` + // Password - Password for the pfx file specified in data. Only applicable in PUT request. + Password *string `json:"password,omitempty"` + // PublicCertData - Base-64 encoded Public cert data corresponding to pfx specified in data. Only applicable in GET request. + PublicCertData *string `json:"publicCertData,omitempty"` + // ProvisioningState - Provisioning state of the SSL certificate resource Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ApplicationGatewaysStartFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type ApplicationGatewaysStartFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future ApplicationGatewaysStartFuture) Result(client ApplicationGatewaysClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysStartFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.ApplicationGatewaysStartFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.StartResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysStartFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysStartFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.StartResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysStartFuture", "Result", resp, "Failure responding to request") + } + return +} + +// ApplicationGatewaysStopFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type ApplicationGatewaysStopFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future ApplicationGatewaysStopFuture) Result(client ApplicationGatewaysClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysStopFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.ApplicationGatewaysStopFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.StopResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysStopFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysStopFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.StopResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ApplicationGatewaysStopFuture", "Result", resp, "Failure responding to request") + } + return +} + +// ApplicationGatewayURLPathMap urlPathMaps give a url path to the backend mapping information for +// PathBasedRouting. +type ApplicationGatewayURLPathMap struct { + *ApplicationGatewayURLPathMapPropertiesFormat `json:"properties,omitempty"` + // Name - Name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ApplicationGatewayURLPathMap. +func (agupm ApplicationGatewayURLPathMap) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if agupm.ApplicationGatewayURLPathMapPropertiesFormat != nil { + objectMap["properties"] = agupm.ApplicationGatewayURLPathMapPropertiesFormat + } + if agupm.Name != nil { + objectMap["name"] = agupm.Name + } + if agupm.Etag != nil { + objectMap["etag"] = agupm.Etag + } + if agupm.ID != nil { + objectMap["id"] = agupm.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ApplicationGatewayURLPathMap struct. +func (agupm *ApplicationGatewayURLPathMap) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var applicationGatewayURLPathMapPropertiesFormat ApplicationGatewayURLPathMapPropertiesFormat + err = json.Unmarshal(*v, &applicationGatewayURLPathMapPropertiesFormat) + if err != nil { + return err + } + agupm.ApplicationGatewayURLPathMapPropertiesFormat = &applicationGatewayURLPathMapPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + agupm.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + agupm.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + agupm.ID = &ID + } + } + } + + return nil +} + +// ApplicationGatewayURLPathMapPropertiesFormat properties of UrlPathMap of the application gateway. +type ApplicationGatewayURLPathMapPropertiesFormat struct { + // DefaultBackendAddressPool - Default backend address pool resource of URL path map. + DefaultBackendAddressPool *SubResource `json:"defaultBackendAddressPool,omitempty"` + // DefaultBackendHTTPSettings - Default backend http settings resource of URL path map. + DefaultBackendHTTPSettings *SubResource `json:"defaultBackendHttpSettings,omitempty"` + // PathRules - Path rule of URL path map resource. + PathRules *[]ApplicationGatewayPathRule `json:"pathRules,omitempty"` + // ProvisioningState - Provisioning state of the backend http settings resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// AuthorizationListResult response for ListAuthorizations API service call retrieves all authorizations that +// belongs to an ExpressRouteCircuit. +type AuthorizationListResult struct { + autorest.Response `json:"-"` + // Value - The authorizations in an ExpressRoute Circuit. + Value *[]ExpressRouteCircuitAuthorization `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// AuthorizationListResultIterator provides access to a complete listing of ExpressRouteCircuitAuthorization +// values. +type AuthorizationListResultIterator struct { + i int + page AuthorizationListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *AuthorizationListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter AuthorizationListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter AuthorizationListResultIterator) Response() AuthorizationListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter AuthorizationListResultIterator) Value() ExpressRouteCircuitAuthorization { + if !iter.page.NotDone() { + return ExpressRouteCircuitAuthorization{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (alr AuthorizationListResult) IsEmpty() bool { + return alr.Value == nil || len(*alr.Value) == 0 +} + +// authorizationListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (alr AuthorizationListResult) authorizationListResultPreparer() (*http.Request, error) { + if alr.NextLink == nil || len(to.String(alr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(alr.NextLink))) +} + +// AuthorizationListResultPage contains a page of ExpressRouteCircuitAuthorization values. +type AuthorizationListResultPage struct { + fn func(AuthorizationListResult) (AuthorizationListResult, error) + alr AuthorizationListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *AuthorizationListResultPage) Next() error { + next, err := page.fn(page.alr) + if err != nil { + return err + } + page.alr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page AuthorizationListResultPage) NotDone() bool { + return !page.alr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page AuthorizationListResultPage) Response() AuthorizationListResult { + return page.alr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page AuthorizationListResultPage) Values() []ExpressRouteCircuitAuthorization { + if page.alr.IsEmpty() { + return nil + } + return *page.alr.Value +} + +// AuthorizationPropertiesFormat ... +type AuthorizationPropertiesFormat struct { + // AuthorizationKey - The authorization key. + AuthorizationKey *string `json:"authorizationKey,omitempty"` + // AuthorizationUseStatus - AuthorizationUseStatus. Possible values are: 'Available' and 'InUse'. Possible values include: 'Available', 'InUse' + AuthorizationUseStatus AuthorizationUseStatus `json:"authorizationUseStatus,omitempty"` + // ProvisioningState - Gets the provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// AzureAsyncOperationResult the response body contains the status of the specified asynchronous operation, +// indicating whether it has succeeded, is in progress, or has failed. Note that this status is distinct from the +// HTTP status code returned for the Get Operation Status operation itself. If the asynchronous operation +// succeeded, the response body includes the HTTP status code for the successful request. If the asynchronous +// operation failed, the response body includes the HTTP status code for the failed request and error information +// regarding the failure. +type AzureAsyncOperationResult struct { + // Status - Status of the Azure async operation. Possible values are: 'InProgress', 'Succeeded', and 'Failed'. Possible values include: 'InProgress', 'Succeeded', 'Failed' + Status OperationStatus `json:"status,omitempty"` + Error *Error `json:"error,omitempty"` +} + +// BackendAddressPool pool of backend IP addresses. +type BackendAddressPool struct { + *BackendAddressPoolPropertiesFormat `json:"properties,omitempty"` + // Name - Gets name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for BackendAddressPool. +func (bap BackendAddressPool) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if bap.BackendAddressPoolPropertiesFormat != nil { + objectMap["properties"] = bap.BackendAddressPoolPropertiesFormat + } + if bap.Name != nil { + objectMap["name"] = bap.Name + } + if bap.Etag != nil { + objectMap["etag"] = bap.Etag + } + if bap.ID != nil { + objectMap["id"] = bap.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for BackendAddressPool struct. +func (bap *BackendAddressPool) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var backendAddressPoolPropertiesFormat BackendAddressPoolPropertiesFormat + err = json.Unmarshal(*v, &backendAddressPoolPropertiesFormat) + if err != nil { + return err + } + bap.BackendAddressPoolPropertiesFormat = &backendAddressPoolPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + bap.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + bap.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + bap.ID = &ID + } + } + } + + return nil +} + +// BackendAddressPoolPropertiesFormat properties of the backend address pool. +type BackendAddressPoolPropertiesFormat struct { + // BackendIPConfigurations - Gets collection of references to IP addresses defined in network interfaces. + BackendIPConfigurations *[]InterfaceIPConfiguration `json:"backendIPConfigurations,omitempty"` + // LoadBalancingRules - Gets load balancing rules that use this backend address pool. + LoadBalancingRules *[]SubResource `json:"loadBalancingRules,omitempty"` + // OutboundNatRule - Gets outbound rules that use this backend address pool. + OutboundNatRule *SubResource `json:"outboundNatRule,omitempty"` + // ProvisioningState - Get provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// BgpSettings ... +type BgpSettings struct { + // Asn - Gets or sets this BGP speaker's ASN + Asn *int64 `json:"asn,omitempty"` + // BgpPeeringAddress - Gets or sets the BGP peering address and BGP identifier of this BGP speaker + BgpPeeringAddress *string `json:"bgpPeeringAddress,omitempty"` + // PeerWeight - Gets or sets the weight added to routes learned from this BGP speaker + PeerWeight *int32 `json:"peerWeight,omitempty"` +} + +// ConnectionResetSharedKey ... +type ConnectionResetSharedKey struct { + autorest.Response `json:"-"` + // KeyLength - The virtual network connection reset shared key length + KeyLength *int64 `json:"keyLength,omitempty"` +} + +// ConnectionSharedKey response for GetConnectionSharedKey Api servive call +type ConnectionSharedKey struct { + autorest.Response `json:"-"` + // Value - The virtual network connection shared key value + Value *string `json:"value,omitempty"` +} + +// ConnectionSharedKeyResult response for CheckConnectionSharedKey Api servive call +type ConnectionSharedKeyResult struct { + autorest.Response `json:"-"` + // Value - The virtual network connection shared key value + Value *string `json:"value,omitempty"` +} + +// DhcpOptions dhcpOptions contains an array of DNS servers available to VMs deployed in the virtual network. +// Standard DHCP option for a subnet overrides VNET DHCP options. +type DhcpOptions struct { + // DNSServers - The list of DNS servers IP addresses. + DNSServers *[]string `json:"dnsServers,omitempty"` +} + +// DNSNameAvailabilityResult response for the CheckDnsNameAvailability API service call. +type DNSNameAvailabilityResult struct { + autorest.Response `json:"-"` + // Available - Domain availability (True/False). + Available *bool `json:"available,omitempty"` +} + +// Error error object properties +type Error struct { + Code *string `json:"code,omitempty"` + Message *string `json:"message,omitempty"` + Target *string `json:"target,omitempty"` + Details *[]ErrorDetails `json:"details,omitempty"` + InnerError *string `json:"innerError,omitempty"` +} + +// ErrorDetails error details properties +type ErrorDetails struct { + Code *string `json:"code,omitempty"` + Target *string `json:"target,omitempty"` + Message *string `json:"message,omitempty"` +} + +// ExpressRouteCircuit expressRouteCircuit resource +type ExpressRouteCircuit struct { + autorest.Response `json:"-"` + // Sku - The SKU. + Sku *ExpressRouteCircuitSku `json:"sku,omitempty"` + *ExpressRouteCircuitPropertiesFormat `json:"properties,omitempty"` + // Etag - Gets a unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for ExpressRouteCircuit. +func (erc ExpressRouteCircuit) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if erc.Sku != nil { + objectMap["sku"] = erc.Sku + } + if erc.ExpressRouteCircuitPropertiesFormat != nil { + objectMap["properties"] = erc.ExpressRouteCircuitPropertiesFormat + } + if erc.Etag != nil { + objectMap["etag"] = erc.Etag + } + if erc.ID != nil { + objectMap["id"] = erc.ID + } + if erc.Name != nil { + objectMap["name"] = erc.Name + } + if erc.Type != nil { + objectMap["type"] = erc.Type + } + if erc.Location != nil { + objectMap["location"] = erc.Location + } + if erc.Tags != nil { + objectMap["tags"] = erc.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ExpressRouteCircuit struct. +func (erc *ExpressRouteCircuit) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "sku": + if v != nil { + var sku ExpressRouteCircuitSku + err = json.Unmarshal(*v, &sku) + if err != nil { + return err + } + erc.Sku = &sku + } + case "properties": + if v != nil { + var expressRouteCircuitPropertiesFormat ExpressRouteCircuitPropertiesFormat + err = json.Unmarshal(*v, &expressRouteCircuitPropertiesFormat) + if err != nil { + return err + } + erc.ExpressRouteCircuitPropertiesFormat = &expressRouteCircuitPropertiesFormat + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + erc.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + erc.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + erc.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + erc.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + erc.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + erc.Tags = tags + } + } + } + + return nil +} + +// ExpressRouteCircuitArpTable the ARP table associated with the ExpressRouteCircuit. +type ExpressRouteCircuitArpTable struct { + // IPAddress - The IP address. + IPAddress *string `json:"ipAddress,omitempty"` + // MacAddress - The MAC address. + MacAddress *string `json:"macAddress,omitempty"` +} + +// ExpressRouteCircuitAuthorization authorization in an ExpressRouteCircuit resource. +type ExpressRouteCircuitAuthorization struct { + autorest.Response `json:"-"` + *AuthorizationPropertiesFormat `json:"properties,omitempty"` + // Name - Gets name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ExpressRouteCircuitAuthorization. +func (erca ExpressRouteCircuitAuthorization) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if erca.AuthorizationPropertiesFormat != nil { + objectMap["properties"] = erca.AuthorizationPropertiesFormat + } + if erca.Name != nil { + objectMap["name"] = erca.Name + } + if erca.Etag != nil { + objectMap["etag"] = erca.Etag + } + if erca.ID != nil { + objectMap["id"] = erca.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ExpressRouteCircuitAuthorization struct. +func (erca *ExpressRouteCircuitAuthorization) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var authorizationPropertiesFormat AuthorizationPropertiesFormat + err = json.Unmarshal(*v, &authorizationPropertiesFormat) + if err != nil { + return err + } + erca.AuthorizationPropertiesFormat = &authorizationPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + erca.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + erca.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + erca.ID = &ID + } + } + } + + return nil +} + +// ExpressRouteCircuitAuthorizationsCreateOrUpdateFuture an abstraction for monitoring and retrieving the results +// of a long-running operation. +type ExpressRouteCircuitAuthorizationsCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future ExpressRouteCircuitAuthorizationsCreateOrUpdateFuture) Result(client ExpressRouteCircuitAuthorizationsClient) (erca ExpressRouteCircuitAuthorization, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return erca, azure.NewAsyncOpIncompleteError("network.ExpressRouteCircuitAuthorizationsCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + erca, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + erca, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// ExpressRouteCircuitAuthorizationsDeleteFuture an abstraction for monitoring and retrieving the results of a +// long-running operation. +type ExpressRouteCircuitAuthorizationsDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future ExpressRouteCircuitAuthorizationsDeleteFuture) Result(client ExpressRouteCircuitAuthorizationsClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.ExpressRouteCircuitAuthorizationsDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitAuthorizationsDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// ExpressRouteCircuitListResult response for ListExpressRouteCircuit API service call. +type ExpressRouteCircuitListResult struct { + autorest.Response `json:"-"` + // Value - A list of ExpressRouteCircuits in a resource group. + Value *[]ExpressRouteCircuit `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// ExpressRouteCircuitListResultIterator provides access to a complete listing of ExpressRouteCircuit values. +type ExpressRouteCircuitListResultIterator struct { + i int + page ExpressRouteCircuitListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *ExpressRouteCircuitListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter ExpressRouteCircuitListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter ExpressRouteCircuitListResultIterator) Response() ExpressRouteCircuitListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter ExpressRouteCircuitListResultIterator) Value() ExpressRouteCircuit { + if !iter.page.NotDone() { + return ExpressRouteCircuit{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (erclr ExpressRouteCircuitListResult) IsEmpty() bool { + return erclr.Value == nil || len(*erclr.Value) == 0 +} + +// expressRouteCircuitListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (erclr ExpressRouteCircuitListResult) expressRouteCircuitListResultPreparer() (*http.Request, error) { + if erclr.NextLink == nil || len(to.String(erclr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(erclr.NextLink))) +} + +// ExpressRouteCircuitListResultPage contains a page of ExpressRouteCircuit values. +type ExpressRouteCircuitListResultPage struct { + fn func(ExpressRouteCircuitListResult) (ExpressRouteCircuitListResult, error) + erclr ExpressRouteCircuitListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *ExpressRouteCircuitListResultPage) Next() error { + next, err := page.fn(page.erclr) + if err != nil { + return err + } + page.erclr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page ExpressRouteCircuitListResultPage) NotDone() bool { + return !page.erclr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page ExpressRouteCircuitListResultPage) Response() ExpressRouteCircuitListResult { + return page.erclr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page ExpressRouteCircuitListResultPage) Values() []ExpressRouteCircuit { + if page.erclr.IsEmpty() { + return nil + } + return *page.erclr.Value +} + +// ExpressRouteCircuitPeering peering in an ExpressRouteCircuit resource. +type ExpressRouteCircuitPeering struct { + autorest.Response `json:"-"` + *ExpressRouteCircuitPeeringPropertiesFormat `json:"properties,omitempty"` + // Name - Gets name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for ExpressRouteCircuitPeering. +func (ercp ExpressRouteCircuitPeering) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if ercp.ExpressRouteCircuitPeeringPropertiesFormat != nil { + objectMap["properties"] = ercp.ExpressRouteCircuitPeeringPropertiesFormat + } + if ercp.Name != nil { + objectMap["name"] = ercp.Name + } + if ercp.Etag != nil { + objectMap["etag"] = ercp.Etag + } + if ercp.ID != nil { + objectMap["id"] = ercp.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ExpressRouteCircuitPeering struct. +func (ercp *ExpressRouteCircuitPeering) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var expressRouteCircuitPeeringPropertiesFormat ExpressRouteCircuitPeeringPropertiesFormat + err = json.Unmarshal(*v, &expressRouteCircuitPeeringPropertiesFormat) + if err != nil { + return err + } + ercp.ExpressRouteCircuitPeeringPropertiesFormat = &expressRouteCircuitPeeringPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + ercp.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + ercp.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + ercp.ID = &ID + } + } + } + + return nil +} + +// ExpressRouteCircuitPeeringConfig specifies the peering configuration. +type ExpressRouteCircuitPeeringConfig struct { + // AdvertisedPublicPrefixes - The reference of AdvertisedPublicPrefixes. + AdvertisedPublicPrefixes *[]string `json:"advertisedPublicPrefixes,omitempty"` + // AdvertisedPublicPrefixesState - AdvertisedPublicPrefixState of the Peering resource. Possible values are 'NotConfigured', 'Configuring', 'Configured', and 'ValidationNeeded'. Possible values include: 'NotConfigured', 'Configuring', 'Configured', 'ValidationNeeded' + AdvertisedPublicPrefixesState ExpressRouteCircuitPeeringAdvertisedPublicPrefixState `json:"advertisedPublicPrefixesState,omitempty"` + // CustomerASN - The CustomerASN of the peering. + CustomerASN *int32 `json:"customerASN,omitempty"` + // RoutingRegistryName - The RoutingRegistryName of the configuration. + RoutingRegistryName *string `json:"routingRegistryName,omitempty"` +} + +// ExpressRouteCircuitPeeringListResult response for ListPeering API service call retrieves all peerings that +// belong to an ExpressRouteCircuit. +type ExpressRouteCircuitPeeringListResult struct { + autorest.Response `json:"-"` + // Value - The peerings in an express route circuit. + Value *[]ExpressRouteCircuitPeering `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// ExpressRouteCircuitPeeringListResultIterator provides access to a complete listing of ExpressRouteCircuitPeering +// values. +type ExpressRouteCircuitPeeringListResultIterator struct { + i int + page ExpressRouteCircuitPeeringListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *ExpressRouteCircuitPeeringListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter ExpressRouteCircuitPeeringListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter ExpressRouteCircuitPeeringListResultIterator) Response() ExpressRouteCircuitPeeringListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter ExpressRouteCircuitPeeringListResultIterator) Value() ExpressRouteCircuitPeering { + if !iter.page.NotDone() { + return ExpressRouteCircuitPeering{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (ercplr ExpressRouteCircuitPeeringListResult) IsEmpty() bool { + return ercplr.Value == nil || len(*ercplr.Value) == 0 +} + +// expressRouteCircuitPeeringListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (ercplr ExpressRouteCircuitPeeringListResult) expressRouteCircuitPeeringListResultPreparer() (*http.Request, error) { + if ercplr.NextLink == nil || len(to.String(ercplr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(ercplr.NextLink))) +} + +// ExpressRouteCircuitPeeringListResultPage contains a page of ExpressRouteCircuitPeering values. +type ExpressRouteCircuitPeeringListResultPage struct { + fn func(ExpressRouteCircuitPeeringListResult) (ExpressRouteCircuitPeeringListResult, error) + ercplr ExpressRouteCircuitPeeringListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *ExpressRouteCircuitPeeringListResultPage) Next() error { + next, err := page.fn(page.ercplr) + if err != nil { + return err + } + page.ercplr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page ExpressRouteCircuitPeeringListResultPage) NotDone() bool { + return !page.ercplr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page ExpressRouteCircuitPeeringListResultPage) Response() ExpressRouteCircuitPeeringListResult { + return page.ercplr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page ExpressRouteCircuitPeeringListResultPage) Values() []ExpressRouteCircuitPeering { + if page.ercplr.IsEmpty() { + return nil + } + return *page.ercplr.Value +} + +// ExpressRouteCircuitPeeringPropertiesFormat ... +type ExpressRouteCircuitPeeringPropertiesFormat struct { + // PeeringType - The PeeringType. Possible values are: 'AzurePublicPeering', 'AzurePrivatePeering', and 'MicrosoftPeering'. Possible values include: 'AzurePublicPeering', 'AzurePrivatePeering', 'MicrosoftPeering' + PeeringType ExpressRouteCircuitPeeringType `json:"peeringType,omitempty"` + // State - The state of peering. Possible values are: 'Disabled' and 'Enabled'. Possible values include: 'ExpressRouteCircuitPeeringStateDisabled', 'ExpressRouteCircuitPeeringStateEnabled' + State ExpressRouteCircuitPeeringState `json:"state,omitempty"` + // AzureASN - The Azure ASN. + AzureASN *int32 `json:"azureASN,omitempty"` + // PeerASN - The peer ASN. + PeerASN *int32 `json:"peerASN,omitempty"` + // PrimaryPeerAddressPrefix - The primary address prefix. + PrimaryPeerAddressPrefix *string `json:"primaryPeerAddressPrefix,omitempty"` + // SecondaryPeerAddressPrefix - The secondary address prefix. + SecondaryPeerAddressPrefix *string `json:"secondaryPeerAddressPrefix,omitempty"` + // PrimaryAzurePort - The primary port. + PrimaryAzurePort *string `json:"primaryAzurePort,omitempty"` + // SecondaryAzurePort - The secondary port. + SecondaryAzurePort *string `json:"secondaryAzurePort,omitempty"` + // SharedKey - The shared key. + SharedKey *string `json:"sharedKey,omitempty"` + // VlanID - The VLAN ID. + VlanID *int32 `json:"vlanId,omitempty"` + // MicrosoftPeeringConfig - The Microsoft peering configuration. + MicrosoftPeeringConfig *ExpressRouteCircuitPeeringConfig `json:"microsoftPeeringConfig,omitempty"` + // Stats - Gets peering stats. + Stats *ExpressRouteCircuitStats `json:"stats,omitempty"` + // ProvisioningState - Gets the provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ExpressRouteCircuitPeeringsCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a +// long-running operation. +type ExpressRouteCircuitPeeringsCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future ExpressRouteCircuitPeeringsCreateOrUpdateFuture) Result(client ExpressRouteCircuitPeeringsClient) (ercp ExpressRouteCircuitPeering, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ercp, azure.NewAsyncOpIncompleteError("network.ExpressRouteCircuitPeeringsCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ercp, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + ercp, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// ExpressRouteCircuitPeeringsDeleteFuture an abstraction for monitoring and retrieving the results of a +// long-running operation. +type ExpressRouteCircuitPeeringsDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future ExpressRouteCircuitPeeringsDeleteFuture) Result(client ExpressRouteCircuitPeeringsClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.ExpressRouteCircuitPeeringsDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitPeeringsDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// ExpressRouteCircuitPropertiesFormat properties of ExpressRouteCircuit. +type ExpressRouteCircuitPropertiesFormat struct { + // CircuitProvisioningState - The CircuitProvisioningState state of the resource. + CircuitProvisioningState *string `json:"circuitProvisioningState,omitempty"` + // ServiceProviderProvisioningState - The ServiceProviderProvisioningState state of the resource. Possible values are 'NotProvisioned', 'Provisioning', 'Provisioned', and 'Deprovisioning'. Possible values include: 'NotProvisioned', 'Provisioning', 'Provisioned', 'Deprovisioning' + ServiceProviderProvisioningState ServiceProviderProvisioningState `json:"serviceProviderProvisioningState,omitempty"` + // Authorizations - The list of authorizations. + Authorizations *[]ExpressRouteCircuitAuthorization `json:"authorizations,omitempty"` + // Peerings - The list of peerings. + Peerings *[]ExpressRouteCircuitPeering `json:"peerings,omitempty"` + // ServiceKey - The ServiceKey. + ServiceKey *string `json:"serviceKey,omitempty"` + // ServiceProviderNotes - The ServiceProviderNotes. + ServiceProviderNotes *string `json:"serviceProviderNotes,omitempty"` + // ServiceProviderProperties - The ServiceProviderProperties. + ServiceProviderProperties *ExpressRouteCircuitServiceProviderProperties `json:"serviceProviderProperties,omitempty"` + // ProvisioningState - Gets the provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// ExpressRouteCircuitRoutesTable the routes table associated with the ExpressRouteCircuit +type ExpressRouteCircuitRoutesTable struct { + // AddressPrefix - Gets AddressPrefix. + AddressPrefix *string `json:"addressPrefix,omitempty"` + // NextHopType - Gets NextHopType. Possible values include: 'RouteNextHopTypeVirtualNetworkGateway', 'RouteNextHopTypeVnetLocal', 'RouteNextHopTypeInternet', 'RouteNextHopTypeVirtualAppliance', 'RouteNextHopTypeNone' + NextHopType RouteNextHopType `json:"nextHopType,omitempty"` + // NextHopIP - Gets NextHopIP. + NextHopIP *string `json:"nextHopIP,omitempty"` + // AsPath - Gets AsPath. + AsPath *string `json:"asPath,omitempty"` +} + +// ExpressRouteCircuitsArpTableListResult response for ListArpTable associated with the Express Route Circuits API. +type ExpressRouteCircuitsArpTableListResult struct { + autorest.Response `json:"-"` + // Value - Gets list of the ARP table. + Value *[]ExpressRouteCircuitArpTable `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// ExpressRouteCircuitsArpTableListResultIterator provides access to a complete listing of +// ExpressRouteCircuitArpTable values. +type ExpressRouteCircuitsArpTableListResultIterator struct { + i int + page ExpressRouteCircuitsArpTableListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *ExpressRouteCircuitsArpTableListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter ExpressRouteCircuitsArpTableListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter ExpressRouteCircuitsArpTableListResultIterator) Response() ExpressRouteCircuitsArpTableListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter ExpressRouteCircuitsArpTableListResultIterator) Value() ExpressRouteCircuitArpTable { + if !iter.page.NotDone() { + return ExpressRouteCircuitArpTable{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (ercatlr ExpressRouteCircuitsArpTableListResult) IsEmpty() bool { + return ercatlr.Value == nil || len(*ercatlr.Value) == 0 +} + +// expressRouteCircuitsArpTableListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (ercatlr ExpressRouteCircuitsArpTableListResult) expressRouteCircuitsArpTableListResultPreparer() (*http.Request, error) { + if ercatlr.NextLink == nil || len(to.String(ercatlr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(ercatlr.NextLink))) +} + +// ExpressRouteCircuitsArpTableListResultPage contains a page of ExpressRouteCircuitArpTable values. +type ExpressRouteCircuitsArpTableListResultPage struct { + fn func(ExpressRouteCircuitsArpTableListResult) (ExpressRouteCircuitsArpTableListResult, error) + ercatlr ExpressRouteCircuitsArpTableListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *ExpressRouteCircuitsArpTableListResultPage) Next() error { + next, err := page.fn(page.ercatlr) + if err != nil { + return err + } + page.ercatlr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page ExpressRouteCircuitsArpTableListResultPage) NotDone() bool { + return !page.ercatlr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page ExpressRouteCircuitsArpTableListResultPage) Response() ExpressRouteCircuitsArpTableListResult { + return page.ercatlr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page ExpressRouteCircuitsArpTableListResultPage) Values() []ExpressRouteCircuitArpTable { + if page.ercatlr.IsEmpty() { + return nil + } + return *page.ercatlr.Value +} + +// ExpressRouteCircuitsCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a +// long-running operation. +type ExpressRouteCircuitsCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future ExpressRouteCircuitsCreateOrUpdateFuture) Result(client ExpressRouteCircuitsClient) (erc ExpressRouteCircuit, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return erc, azure.NewAsyncOpIncompleteError("network.ExpressRouteCircuitsCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + erc, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + erc, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// ExpressRouteCircuitsDeleteFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type ExpressRouteCircuitsDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future ExpressRouteCircuitsDeleteFuture) Result(client ExpressRouteCircuitsClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.ExpressRouteCircuitsDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.ExpressRouteCircuitsDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// ExpressRouteCircuitServiceProviderProperties contains ServiceProviderProperties in an ExpressRouteCircuit. +type ExpressRouteCircuitServiceProviderProperties struct { + // ServiceProviderName - The serviceProviderName. + ServiceProviderName *string `json:"serviceProviderName,omitempty"` + // PeeringLocation - The peering location. + PeeringLocation *string `json:"peeringLocation,omitempty"` + // BandwidthInMbps - The BandwidthInMbps. + BandwidthInMbps *int32 `json:"bandwidthInMbps,omitempty"` +} + +// ExpressRouteCircuitSku contains SKU in an ExpressRouteCircuit. +type ExpressRouteCircuitSku struct { + // Name - The name of the SKU. + Name *string `json:"name,omitempty"` + // Tier - The tier of the SKU. Possible values are 'Standard' and 'Premium'. Possible values include: 'ExpressRouteCircuitSkuTierStandard', 'ExpressRouteCircuitSkuTierPremium' + Tier ExpressRouteCircuitSkuTier `json:"tier,omitempty"` + // Family - The family of the SKU. Possible values are: 'UnlimitedData' and 'MeteredData'. Possible values include: 'UnlimitedData', 'MeteredData' + Family ExpressRouteCircuitSkuFamily `json:"family,omitempty"` +} + +// ExpressRouteCircuitsRoutesTableListResult response for ListRoutesTable associated with the Express Route +// Circuits API. +type ExpressRouteCircuitsRoutesTableListResult struct { + autorest.Response `json:"-"` + // Value - The list of routes table. + Value *[]ExpressRouteCircuitRoutesTable `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// ExpressRouteCircuitsRoutesTableListResultIterator provides access to a complete listing of +// ExpressRouteCircuitRoutesTable values. +type ExpressRouteCircuitsRoutesTableListResultIterator struct { + i int + page ExpressRouteCircuitsRoutesTableListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *ExpressRouteCircuitsRoutesTableListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter ExpressRouteCircuitsRoutesTableListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter ExpressRouteCircuitsRoutesTableListResultIterator) Response() ExpressRouteCircuitsRoutesTableListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter ExpressRouteCircuitsRoutesTableListResultIterator) Value() ExpressRouteCircuitRoutesTable { + if !iter.page.NotDone() { + return ExpressRouteCircuitRoutesTable{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (ercrtlr ExpressRouteCircuitsRoutesTableListResult) IsEmpty() bool { + return ercrtlr.Value == nil || len(*ercrtlr.Value) == 0 +} + +// expressRouteCircuitsRoutesTableListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (ercrtlr ExpressRouteCircuitsRoutesTableListResult) expressRouteCircuitsRoutesTableListResultPreparer() (*http.Request, error) { + if ercrtlr.NextLink == nil || len(to.String(ercrtlr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(ercrtlr.NextLink))) +} + +// ExpressRouteCircuitsRoutesTableListResultPage contains a page of ExpressRouteCircuitRoutesTable values. +type ExpressRouteCircuitsRoutesTableListResultPage struct { + fn func(ExpressRouteCircuitsRoutesTableListResult) (ExpressRouteCircuitsRoutesTableListResult, error) + ercrtlr ExpressRouteCircuitsRoutesTableListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *ExpressRouteCircuitsRoutesTableListResultPage) Next() error { + next, err := page.fn(page.ercrtlr) + if err != nil { + return err + } + page.ercrtlr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page ExpressRouteCircuitsRoutesTableListResultPage) NotDone() bool { + return !page.ercrtlr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page ExpressRouteCircuitsRoutesTableListResultPage) Response() ExpressRouteCircuitsRoutesTableListResult { + return page.ercrtlr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page ExpressRouteCircuitsRoutesTableListResultPage) Values() []ExpressRouteCircuitRoutesTable { + if page.ercrtlr.IsEmpty() { + return nil + } + return *page.ercrtlr.Value +} + +// ExpressRouteCircuitsStatsListResult response for ListStats from Express Route Circuits Api service call +type ExpressRouteCircuitsStatsListResult struct { + autorest.Response `json:"-"` + // Value - Gets List of Stats + Value *[]ExpressRouteCircuitStats `json:"value,omitempty"` + // NextLink - Gets the URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// ExpressRouteCircuitsStatsListResultIterator provides access to a complete listing of ExpressRouteCircuitStats +// values. +type ExpressRouteCircuitsStatsListResultIterator struct { + i int + page ExpressRouteCircuitsStatsListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *ExpressRouteCircuitsStatsListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter ExpressRouteCircuitsStatsListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter ExpressRouteCircuitsStatsListResultIterator) Response() ExpressRouteCircuitsStatsListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter ExpressRouteCircuitsStatsListResultIterator) Value() ExpressRouteCircuitStats { + if !iter.page.NotDone() { + return ExpressRouteCircuitStats{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (ercslr ExpressRouteCircuitsStatsListResult) IsEmpty() bool { + return ercslr.Value == nil || len(*ercslr.Value) == 0 +} + +// expressRouteCircuitsStatsListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (ercslr ExpressRouteCircuitsStatsListResult) expressRouteCircuitsStatsListResultPreparer() (*http.Request, error) { + if ercslr.NextLink == nil || len(to.String(ercslr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(ercslr.NextLink))) +} + +// ExpressRouteCircuitsStatsListResultPage contains a page of ExpressRouteCircuitStats values. +type ExpressRouteCircuitsStatsListResultPage struct { + fn func(ExpressRouteCircuitsStatsListResult) (ExpressRouteCircuitsStatsListResult, error) + ercslr ExpressRouteCircuitsStatsListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *ExpressRouteCircuitsStatsListResultPage) Next() error { + next, err := page.fn(page.ercslr) + if err != nil { + return err + } + page.ercslr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page ExpressRouteCircuitsStatsListResultPage) NotDone() bool { + return !page.ercslr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page ExpressRouteCircuitsStatsListResultPage) Response() ExpressRouteCircuitsStatsListResult { + return page.ercslr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page ExpressRouteCircuitsStatsListResultPage) Values() []ExpressRouteCircuitStats { + if page.ercslr.IsEmpty() { + return nil + } + return *page.ercslr.Value +} + +// ExpressRouteCircuitStats contains stats associated with the peering. +type ExpressRouteCircuitStats struct { + // BytesIn - Gets BytesIn of the peering. + BytesIn *int32 `json:"bytesIn,omitempty"` + // BytesOut - Gets BytesOut of the peering. + BytesOut *int32 `json:"bytesOut,omitempty"` +} + +// ExpressRouteServiceProvider a ExpressRouteResourceProvider object. +type ExpressRouteServiceProvider struct { + *ExpressRouteServiceProviderPropertiesFormat `json:"properties,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for ExpressRouteServiceProvider. +func (ersp ExpressRouteServiceProvider) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if ersp.ExpressRouteServiceProviderPropertiesFormat != nil { + objectMap["properties"] = ersp.ExpressRouteServiceProviderPropertiesFormat + } + if ersp.ID != nil { + objectMap["id"] = ersp.ID + } + if ersp.Name != nil { + objectMap["name"] = ersp.Name + } + if ersp.Type != nil { + objectMap["type"] = ersp.Type + } + if ersp.Location != nil { + objectMap["location"] = ersp.Location + } + if ersp.Tags != nil { + objectMap["tags"] = ersp.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for ExpressRouteServiceProvider struct. +func (ersp *ExpressRouteServiceProvider) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var expressRouteServiceProviderPropertiesFormat ExpressRouteServiceProviderPropertiesFormat + err = json.Unmarshal(*v, &expressRouteServiceProviderPropertiesFormat) + if err != nil { + return err + } + ersp.ExpressRouteServiceProviderPropertiesFormat = &expressRouteServiceProviderPropertiesFormat + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + ersp.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + ersp.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + ersp.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + ersp.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + ersp.Tags = tags + } + } + } + + return nil +} + +// ExpressRouteServiceProviderBandwidthsOffered contains bandwidths offered in ExpressRouteServiceProvider +// resources. +type ExpressRouteServiceProviderBandwidthsOffered struct { + // OfferName - The OfferName. + OfferName *string `json:"offerName,omitempty"` + // ValueInMbps - The ValueInMbps. + ValueInMbps *int32 `json:"valueInMbps,omitempty"` +} + +// ExpressRouteServiceProviderListResult response for the ListExpressRouteServiceProvider API service call. +type ExpressRouteServiceProviderListResult struct { + autorest.Response `json:"-"` + // Value - A list of ExpressRouteResourceProvider resources. + Value *[]ExpressRouteServiceProvider `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// ExpressRouteServiceProviderListResultIterator provides access to a complete listing of +// ExpressRouteServiceProvider values. +type ExpressRouteServiceProviderListResultIterator struct { + i int + page ExpressRouteServiceProviderListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *ExpressRouteServiceProviderListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter ExpressRouteServiceProviderListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter ExpressRouteServiceProviderListResultIterator) Response() ExpressRouteServiceProviderListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter ExpressRouteServiceProviderListResultIterator) Value() ExpressRouteServiceProvider { + if !iter.page.NotDone() { + return ExpressRouteServiceProvider{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (ersplr ExpressRouteServiceProviderListResult) IsEmpty() bool { + return ersplr.Value == nil || len(*ersplr.Value) == 0 +} + +// expressRouteServiceProviderListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (ersplr ExpressRouteServiceProviderListResult) expressRouteServiceProviderListResultPreparer() (*http.Request, error) { + if ersplr.NextLink == nil || len(to.String(ersplr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(ersplr.NextLink))) +} + +// ExpressRouteServiceProviderListResultPage contains a page of ExpressRouteServiceProvider values. +type ExpressRouteServiceProviderListResultPage struct { + fn func(ExpressRouteServiceProviderListResult) (ExpressRouteServiceProviderListResult, error) + ersplr ExpressRouteServiceProviderListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *ExpressRouteServiceProviderListResultPage) Next() error { + next, err := page.fn(page.ersplr) + if err != nil { + return err + } + page.ersplr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page ExpressRouteServiceProviderListResultPage) NotDone() bool { + return !page.ersplr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page ExpressRouteServiceProviderListResultPage) Response() ExpressRouteServiceProviderListResult { + return page.ersplr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page ExpressRouteServiceProviderListResultPage) Values() []ExpressRouteServiceProvider { + if page.ersplr.IsEmpty() { + return nil + } + return *page.ersplr.Value +} + +// ExpressRouteServiceProviderPropertiesFormat properties of ExpressRouteServiceProvider. +type ExpressRouteServiceProviderPropertiesFormat struct { + // PeeringLocations - Get a list of peering locations. + PeeringLocations *[]string `json:"peeringLocations,omitempty"` + // BandwidthsOffered - Gets bandwidths offered. + BandwidthsOffered *[]ExpressRouteServiceProviderBandwidthsOffered `json:"bandwidthsOffered,omitempty"` + // ProvisioningState - Gets the provisioning state of the resource. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// FrontendIPConfiguration frontend IP address of the load balancer. +type FrontendIPConfiguration struct { + *FrontendIPConfigurationPropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for FrontendIPConfiguration. +func (fic FrontendIPConfiguration) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if fic.FrontendIPConfigurationPropertiesFormat != nil { + objectMap["properties"] = fic.FrontendIPConfigurationPropertiesFormat + } + if fic.Name != nil { + objectMap["name"] = fic.Name + } + if fic.Etag != nil { + objectMap["etag"] = fic.Etag + } + if fic.ID != nil { + objectMap["id"] = fic.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for FrontendIPConfiguration struct. +func (fic *FrontendIPConfiguration) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var frontendIPConfigurationPropertiesFormat FrontendIPConfigurationPropertiesFormat + err = json.Unmarshal(*v, &frontendIPConfigurationPropertiesFormat) + if err != nil { + return err + } + fic.FrontendIPConfigurationPropertiesFormat = &frontendIPConfigurationPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + fic.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + fic.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + fic.ID = &ID + } + } + } + + return nil +} + +// FrontendIPConfigurationPropertiesFormat properties of Frontend IP Configuration of the load balancer. +type FrontendIPConfigurationPropertiesFormat struct { + // InboundNatRules - Read only. Inbound rules URIs that use this frontend IP. + InboundNatRules *[]SubResource `json:"inboundNatRules,omitempty"` + // InboundNatPools - Read only. Inbound pools URIs that use this frontend IP. + InboundNatPools *[]SubResource `json:"inboundNatPools,omitempty"` + // OutboundNatRules - Read only. Outbound rules URIs that use this frontend IP. + OutboundNatRules *[]SubResource `json:"outboundNatRules,omitempty"` + // LoadBalancingRules - Gets load balancing rules URIs that use this frontend IP. + LoadBalancingRules *[]SubResource `json:"loadBalancingRules,omitempty"` + // PrivateIPAddress - The private IP address of the IP configuration. + PrivateIPAddress *string `json:"privateIPAddress,omitempty"` + // PrivateIPAllocationMethod - The Private IP allocation method. Possible values are: 'Static' and 'Dynamic'. Possible values include: 'Static', 'Dynamic' + PrivateIPAllocationMethod IPAllocationMethod `json:"privateIPAllocationMethod,omitempty"` + // Subnet - The reference of the subnet resource. + Subnet *Subnet `json:"subnet,omitempty"` + // PublicIPAddress - The reference of the Public IP resource. + PublicIPAddress *PublicIPAddress `json:"publicIPAddress,omitempty"` + // ProvisioningState - Gets the provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// InboundNatPool inbound NAT pool of the load balancer. +type InboundNatPool struct { + *InboundNatPoolPropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for InboundNatPool. +func (inp InboundNatPool) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if inp.InboundNatPoolPropertiesFormat != nil { + objectMap["properties"] = inp.InboundNatPoolPropertiesFormat + } + if inp.Name != nil { + objectMap["name"] = inp.Name + } + if inp.Etag != nil { + objectMap["etag"] = inp.Etag + } + if inp.ID != nil { + objectMap["id"] = inp.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for InboundNatPool struct. +func (inp *InboundNatPool) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var inboundNatPoolPropertiesFormat InboundNatPoolPropertiesFormat + err = json.Unmarshal(*v, &inboundNatPoolPropertiesFormat) + if err != nil { + return err + } + inp.InboundNatPoolPropertiesFormat = &inboundNatPoolPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + inp.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + inp.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + inp.ID = &ID + } + } + } + + return nil +} + +// InboundNatPoolPropertiesFormat properties of Inbound NAT pool. +type InboundNatPoolPropertiesFormat struct { + // FrontendIPConfiguration - A reference to frontend IP addresses. + FrontendIPConfiguration *SubResource `json:"frontendIPConfiguration,omitempty"` + // Protocol - The transport protocol for the endpoint. Possible values are: 'Udp' or 'Tcp'. Possible values include: 'TransportProtocolUDP', 'TransportProtocolTCP' + Protocol TransportProtocol `json:"protocol,omitempty"` + // FrontendPortRangeStart - The first port number in the range of external ports that will be used to provide Inbound Nat to NICs associated with a load balancer. Acceptable values range between 1 and 65534. + FrontendPortRangeStart *int32 `json:"frontendPortRangeStart,omitempty"` + // FrontendPortRangeEnd - The last port number in the range of external ports that will be used to provide Inbound Nat to NICs associated with a load balancer. Acceptable values range between 1 and 65535. + FrontendPortRangeEnd *int32 `json:"frontendPortRangeEnd,omitempty"` + // BackendPort - The port used for internal connections on the endpoint. Acceptable values are between 1 and 65535. + BackendPort *int32 `json:"backendPort,omitempty"` + // ProvisioningState - Gets the provisioning state of the PublicIP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// InboundNatRule inbound NAT rule of the load balancer. +type InboundNatRule struct { + *InboundNatRulePropertiesFormat `json:"properties,omitempty"` + // Name - Gets name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for InboundNatRule. +func (inr InboundNatRule) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if inr.InboundNatRulePropertiesFormat != nil { + objectMap["properties"] = inr.InboundNatRulePropertiesFormat + } + if inr.Name != nil { + objectMap["name"] = inr.Name + } + if inr.Etag != nil { + objectMap["etag"] = inr.Etag + } + if inr.ID != nil { + objectMap["id"] = inr.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for InboundNatRule struct. +func (inr *InboundNatRule) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var inboundNatRulePropertiesFormat InboundNatRulePropertiesFormat + err = json.Unmarshal(*v, &inboundNatRulePropertiesFormat) + if err != nil { + return err + } + inr.InboundNatRulePropertiesFormat = &inboundNatRulePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + inr.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + inr.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + inr.ID = &ID + } + } + } + + return nil +} + +// InboundNatRulePropertiesFormat properties of the inbound NAT rule. +type InboundNatRulePropertiesFormat struct { + // FrontendIPConfiguration - A reference to frontend IP addresses. + FrontendIPConfiguration *SubResource `json:"frontendIPConfiguration,omitempty"` + // BackendIPConfiguration - A reference to a private IP address defined on a network interface of a VM. Traffic sent to the frontend port of each of the frontend IP configurations is forwarded to the backed IP. + BackendIPConfiguration *InterfaceIPConfiguration `json:"backendIPConfiguration,omitempty"` + // Protocol - The transport protocol for the endpoint. Possible values are: 'Udp' or 'Tcp'. Possible values include: 'TransportProtocolUDP', 'TransportProtocolTCP' + Protocol TransportProtocol `json:"protocol,omitempty"` + // FrontendPort - The port for the external endpoint. Port numbers for each rule must be unique within the Load Balancer. Acceptable values range from 1 to 65534. + FrontendPort *int32 `json:"frontendPort,omitempty"` + // BackendPort - The port used for the internal endpoint. Acceptable values range from 1 to 65535. + BackendPort *int32 `json:"backendPort,omitempty"` + // IdleTimeoutInMinutes - The timeout for the TCP idle connection. The value can be set between 4 and 30 minutes. The default value is 4 minutes. This element is only used when the protocol is set to TCP. + IdleTimeoutInMinutes *int32 `json:"idleTimeoutInMinutes,omitempty"` + // EnableFloatingIP - Configures a virtual machine's endpoint for the floating IP capability required to configure a SQL AlwaysOn Availability Group. This setting is required when using the SQL AlwaysOn Availability Groups in SQL server. This setting can't be changed after you create the endpoint. + EnableFloatingIP *bool `json:"enableFloatingIP,omitempty"` + // ProvisioningState - Gets the provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// Interface a network interface in a resource group. +type Interface struct { + autorest.Response `json:"-"` + *InterfacePropertiesFormat `json:"properties,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for Interface. +func (i Interface) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if i.InterfacePropertiesFormat != nil { + objectMap["properties"] = i.InterfacePropertiesFormat + } + if i.Etag != nil { + objectMap["etag"] = i.Etag + } + if i.ID != nil { + objectMap["id"] = i.ID + } + if i.Name != nil { + objectMap["name"] = i.Name + } + if i.Type != nil { + objectMap["type"] = i.Type + } + if i.Location != nil { + objectMap["location"] = i.Location + } + if i.Tags != nil { + objectMap["tags"] = i.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for Interface struct. +func (i *Interface) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var interfacePropertiesFormat InterfacePropertiesFormat + err = json.Unmarshal(*v, &interfacePropertiesFormat) + if err != nil { + return err + } + i.InterfacePropertiesFormat = &interfacePropertiesFormat + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + i.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + i.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + i.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + i.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + i.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + i.Tags = tags + } + } + } + + return nil +} + +// InterfaceDNSSettings DNS settings of a network interface. +type InterfaceDNSSettings struct { + // DNSServers - List of DNS servers IP addresses. Use 'AzureProvidedDNS' to switch to azure provided DNS resolution. 'AzureProvidedDNS' value cannot be combined with other IPs, it must be the only value in dnsServers collection. + DNSServers *[]string `json:"dnsServers,omitempty"` + // AppliedDNSServers - If the VM that uses this NIC is part of an Availability Set, then this list will have the union of all DNS servers from all NICs that are part of the Availability Set. This property is what is configured on each of those VMs. + AppliedDNSServers *[]string `json:"appliedDnsServers,omitempty"` + // InternalDNSNameLabel - Relative DNS name for this NIC used for internal communications between VMs in the same virtual network. + InternalDNSNameLabel *string `json:"internalDnsNameLabel,omitempty"` + // InternalFqdn - Fully qualified DNS name supporting internal communications between VMs in the same virtual network. + InternalFqdn *string `json:"internalFqdn,omitempty"` +} + +// InterfaceIPConfiguration iPConfiguration in a network interface. +type InterfaceIPConfiguration struct { + *InterfaceIPConfigurationPropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for InterfaceIPConfiguration. +func (iic InterfaceIPConfiguration) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if iic.InterfaceIPConfigurationPropertiesFormat != nil { + objectMap["properties"] = iic.InterfaceIPConfigurationPropertiesFormat + } + if iic.Name != nil { + objectMap["name"] = iic.Name + } + if iic.Etag != nil { + objectMap["etag"] = iic.Etag + } + if iic.ID != nil { + objectMap["id"] = iic.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for InterfaceIPConfiguration struct. +func (iic *InterfaceIPConfiguration) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var interfaceIPConfigurationPropertiesFormat InterfaceIPConfigurationPropertiesFormat + err = json.Unmarshal(*v, &interfaceIPConfigurationPropertiesFormat) + if err != nil { + return err + } + iic.InterfaceIPConfigurationPropertiesFormat = &interfaceIPConfigurationPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + iic.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + iic.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + iic.ID = &ID + } + } + } + + return nil +} + +// InterfaceIPConfigurationPropertiesFormat properties of IP configuration. +type InterfaceIPConfigurationPropertiesFormat struct { + // LoadBalancerBackendAddressPools - The reference of LoadBalancerBackendAddressPool resource. + LoadBalancerBackendAddressPools *[]BackendAddressPool `json:"loadBalancerBackendAddressPools,omitempty"` + // LoadBalancerInboundNatRules - A list of references of LoadBalancerInboundNatRules. + LoadBalancerInboundNatRules *[]InboundNatRule `json:"loadBalancerInboundNatRules,omitempty"` + PrivateIPAddress *string `json:"privateIPAddress,omitempty"` + // PrivateIPAllocationMethod - Defines how a private IP address is assigned. Possible values are: 'Static' and 'Dynamic'. Possible values include: 'Static', 'Dynamic' + PrivateIPAllocationMethod IPAllocationMethod `json:"privateIPAllocationMethod,omitempty"` + Subnet *Subnet `json:"subnet,omitempty"` + PublicIPAddress *PublicIPAddress `json:"publicIPAddress,omitempty"` + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// InterfaceListResult response for the ListNetworkInterface API service call. +type InterfaceListResult struct { + autorest.Response `json:"-"` + // Value - A list of network interfaces in a resource group. + Value *[]Interface `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// InterfaceListResultIterator provides access to a complete listing of Interface values. +type InterfaceListResultIterator struct { + i int + page InterfaceListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *InterfaceListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter InterfaceListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter InterfaceListResultIterator) Response() InterfaceListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter InterfaceListResultIterator) Value() Interface { + if !iter.page.NotDone() { + return Interface{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (ilr InterfaceListResult) IsEmpty() bool { + return ilr.Value == nil || len(*ilr.Value) == 0 +} + +// interfaceListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (ilr InterfaceListResult) interfaceListResultPreparer() (*http.Request, error) { + if ilr.NextLink == nil || len(to.String(ilr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(ilr.NextLink))) +} + +// InterfaceListResultPage contains a page of Interface values. +type InterfaceListResultPage struct { + fn func(InterfaceListResult) (InterfaceListResult, error) + ilr InterfaceListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *InterfaceListResultPage) Next() error { + next, err := page.fn(page.ilr) + if err != nil { + return err + } + page.ilr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page InterfaceListResultPage) NotDone() bool { + return !page.ilr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page InterfaceListResultPage) Response() InterfaceListResult { + return page.ilr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page InterfaceListResultPage) Values() []Interface { + if page.ilr.IsEmpty() { + return nil + } + return *page.ilr.Value +} + +// InterfacePropertiesFormat networkInterface properties. +type InterfacePropertiesFormat struct { + // VirtualMachine - The reference of a virtual machine. + VirtualMachine *SubResource `json:"virtualMachine,omitempty"` + // NetworkSecurityGroup - The reference of the NetworkSecurityGroup resource. + NetworkSecurityGroup *SecurityGroup `json:"networkSecurityGroup,omitempty"` + // IPConfigurations - A list of IPConfigurations of the network interface. + IPConfigurations *[]InterfaceIPConfiguration `json:"ipConfigurations,omitempty"` + // DNSSettings - The DNS settings in network interface. + DNSSettings *InterfaceDNSSettings `json:"dnsSettings,omitempty"` + // MacAddress - The MAC address of the network interface. + MacAddress *string `json:"macAddress,omitempty"` + // Primary - Gets whether this is a primary network interface on a virtual machine. + Primary *bool `json:"primary,omitempty"` + // EnableIPForwarding - Indicates whether IP forwarding is enabled on this network interface. + EnableIPForwarding *bool `json:"enableIPForwarding,omitempty"` + // ResourceGUID - The resource GUID property of the network interface resource. + ResourceGUID *string `json:"resourceGuid,omitempty"` + // ProvisioningState - The provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// InterfacesCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type InterfacesCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future InterfacesCreateOrUpdateFuture) Result(client InterfacesClient) (i Interface, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return i, azure.NewAsyncOpIncompleteError("network.InterfacesCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + i, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + i, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// InterfacesDeleteFuture an abstraction for monitoring and retrieving the results of a long-running operation. +type InterfacesDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future InterfacesDeleteFuture) Result(client InterfacesClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.InterfacesDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.InterfacesDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// IPConfiguration iPConfiguration +type IPConfiguration struct { + *IPConfigurationPropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for IPConfiguration. +func (ic IPConfiguration) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if ic.IPConfigurationPropertiesFormat != nil { + objectMap["properties"] = ic.IPConfigurationPropertiesFormat + } + if ic.Name != nil { + objectMap["name"] = ic.Name + } + if ic.Etag != nil { + objectMap["etag"] = ic.Etag + } + if ic.ID != nil { + objectMap["id"] = ic.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for IPConfiguration struct. +func (ic *IPConfiguration) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var IPConfigurationPropertiesFormat IPConfigurationPropertiesFormat + err = json.Unmarshal(*v, &IPConfigurationPropertiesFormat) + if err != nil { + return err + } + ic.IPConfigurationPropertiesFormat = &IPConfigurationPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + ic.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + ic.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + ic.ID = &ID + } + } + } + + return nil +} + +// IPConfigurationPropertiesFormat properties of IP configuration. +type IPConfigurationPropertiesFormat struct { + // PrivateIPAddress - The private IP address of the IP configuration. + PrivateIPAddress *string `json:"privateIPAddress,omitempty"` + // PrivateIPAllocationMethod - The private IP allocation method. Possible values are 'Static' and 'Dynamic'. Possible values include: 'Static', 'Dynamic' + PrivateIPAllocationMethod IPAllocationMethod `json:"privateIPAllocationMethod,omitempty"` + // Subnet - The reference of the subnet resource. + Subnet *Subnet `json:"subnet,omitempty"` + // PublicIPAddress - The reference of the public IP resource. + PublicIPAddress *PublicIPAddress `json:"publicIPAddress,omitempty"` + // ProvisioningState - Gets the provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// LoadBalancer loadBalancer resource +type LoadBalancer struct { + autorest.Response `json:"-"` + *LoadBalancerPropertiesFormat `json:"properties,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for LoadBalancer. +func (lb LoadBalancer) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if lb.LoadBalancerPropertiesFormat != nil { + objectMap["properties"] = lb.LoadBalancerPropertiesFormat + } + if lb.Etag != nil { + objectMap["etag"] = lb.Etag + } + if lb.ID != nil { + objectMap["id"] = lb.ID + } + if lb.Name != nil { + objectMap["name"] = lb.Name + } + if lb.Type != nil { + objectMap["type"] = lb.Type + } + if lb.Location != nil { + objectMap["location"] = lb.Location + } + if lb.Tags != nil { + objectMap["tags"] = lb.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for LoadBalancer struct. +func (lb *LoadBalancer) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var loadBalancerPropertiesFormat LoadBalancerPropertiesFormat + err = json.Unmarshal(*v, &loadBalancerPropertiesFormat) + if err != nil { + return err + } + lb.LoadBalancerPropertiesFormat = &loadBalancerPropertiesFormat + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + lb.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + lb.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + lb.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + lb.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + lb.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + lb.Tags = tags + } + } + } + + return nil +} + +// LoadBalancerListResult response for ListLoadBalancers API service call. +type LoadBalancerListResult struct { + autorest.Response `json:"-"` + // Value - A list of load balancers in a resource group. + Value *[]LoadBalancer `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// LoadBalancerListResultIterator provides access to a complete listing of LoadBalancer values. +type LoadBalancerListResultIterator struct { + i int + page LoadBalancerListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *LoadBalancerListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter LoadBalancerListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter LoadBalancerListResultIterator) Response() LoadBalancerListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter LoadBalancerListResultIterator) Value() LoadBalancer { + if !iter.page.NotDone() { + return LoadBalancer{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (lblr LoadBalancerListResult) IsEmpty() bool { + return lblr.Value == nil || len(*lblr.Value) == 0 +} + +// loadBalancerListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (lblr LoadBalancerListResult) loadBalancerListResultPreparer() (*http.Request, error) { + if lblr.NextLink == nil || len(to.String(lblr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(lblr.NextLink))) +} + +// LoadBalancerListResultPage contains a page of LoadBalancer values. +type LoadBalancerListResultPage struct { + fn func(LoadBalancerListResult) (LoadBalancerListResult, error) + lblr LoadBalancerListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *LoadBalancerListResultPage) Next() error { + next, err := page.fn(page.lblr) + if err != nil { + return err + } + page.lblr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page LoadBalancerListResultPage) NotDone() bool { + return !page.lblr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page LoadBalancerListResultPage) Response() LoadBalancerListResult { + return page.lblr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page LoadBalancerListResultPage) Values() []LoadBalancer { + if page.lblr.IsEmpty() { + return nil + } + return *page.lblr.Value +} + +// LoadBalancerPropertiesFormat properties of the load balancer. +type LoadBalancerPropertiesFormat struct { + // FrontendIPConfigurations - Object representing the frontend IPs to be used for the load balancer + FrontendIPConfigurations *[]FrontendIPConfiguration `json:"frontendIPConfigurations,omitempty"` + // BackendAddressPools - Collection of backend address pools used by a load balancer + BackendAddressPools *[]BackendAddressPool `json:"backendAddressPools,omitempty"` + // LoadBalancingRules - Object collection representing the load balancing rules Gets the provisioning + LoadBalancingRules *[]LoadBalancingRule `json:"loadBalancingRules,omitempty"` + // Probes - Collection of probe objects used in the load balancer + Probes *[]Probe `json:"probes,omitempty"` + // InboundNatRules - Collection of inbound NAT Rules used by a load balancer. Defining inbound NAT rules on your load balancer is mutually exclusive with defining an inbound NAT pool. Inbound NAT pools are referenced from virtual machine scale sets. NICs that are associated with individual virtual machines cannot reference an Inbound NAT pool. They have to reference individual inbound NAT rules. + InboundNatRules *[]InboundNatRule `json:"inboundNatRules,omitempty"` + // InboundNatPools - Defines an external port range for inbound NAT to a single backend port on NICs associated with a load balancer. Inbound NAT rules are created automatically for each NIC associated with the Load Balancer using an external port from this range. Defining an Inbound NAT pool on your Load Balancer is mutually exclusive with defining inbound Nat rules. Inbound NAT pools are referenced from virtual machine scale sets. NICs that are associated with individual virtual machines cannot reference an inbound NAT pool. They have to reference individual inbound NAT rules. + InboundNatPools *[]InboundNatPool `json:"inboundNatPools,omitempty"` + // OutboundNatRules - The outbound NAT rules. + OutboundNatRules *[]OutboundNatRule `json:"outboundNatRules,omitempty"` + // ResourceGUID - The resource GUID property of the load balancer resource. + ResourceGUID *string `json:"resourceGuid,omitempty"` + // ProvisioningState - Gets the provisioning state of the PublicIP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// LoadBalancersCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type LoadBalancersCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future LoadBalancersCreateOrUpdateFuture) Result(client LoadBalancersClient) (lb LoadBalancer, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return lb, azure.NewAsyncOpIncompleteError("network.LoadBalancersCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + lb, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + lb, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// LoadBalancersDeleteFuture an abstraction for monitoring and retrieving the results of a long-running operation. +type LoadBalancersDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future LoadBalancersDeleteFuture) Result(client LoadBalancersClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.LoadBalancersDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LoadBalancersDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// LoadBalancingRule a loag balancing rule for a load balancer. +type LoadBalancingRule struct { + *LoadBalancingRulePropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for LoadBalancingRule. +func (lbr LoadBalancingRule) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if lbr.LoadBalancingRulePropertiesFormat != nil { + objectMap["properties"] = lbr.LoadBalancingRulePropertiesFormat + } + if lbr.Name != nil { + objectMap["name"] = lbr.Name + } + if lbr.Etag != nil { + objectMap["etag"] = lbr.Etag + } + if lbr.ID != nil { + objectMap["id"] = lbr.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for LoadBalancingRule struct. +func (lbr *LoadBalancingRule) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var loadBalancingRulePropertiesFormat LoadBalancingRulePropertiesFormat + err = json.Unmarshal(*v, &loadBalancingRulePropertiesFormat) + if err != nil { + return err + } + lbr.LoadBalancingRulePropertiesFormat = &loadBalancingRulePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + lbr.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + lbr.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + lbr.ID = &ID + } + } + } + + return nil +} + +// LoadBalancingRulePropertiesFormat properties of the load balancer. +type LoadBalancingRulePropertiesFormat struct { + // FrontendIPConfiguration - A reference to frontend IP addresses. + FrontendIPConfiguration *SubResource `json:"frontendIPConfiguration,omitempty"` + // BackendAddressPool - A reference to a pool of DIPs. Inbound traffic is randomly load balanced across IPs in the backend IPs. + BackendAddressPool *SubResource `json:"backendAddressPool,omitempty"` + // Probe - The reference of the load balancer probe used by the load balancing rule. + Probe *SubResource `json:"probe,omitempty"` + // Protocol - The transport protocol for the external endpoint. Possible values are 'Udp' or 'Tcp'. Possible values include: 'TransportProtocolUDP', 'TransportProtocolTCP' + Protocol TransportProtocol `json:"protocol,omitempty"` + // LoadDistribution - The load distribution policy for this rule. Possible values are 'Default', 'SourceIP', and 'SourceIPProtocol'. Possible values include: 'Default', 'SourceIP', 'SourceIPProtocol' + LoadDistribution LoadDistribution `json:"loadDistribution,omitempty"` + // FrontendPort - The port for the external endpoint. Port numbers for each rule must be unique within the Load Balancer. Acceptable values are between 1 and 65534. + FrontendPort *int32 `json:"frontendPort,omitempty"` + // BackendPort - The port used for internal connections on the endpoint. Acceptable values are between 1 and 65535. + BackendPort *int32 `json:"backendPort,omitempty"` + // IdleTimeoutInMinutes - The timeout for the TCP idle connection. The value can be set between 4 and 30 minutes. The default value is 4 minutes. This element is only used when the protocol is set to TCP. + IdleTimeoutInMinutes *int32 `json:"idleTimeoutInMinutes,omitempty"` + // EnableFloatingIP - Configures a virtual machine's endpoint for the floating IP capability required to configure a SQL AlwaysOn Availability Group. This setting is required when using the SQL AlwaysOn Availability Groups in SQL server. This setting can't be changed after you create the endpoint. + EnableFloatingIP *bool `json:"enableFloatingIP,omitempty"` + // ProvisioningState - Gets the provisioning state of the PublicIP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// LocalNetworkGateway a common class for general resource information +type LocalNetworkGateway struct { + autorest.Response `json:"-"` + *LocalNetworkGatewayPropertiesFormat `json:"properties,omitempty"` + // Etag - Gets a unique read-only string that changes whenever the resource is updated + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for LocalNetworkGateway. +func (lng LocalNetworkGateway) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if lng.LocalNetworkGatewayPropertiesFormat != nil { + objectMap["properties"] = lng.LocalNetworkGatewayPropertiesFormat + } + if lng.Etag != nil { + objectMap["etag"] = lng.Etag + } + if lng.ID != nil { + objectMap["id"] = lng.ID + } + if lng.Name != nil { + objectMap["name"] = lng.Name + } + if lng.Type != nil { + objectMap["type"] = lng.Type + } + if lng.Location != nil { + objectMap["location"] = lng.Location + } + if lng.Tags != nil { + objectMap["tags"] = lng.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for LocalNetworkGateway struct. +func (lng *LocalNetworkGateway) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var localNetworkGatewayPropertiesFormat LocalNetworkGatewayPropertiesFormat + err = json.Unmarshal(*v, &localNetworkGatewayPropertiesFormat) + if err != nil { + return err + } + lng.LocalNetworkGatewayPropertiesFormat = &localNetworkGatewayPropertiesFormat + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + lng.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + lng.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + lng.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + lng.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + lng.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + lng.Tags = tags + } + } + } + + return nil +} + +// LocalNetworkGatewayListResult response for ListLocalNetworkGateways API service call. +type LocalNetworkGatewayListResult struct { + autorest.Response `json:"-"` + // Value - A list of local network gateways that exists in a resource group. + Value *[]LocalNetworkGateway `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// LocalNetworkGatewayListResultIterator provides access to a complete listing of LocalNetworkGateway values. +type LocalNetworkGatewayListResultIterator struct { + i int + page LocalNetworkGatewayListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *LocalNetworkGatewayListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter LocalNetworkGatewayListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter LocalNetworkGatewayListResultIterator) Response() LocalNetworkGatewayListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter LocalNetworkGatewayListResultIterator) Value() LocalNetworkGateway { + if !iter.page.NotDone() { + return LocalNetworkGateway{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (lnglr LocalNetworkGatewayListResult) IsEmpty() bool { + return lnglr.Value == nil || len(*lnglr.Value) == 0 +} + +// localNetworkGatewayListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (lnglr LocalNetworkGatewayListResult) localNetworkGatewayListResultPreparer() (*http.Request, error) { + if lnglr.NextLink == nil || len(to.String(lnglr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(lnglr.NextLink))) +} + +// LocalNetworkGatewayListResultPage contains a page of LocalNetworkGateway values. +type LocalNetworkGatewayListResultPage struct { + fn func(LocalNetworkGatewayListResult) (LocalNetworkGatewayListResult, error) + lnglr LocalNetworkGatewayListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *LocalNetworkGatewayListResultPage) Next() error { + next, err := page.fn(page.lnglr) + if err != nil { + return err + } + page.lnglr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page LocalNetworkGatewayListResultPage) NotDone() bool { + return !page.lnglr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page LocalNetworkGatewayListResultPage) Response() LocalNetworkGatewayListResult { + return page.lnglr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page LocalNetworkGatewayListResultPage) Values() []LocalNetworkGateway { + if page.lnglr.IsEmpty() { + return nil + } + return *page.lnglr.Value +} + +// LocalNetworkGatewayPropertiesFormat localNetworkGateway properties +type LocalNetworkGatewayPropertiesFormat struct { + // LocalNetworkAddressSpace - Local network site address space. + LocalNetworkAddressSpace *AddressSpace `json:"localNetworkAddressSpace,omitempty"` + // GatewayIPAddress - IP address of local network gateway. + GatewayIPAddress *string `json:"gatewayIpAddress,omitempty"` + // BgpSettings - Local network gateway's BGP speaker settings. + BgpSettings *BgpSettings `json:"bgpSettings,omitempty"` + // ResourceGUID - The resource GUID property of the LocalNetworkGateway resource. + ResourceGUID *string `json:"resourceGuid,omitempty"` + // ProvisioningState - Gets or sets Provisioning state of the LocalNetworkGateway resource Updating/Deleting/Failed + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// LocalNetworkGatewaysCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a +// long-running operation. +type LocalNetworkGatewaysCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future LocalNetworkGatewaysCreateOrUpdateFuture) Result(client LocalNetworkGatewaysClient) (lng LocalNetworkGateway, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return lng, azure.NewAsyncOpIncompleteError("network.LocalNetworkGatewaysCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + lng, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + lng, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// LocalNetworkGatewaysDeleteFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type LocalNetworkGatewaysDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future LocalNetworkGatewaysDeleteFuture) Result(client LocalNetworkGatewaysClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.LocalNetworkGatewaysDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.LocalNetworkGatewaysDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// OutboundNatRule outbound NAT pool of the load balancer. +type OutboundNatRule struct { + *OutboundNatRulePropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for OutboundNatRule. +func (onr OutboundNatRule) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if onr.OutboundNatRulePropertiesFormat != nil { + objectMap["properties"] = onr.OutboundNatRulePropertiesFormat + } + if onr.Name != nil { + objectMap["name"] = onr.Name + } + if onr.Etag != nil { + objectMap["etag"] = onr.Etag + } + if onr.ID != nil { + objectMap["id"] = onr.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for OutboundNatRule struct. +func (onr *OutboundNatRule) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var outboundNatRulePropertiesFormat OutboundNatRulePropertiesFormat + err = json.Unmarshal(*v, &outboundNatRulePropertiesFormat) + if err != nil { + return err + } + onr.OutboundNatRulePropertiesFormat = &outboundNatRulePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + onr.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + onr.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + onr.ID = &ID + } + } + } + + return nil +} + +// OutboundNatRulePropertiesFormat outbound NAT pool of the load balancer. +type OutboundNatRulePropertiesFormat struct { + // AllocatedOutboundPorts - The number of outbound ports to be used for NAT. + AllocatedOutboundPorts *int32 `json:"allocatedOutboundPorts,omitempty"` + // FrontendIPConfigurations - The Frontend IP addresses of the load balancer. + FrontendIPConfigurations *[]SubResource `json:"frontendIPConfigurations,omitempty"` + // BackendAddressPool - A reference to a pool of DIPs. Outbound traffic is randomly load balanced across IPs in the backend IPs. + BackendAddressPool *SubResource `json:"backendAddressPool,omitempty"` + // ProvisioningState - Gets the provisioning state of the PublicIP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// Probe a load balancer probe. +type Probe struct { + *ProbePropertiesFormat `json:"properties,omitempty"` + // Name - Gets name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for Probe. +func (p Probe) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if p.ProbePropertiesFormat != nil { + objectMap["properties"] = p.ProbePropertiesFormat + } + if p.Name != nil { + objectMap["name"] = p.Name + } + if p.Etag != nil { + objectMap["etag"] = p.Etag + } + if p.ID != nil { + objectMap["id"] = p.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for Probe struct. +func (p *Probe) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var probePropertiesFormat ProbePropertiesFormat + err = json.Unmarshal(*v, &probePropertiesFormat) + if err != nil { + return err + } + p.ProbePropertiesFormat = &probePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + p.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + p.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + p.ID = &ID + } + } + } + + return nil +} + +// ProbePropertiesFormat ... +type ProbePropertiesFormat struct { + // LoadBalancingRules - The load balancer rules that use this probe. + LoadBalancingRules *[]SubResource `json:"loadBalancingRules,omitempty"` + // Protocol - The protocol of the end point. Possible values are: 'Http' or 'Tcp'. If 'Tcp' is specified, a received ACK is required for the probe to be successful. If 'Http' is specified, a 200 OK response from the specifies URI is required for the probe to be successful. Possible values include: 'ProbeProtocolHTTP', 'ProbeProtocolTCP' + Protocol ProbeProtocol `json:"protocol,omitempty"` + // Port - The port for communicating the probe. Possible values range from 1 to 65535, inclusive. + Port *int32 `json:"port,omitempty"` + // IntervalInSeconds - The interval, in seconds, for how frequently to probe the endpoint for health status. Typically, the interval is slightly less than half the allocated timeout period (in seconds) which allows two full probes before taking the instance out of rotation. The default value is 15, the minimum value is 5. + IntervalInSeconds *int32 `json:"intervalInSeconds,omitempty"` + // NumberOfProbes - The number of probes where if no response, will result in stopping further traffic from being delivered to the endpoint. This values allows endpoints to be taken out of rotation faster or slower than the typical times used in Azure. + NumberOfProbes *int32 `json:"numberOfProbes,omitempty"` + // RequestPath - The URI used for requesting health status from the VM. Path is required if a protocol is set to http. Otherwise, it is not allowed. There is no default value. + RequestPath *string `json:"requestPath,omitempty"` + // ProvisioningState - Gets the provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// PublicIPAddress public IP address resource. +type PublicIPAddress struct { + autorest.Response `json:"-"` + *PublicIPAddressPropertiesFormat `json:"properties,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for PublicIPAddress. +func (pia PublicIPAddress) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if pia.PublicIPAddressPropertiesFormat != nil { + objectMap["properties"] = pia.PublicIPAddressPropertiesFormat + } + if pia.Etag != nil { + objectMap["etag"] = pia.Etag + } + if pia.ID != nil { + objectMap["id"] = pia.ID + } + if pia.Name != nil { + objectMap["name"] = pia.Name + } + if pia.Type != nil { + objectMap["type"] = pia.Type + } + if pia.Location != nil { + objectMap["location"] = pia.Location + } + if pia.Tags != nil { + objectMap["tags"] = pia.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for PublicIPAddress struct. +func (pia *PublicIPAddress) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var publicIPAddressPropertiesFormat PublicIPAddressPropertiesFormat + err = json.Unmarshal(*v, &publicIPAddressPropertiesFormat) + if err != nil { + return err + } + pia.PublicIPAddressPropertiesFormat = &publicIPAddressPropertiesFormat + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + pia.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + pia.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + pia.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + pia.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + pia.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + pia.Tags = tags + } + } + } + + return nil +} + +// PublicIPAddressDNSSettings contains FQDN of the DNS record associated with the public IP address +type PublicIPAddressDNSSettings struct { + // DomainNameLabel - Gets or sets the Domain name label.The concatenation of the domain name label and the regionalized DNS zone make up the fully qualified domain name associated with the public IP address. If a domain name label is specified, an A DNS record is created for the public IP in the Microsoft Azure DNS system. + DomainNameLabel *string `json:"domainNameLabel,omitempty"` + // Fqdn - Gets the FQDN, Fully qualified domain name of the A DNS record associated with the public IP. This is the concatenation of the domainNameLabel and the regionalized DNS zone. + Fqdn *string `json:"fqdn,omitempty"` + // ReverseFqdn - Gets or Sets the Reverse FQDN. A user-visible, fully qualified domain name that resolves to this public IP address. If the reverseFqdn is specified, then a PTR DNS record is created pointing from the IP address in the in-addr.arpa domain to the reverse FQDN. + ReverseFqdn *string `json:"reverseFqdn,omitempty"` +} + +// PublicIPAddressesCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type PublicIPAddressesCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future PublicIPAddressesCreateOrUpdateFuture) Result(client PublicIPAddressesClient) (pia PublicIPAddress, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return pia, azure.NewAsyncOpIncompleteError("network.PublicIPAddressesCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + pia, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + pia, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// PublicIPAddressesDeleteFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type PublicIPAddressesDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future PublicIPAddressesDeleteFuture) Result(client PublicIPAddressesClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.PublicIPAddressesDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// PublicIPAddressListResult response for ListPublicIpAddresses API service call. +type PublicIPAddressListResult struct { + autorest.Response `json:"-"` + // Value - A list of public IP addresses that exists in a resource group. + Value *[]PublicIPAddress `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// PublicIPAddressListResultIterator provides access to a complete listing of PublicIPAddress values. +type PublicIPAddressListResultIterator struct { + i int + page PublicIPAddressListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *PublicIPAddressListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter PublicIPAddressListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter PublicIPAddressListResultIterator) Response() PublicIPAddressListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter PublicIPAddressListResultIterator) Value() PublicIPAddress { + if !iter.page.NotDone() { + return PublicIPAddress{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (pialr PublicIPAddressListResult) IsEmpty() bool { + return pialr.Value == nil || len(*pialr.Value) == 0 +} + +// publicIPAddressListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (pialr PublicIPAddressListResult) publicIPAddressListResultPreparer() (*http.Request, error) { + if pialr.NextLink == nil || len(to.String(pialr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(pialr.NextLink))) +} + +// PublicIPAddressListResultPage contains a page of PublicIPAddress values. +type PublicIPAddressListResultPage struct { + fn func(PublicIPAddressListResult) (PublicIPAddressListResult, error) + pialr PublicIPAddressListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *PublicIPAddressListResultPage) Next() error { + next, err := page.fn(page.pialr) + if err != nil { + return err + } + page.pialr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page PublicIPAddressListResultPage) NotDone() bool { + return !page.pialr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page PublicIPAddressListResultPage) Response() PublicIPAddressListResult { + return page.pialr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page PublicIPAddressListResultPage) Values() []PublicIPAddress { + if page.pialr.IsEmpty() { + return nil + } + return *page.pialr.Value +} + +// PublicIPAddressPropertiesFormat public IP address properties. +type PublicIPAddressPropertiesFormat struct { + // PublicIPAllocationMethod - The public IP allocation method. Possible values are: 'Static' and 'Dynamic'. Possible values include: 'Static', 'Dynamic' + PublicIPAllocationMethod IPAllocationMethod `json:"publicIPAllocationMethod,omitempty"` + IPConfiguration *IPConfiguration `json:"ipConfiguration,omitempty"` + // DNSSettings - The FQDN of the DNS record associated with the public IP address. + DNSSettings *PublicIPAddressDNSSettings `json:"dnsSettings,omitempty"` + IPAddress *string `json:"ipAddress,omitempty"` + // IdleTimeoutInMinutes - The idle timeout of the public IP address. + IdleTimeoutInMinutes *int32 `json:"idleTimeoutInMinutes,omitempty"` + // ResourceGUID - The resource GUID property of the public IP resource. + ResourceGUID *string `json:"resourceGuid,omitempty"` + // ProvisioningState - The provisioning state of the PublicIP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// Resource azure resource manager resource properties. +type Resource struct { + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for Resource. +func (r Resource) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if r.ID != nil { + objectMap["id"] = r.ID + } + if r.Name != nil { + objectMap["name"] = r.Name + } + if r.Type != nil { + objectMap["type"] = r.Type + } + if r.Location != nil { + objectMap["location"] = r.Location + } + if r.Tags != nil { + objectMap["tags"] = r.Tags + } + return json.Marshal(objectMap) +} + +// Route route resource +type Route struct { + autorest.Response `json:"-"` + *RoutePropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for Route. +func (r Route) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if r.RoutePropertiesFormat != nil { + objectMap["properties"] = r.RoutePropertiesFormat + } + if r.Name != nil { + objectMap["name"] = r.Name + } + if r.Etag != nil { + objectMap["etag"] = r.Etag + } + if r.ID != nil { + objectMap["id"] = r.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for Route struct. +func (r *Route) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var routePropertiesFormat RoutePropertiesFormat + err = json.Unmarshal(*v, &routePropertiesFormat) + if err != nil { + return err + } + r.RoutePropertiesFormat = &routePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + r.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + r.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + r.ID = &ID + } + } + } + + return nil +} + +// RouteListResult response for the ListRoute API service call +type RouteListResult struct { + autorest.Response `json:"-"` + // Value - Gets a list of routes in a resource group. + Value *[]Route `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// RouteListResultIterator provides access to a complete listing of Route values. +type RouteListResultIterator struct { + i int + page RouteListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *RouteListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter RouteListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter RouteListResultIterator) Response() RouteListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter RouteListResultIterator) Value() Route { + if !iter.page.NotDone() { + return Route{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (rlr RouteListResult) IsEmpty() bool { + return rlr.Value == nil || len(*rlr.Value) == 0 +} + +// routeListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (rlr RouteListResult) routeListResultPreparer() (*http.Request, error) { + if rlr.NextLink == nil || len(to.String(rlr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(rlr.NextLink))) +} + +// RouteListResultPage contains a page of Route values. +type RouteListResultPage struct { + fn func(RouteListResult) (RouteListResult, error) + rlr RouteListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *RouteListResultPage) Next() error { + next, err := page.fn(page.rlr) + if err != nil { + return err + } + page.rlr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page RouteListResultPage) NotDone() bool { + return !page.rlr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page RouteListResultPage) Response() RouteListResult { + return page.rlr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page RouteListResultPage) Values() []Route { + if page.rlr.IsEmpty() { + return nil + } + return *page.rlr.Value +} + +// RoutePropertiesFormat route resource +type RoutePropertiesFormat struct { + // AddressPrefix - The destination CIDR to which the route applies. + AddressPrefix *string `json:"addressPrefix,omitempty"` + // NextHopType - The type of Azure hop the packet should be sent to. Possible values are: 'VirtualNetworkGateway', 'VnetLocal', 'Internet', 'VirtualAppliance', and 'None'. Possible values include: 'RouteNextHopTypeVirtualNetworkGateway', 'RouteNextHopTypeVnetLocal', 'RouteNextHopTypeInternet', 'RouteNextHopTypeVirtualAppliance', 'RouteNextHopTypeNone' + NextHopType RouteNextHopType `json:"nextHopType,omitempty"` + // NextHopIPAddress - The IP address packets should be forwarded to. Next hop values are only allowed in routes where the next hop type is VirtualAppliance. + NextHopIPAddress *string `json:"nextHopIpAddress,omitempty"` + // ProvisioningState - The provisioning state of the resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// RoutesCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a long-running operation. +type RoutesCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future RoutesCreateOrUpdateFuture) Result(client RoutesClient) (r Route, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return r, azure.NewAsyncOpIncompleteError("network.RoutesCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + r, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + r, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// RoutesDeleteFuture an abstraction for monitoring and retrieving the results of a long-running operation. +type RoutesDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future RoutesDeleteFuture) Result(client RoutesClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.RoutesDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// RouteTable route table resource. +type RouteTable struct { + autorest.Response `json:"-"` + *RouteTablePropertiesFormat `json:"properties,omitempty"` + // Etag - Gets a unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for RouteTable. +func (rt RouteTable) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if rt.RouteTablePropertiesFormat != nil { + objectMap["properties"] = rt.RouteTablePropertiesFormat + } + if rt.Etag != nil { + objectMap["etag"] = rt.Etag + } + if rt.ID != nil { + objectMap["id"] = rt.ID + } + if rt.Name != nil { + objectMap["name"] = rt.Name + } + if rt.Type != nil { + objectMap["type"] = rt.Type + } + if rt.Location != nil { + objectMap["location"] = rt.Location + } + if rt.Tags != nil { + objectMap["tags"] = rt.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for RouteTable struct. +func (rt *RouteTable) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var routeTablePropertiesFormat RouteTablePropertiesFormat + err = json.Unmarshal(*v, &routeTablePropertiesFormat) + if err != nil { + return err + } + rt.RouteTablePropertiesFormat = &routeTablePropertiesFormat + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + rt.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + rt.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + rt.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + rt.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + rt.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + rt.Tags = tags + } + } + } + + return nil +} + +// RouteTableListResult response for the ListRouteTable API service call. +type RouteTableListResult struct { + autorest.Response `json:"-"` + // Value - Gets a list of route tables in a resource group. + Value *[]RouteTable `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// RouteTableListResultIterator provides access to a complete listing of RouteTable values. +type RouteTableListResultIterator struct { + i int + page RouteTableListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *RouteTableListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter RouteTableListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter RouteTableListResultIterator) Response() RouteTableListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter RouteTableListResultIterator) Value() RouteTable { + if !iter.page.NotDone() { + return RouteTable{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (rtlr RouteTableListResult) IsEmpty() bool { + return rtlr.Value == nil || len(*rtlr.Value) == 0 +} + +// routeTableListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (rtlr RouteTableListResult) routeTableListResultPreparer() (*http.Request, error) { + if rtlr.NextLink == nil || len(to.String(rtlr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(rtlr.NextLink))) +} + +// RouteTableListResultPage contains a page of RouteTable values. +type RouteTableListResultPage struct { + fn func(RouteTableListResult) (RouteTableListResult, error) + rtlr RouteTableListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *RouteTableListResultPage) Next() error { + next, err := page.fn(page.rtlr) + if err != nil { + return err + } + page.rtlr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page RouteTableListResultPage) NotDone() bool { + return !page.rtlr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page RouteTableListResultPage) Response() RouteTableListResult { + return page.rtlr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page RouteTableListResultPage) Values() []RouteTable { + if page.rtlr.IsEmpty() { + return nil + } + return *page.rtlr.Value +} + +// RouteTablePropertiesFormat route Table resource +type RouteTablePropertiesFormat struct { + // Routes - Collection of routes contained within a route table. + Routes *[]Route `json:"routes,omitempty"` + // Subnets - A collection of references to subnets. + Subnets *[]Subnet `json:"subnets,omitempty"` + // ProvisioningState - The provisioning state of the resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// RouteTablesCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type RouteTablesCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future RouteTablesCreateOrUpdateFuture) Result(client RouteTablesClient) (rt RouteTable, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return rt, azure.NewAsyncOpIncompleteError("network.RouteTablesCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + rt, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + rt, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// RouteTablesDeleteFuture an abstraction for monitoring and retrieving the results of a long-running operation. +type RouteTablesDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future RouteTablesDeleteFuture) Result(client RouteTablesClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.RouteTablesDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// SecurityGroup networkSecurityGroup resource. +type SecurityGroup struct { + autorest.Response `json:"-"` + *SecurityGroupPropertiesFormat `json:"properties,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for SecurityGroup. +func (sg SecurityGroup) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if sg.SecurityGroupPropertiesFormat != nil { + objectMap["properties"] = sg.SecurityGroupPropertiesFormat + } + if sg.Etag != nil { + objectMap["etag"] = sg.Etag + } + if sg.ID != nil { + objectMap["id"] = sg.ID + } + if sg.Name != nil { + objectMap["name"] = sg.Name + } + if sg.Type != nil { + objectMap["type"] = sg.Type + } + if sg.Location != nil { + objectMap["location"] = sg.Location + } + if sg.Tags != nil { + objectMap["tags"] = sg.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for SecurityGroup struct. +func (sg *SecurityGroup) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var securityGroupPropertiesFormat SecurityGroupPropertiesFormat + err = json.Unmarshal(*v, &securityGroupPropertiesFormat) + if err != nil { + return err + } + sg.SecurityGroupPropertiesFormat = &securityGroupPropertiesFormat + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + sg.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + sg.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + sg.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + sg.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + sg.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + sg.Tags = tags + } + } + } + + return nil +} + +// SecurityGroupListResult response for ListNetworkSecurityGroups API service call. +type SecurityGroupListResult struct { + autorest.Response `json:"-"` + // Value - A list of NetworkSecurityGroup resources. + Value *[]SecurityGroup `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// SecurityGroupListResultIterator provides access to a complete listing of SecurityGroup values. +type SecurityGroupListResultIterator struct { + i int + page SecurityGroupListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *SecurityGroupListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter SecurityGroupListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter SecurityGroupListResultIterator) Response() SecurityGroupListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter SecurityGroupListResultIterator) Value() SecurityGroup { + if !iter.page.NotDone() { + return SecurityGroup{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (sglr SecurityGroupListResult) IsEmpty() bool { + return sglr.Value == nil || len(*sglr.Value) == 0 +} + +// securityGroupListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (sglr SecurityGroupListResult) securityGroupListResultPreparer() (*http.Request, error) { + if sglr.NextLink == nil || len(to.String(sglr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(sglr.NextLink))) +} + +// SecurityGroupListResultPage contains a page of SecurityGroup values. +type SecurityGroupListResultPage struct { + fn func(SecurityGroupListResult) (SecurityGroupListResult, error) + sglr SecurityGroupListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *SecurityGroupListResultPage) Next() error { + next, err := page.fn(page.sglr) + if err != nil { + return err + } + page.sglr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page SecurityGroupListResultPage) NotDone() bool { + return !page.sglr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page SecurityGroupListResultPage) Response() SecurityGroupListResult { + return page.sglr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page SecurityGroupListResultPage) Values() []SecurityGroup { + if page.sglr.IsEmpty() { + return nil + } + return *page.sglr.Value +} + +// SecurityGroupPropertiesFormat network Security Group resource. +type SecurityGroupPropertiesFormat struct { + // SecurityRules - A collection of security rules of the network security group. + SecurityRules *[]SecurityRule `json:"securityRules,omitempty"` + // DefaultSecurityRules - The default security rules of network security group. + DefaultSecurityRules *[]SecurityRule `json:"defaultSecurityRules,omitempty"` + // NetworkInterfaces - A collection of references to network interfaces. + NetworkInterfaces *[]Interface `json:"networkInterfaces,omitempty"` + // Subnets - A collection of references to subnets. + Subnets *[]Subnet `json:"subnets,omitempty"` + // ResourceGUID - The resource GUID property of the network security group resource. + ResourceGUID *string `json:"resourceGuid,omitempty"` + // ProvisioningState - The provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// SecurityGroupsCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type SecurityGroupsCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future SecurityGroupsCreateOrUpdateFuture) Result(client SecurityGroupsClient) (sg SecurityGroup, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return sg, azure.NewAsyncOpIncompleteError("network.SecurityGroupsCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + sg, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + sg, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// SecurityGroupsDeleteFuture an abstraction for monitoring and retrieving the results of a long-running operation. +type SecurityGroupsDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future SecurityGroupsDeleteFuture) Result(client SecurityGroupsClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.SecurityGroupsDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// SecurityRule network security rule. +type SecurityRule struct { + autorest.Response `json:"-"` + *SecurityRulePropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for SecurityRule. +func (sr SecurityRule) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if sr.SecurityRulePropertiesFormat != nil { + objectMap["properties"] = sr.SecurityRulePropertiesFormat + } + if sr.Name != nil { + objectMap["name"] = sr.Name + } + if sr.Etag != nil { + objectMap["etag"] = sr.Etag + } + if sr.ID != nil { + objectMap["id"] = sr.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for SecurityRule struct. +func (sr *SecurityRule) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var securityRulePropertiesFormat SecurityRulePropertiesFormat + err = json.Unmarshal(*v, &securityRulePropertiesFormat) + if err != nil { + return err + } + sr.SecurityRulePropertiesFormat = &securityRulePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + sr.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + sr.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + sr.ID = &ID + } + } + } + + return nil +} + +// SecurityRuleListResult response for ListSecurityRule API service call. Retrieves all security rules that belongs +// to a network security group. +type SecurityRuleListResult struct { + autorest.Response `json:"-"` + // Value - The security rules in a network security group. + Value *[]SecurityRule `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// SecurityRuleListResultIterator provides access to a complete listing of SecurityRule values. +type SecurityRuleListResultIterator struct { + i int + page SecurityRuleListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *SecurityRuleListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter SecurityRuleListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter SecurityRuleListResultIterator) Response() SecurityRuleListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter SecurityRuleListResultIterator) Value() SecurityRule { + if !iter.page.NotDone() { + return SecurityRule{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (srlr SecurityRuleListResult) IsEmpty() bool { + return srlr.Value == nil || len(*srlr.Value) == 0 +} + +// securityRuleListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (srlr SecurityRuleListResult) securityRuleListResultPreparer() (*http.Request, error) { + if srlr.NextLink == nil || len(to.String(srlr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(srlr.NextLink))) +} + +// SecurityRuleListResultPage contains a page of SecurityRule values. +type SecurityRuleListResultPage struct { + fn func(SecurityRuleListResult) (SecurityRuleListResult, error) + srlr SecurityRuleListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *SecurityRuleListResultPage) Next() error { + next, err := page.fn(page.srlr) + if err != nil { + return err + } + page.srlr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page SecurityRuleListResultPage) NotDone() bool { + return !page.srlr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page SecurityRuleListResultPage) Response() SecurityRuleListResult { + return page.srlr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page SecurityRuleListResultPage) Values() []SecurityRule { + if page.srlr.IsEmpty() { + return nil + } + return *page.srlr.Value +} + +// SecurityRulePropertiesFormat ... +type SecurityRulePropertiesFormat struct { + // Description - A description for this rule. Restricted to 140 chars. + Description *string `json:"description,omitempty"` + // Protocol - Network protocol this rule applies to. Possible values are 'Tcp', 'Udp', and '*'. Possible values include: 'TCP', 'UDP', 'Asterisk' + Protocol SecurityRuleProtocol `json:"protocol,omitempty"` + // SourcePortRange - The source port or range. Integer or range between 0 and 65535. Asterix '*' can also be used to match all ports. + SourcePortRange *string `json:"sourcePortRange,omitempty"` + // DestinationPortRange - The destination port or range. Integer or range between 0 and 65535. Asterix '*' can also be used to match all ports. + DestinationPortRange *string `json:"destinationPortRange,omitempty"` + // SourceAddressPrefix - The CIDR or source IP range. Asterix '*' can also be used to match all source IPs. Default tags such as 'VirtualNetwork', 'AzureLoadBalancer' and 'Internet' can also be used. If this is an ingress rule, specifies where network traffic originates from. + SourceAddressPrefix *string `json:"sourceAddressPrefix,omitempty"` + // DestinationAddressPrefix - The destination address prefix. CIDR or source IP range. Asterix '*' can also be used to match all source IPs. Default tags such as 'VirtualNetwork', 'AzureLoadBalancer' and 'Internet' can also be used. + DestinationAddressPrefix *string `json:"destinationAddressPrefix,omitempty"` + // Access - The network traffic is allowed or denied. Possible values are: 'Allow' and 'Deny'. Possible values include: 'Allow', 'Deny' + Access SecurityRuleAccess `json:"access,omitempty"` + // Priority - The priority of the rule. The value can be between 100 and 4096. The priority number must be unique for each rule in the collection. The lower the priority number, the higher the priority of the rule. + Priority *int32 `json:"priority,omitempty"` + // Direction - The direction of the rule. The direction specifies if rule will be evaluated on incoming or outcoming traffic. Possible values are: 'Inbound' and 'Outbound'. Possible values include: 'Inbound', 'Outbound' + Direction SecurityRuleDirection `json:"direction,omitempty"` + // ProvisioningState - The provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// SecurityRulesCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type SecurityRulesCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future SecurityRulesCreateOrUpdateFuture) Result(client SecurityRulesClient) (sr SecurityRule, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return sr, azure.NewAsyncOpIncompleteError("network.SecurityRulesCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + sr, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + sr, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// SecurityRulesDeleteFuture an abstraction for monitoring and retrieving the results of a long-running operation. +type SecurityRulesDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future SecurityRulesDeleteFuture) Result(client SecurityRulesClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.SecurityRulesDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// String ... +type String struct { + autorest.Response `json:"-"` + Value *string `json:"value,omitempty"` +} + +// Subnet subnet in a virtual network resource. +type Subnet struct { + autorest.Response `json:"-"` + *SubnetPropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for Subnet. +func (s Subnet) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if s.SubnetPropertiesFormat != nil { + objectMap["properties"] = s.SubnetPropertiesFormat + } + if s.Name != nil { + objectMap["name"] = s.Name + } + if s.Etag != nil { + objectMap["etag"] = s.Etag + } + if s.ID != nil { + objectMap["id"] = s.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for Subnet struct. +func (s *Subnet) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var subnetPropertiesFormat SubnetPropertiesFormat + err = json.Unmarshal(*v, &subnetPropertiesFormat) + if err != nil { + return err + } + s.SubnetPropertiesFormat = &subnetPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + s.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + s.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + s.ID = &ID + } + } + } + + return nil +} + +// SubnetListResult response for ListSubnets API service callRetrieves all subnet that belongs to a virtual network +type SubnetListResult struct { + autorest.Response `json:"-"` + // Value - The subnets in a virtual network. + Value *[]Subnet `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// SubnetListResultIterator provides access to a complete listing of Subnet values. +type SubnetListResultIterator struct { + i int + page SubnetListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *SubnetListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter SubnetListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter SubnetListResultIterator) Response() SubnetListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter SubnetListResultIterator) Value() Subnet { + if !iter.page.NotDone() { + return Subnet{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (slr SubnetListResult) IsEmpty() bool { + return slr.Value == nil || len(*slr.Value) == 0 +} + +// subnetListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (slr SubnetListResult) subnetListResultPreparer() (*http.Request, error) { + if slr.NextLink == nil || len(to.String(slr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(slr.NextLink))) +} + +// SubnetListResultPage contains a page of Subnet values. +type SubnetListResultPage struct { + fn func(SubnetListResult) (SubnetListResult, error) + slr SubnetListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *SubnetListResultPage) Next() error { + next, err := page.fn(page.slr) + if err != nil { + return err + } + page.slr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page SubnetListResultPage) NotDone() bool { + return !page.slr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page SubnetListResultPage) Response() SubnetListResult { + return page.slr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page SubnetListResultPage) Values() []Subnet { + if page.slr.IsEmpty() { + return nil + } + return *page.slr.Value +} + +// SubnetPropertiesFormat ... +type SubnetPropertiesFormat struct { + // AddressPrefix - The address prefix for the subnet. + AddressPrefix *string `json:"addressPrefix,omitempty"` + // NetworkSecurityGroup - The reference of the NetworkSecurityGroup resource. + NetworkSecurityGroup *SecurityGroup `json:"networkSecurityGroup,omitempty"` + // RouteTable - The reference of the RouteTable resource. + RouteTable *RouteTable `json:"routeTable,omitempty"` + // IPConfigurations - Gets an array of references to the network interface IP configurations using subnet. + IPConfigurations *[]IPConfiguration `json:"ipConfigurations,omitempty"` + // ProvisioningState - The provisioning state of the resource. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// SubnetsCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type SubnetsCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future SubnetsCreateOrUpdateFuture) Result(client SubnetsClient) (s Subnet, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return s, azure.NewAsyncOpIncompleteError("network.SubnetsCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + s, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + s, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// SubnetsDeleteFuture an abstraction for monitoring and retrieving the results of a long-running operation. +type SubnetsDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future SubnetsDeleteFuture) Result(client SubnetsClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.SubnetsDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// SubResource azure resource manager sub resource properties. +type SubResource struct { + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// Usage describes network resource usage. +type Usage struct { + // Unit - An enum describing the unit of measurement. + Unit *string `json:"unit,omitempty"` + // CurrentValue - The current value of the usage. + CurrentValue *int64 `json:"currentValue,omitempty"` + // Limit - The limit of usage. + Limit *int64 `json:"limit,omitempty"` + // Name - The name of the type of usage. + Name *UsageName `json:"name,omitempty"` +} + +// UsageName the usage names. +type UsageName struct { + // Value - A string describing the resource name. + Value *string `json:"value,omitempty"` + // LocalizedValue - A localized string describing the resource name. + LocalizedValue *string `json:"localizedValue,omitempty"` +} + +// UsagesListResult the list usages operation response. +type UsagesListResult struct { + autorest.Response `json:"-"` + // Value - The list network resource usages. + Value *[]Usage `json:"value,omitempty"` + // NextLink - URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// UsagesListResultIterator provides access to a complete listing of Usage values. +type UsagesListResultIterator struct { + i int + page UsagesListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *UsagesListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter UsagesListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter UsagesListResultIterator) Response() UsagesListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter UsagesListResultIterator) Value() Usage { + if !iter.page.NotDone() { + return Usage{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (ulr UsagesListResult) IsEmpty() bool { + return ulr.Value == nil || len(*ulr.Value) == 0 +} + +// usagesListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (ulr UsagesListResult) usagesListResultPreparer() (*http.Request, error) { + if ulr.NextLink == nil || len(to.String(ulr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(ulr.NextLink))) +} + +// UsagesListResultPage contains a page of Usage values. +type UsagesListResultPage struct { + fn func(UsagesListResult) (UsagesListResult, error) + ulr UsagesListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *UsagesListResultPage) Next() error { + next, err := page.fn(page.ulr) + if err != nil { + return err + } + page.ulr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page UsagesListResultPage) NotDone() bool { + return !page.ulr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page UsagesListResultPage) Response() UsagesListResult { + return page.ulr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page UsagesListResultPage) Values() []Usage { + if page.ulr.IsEmpty() { + return nil + } + return *page.ulr.Value +} + +// VirtualNetwork virtual Network resource. +type VirtualNetwork struct { + autorest.Response `json:"-"` + *VirtualNetworkPropertiesFormat `json:"properties,omitempty"` + // Etag - Gets a unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for VirtualNetwork. +func (vn VirtualNetwork) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if vn.VirtualNetworkPropertiesFormat != nil { + objectMap["properties"] = vn.VirtualNetworkPropertiesFormat + } + if vn.Etag != nil { + objectMap["etag"] = vn.Etag + } + if vn.ID != nil { + objectMap["id"] = vn.ID + } + if vn.Name != nil { + objectMap["name"] = vn.Name + } + if vn.Type != nil { + objectMap["type"] = vn.Type + } + if vn.Location != nil { + objectMap["location"] = vn.Location + } + if vn.Tags != nil { + objectMap["tags"] = vn.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for VirtualNetwork struct. +func (vn *VirtualNetwork) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var virtualNetworkPropertiesFormat VirtualNetworkPropertiesFormat + err = json.Unmarshal(*v, &virtualNetworkPropertiesFormat) + if err != nil { + return err + } + vn.VirtualNetworkPropertiesFormat = &virtualNetworkPropertiesFormat + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + vn.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + vn.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + vn.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + vn.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + vn.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + vn.Tags = tags + } + } + } + + return nil +} + +// VirtualNetworkGateway a common class for general resource information +type VirtualNetworkGateway struct { + autorest.Response `json:"-"` + *VirtualNetworkGatewayPropertiesFormat `json:"properties,omitempty"` + // Etag - Gets a unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for VirtualNetworkGateway. +func (vng VirtualNetworkGateway) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if vng.VirtualNetworkGatewayPropertiesFormat != nil { + objectMap["properties"] = vng.VirtualNetworkGatewayPropertiesFormat + } + if vng.Etag != nil { + objectMap["etag"] = vng.Etag + } + if vng.ID != nil { + objectMap["id"] = vng.ID + } + if vng.Name != nil { + objectMap["name"] = vng.Name + } + if vng.Type != nil { + objectMap["type"] = vng.Type + } + if vng.Location != nil { + objectMap["location"] = vng.Location + } + if vng.Tags != nil { + objectMap["tags"] = vng.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for VirtualNetworkGateway struct. +func (vng *VirtualNetworkGateway) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var virtualNetworkGatewayPropertiesFormat VirtualNetworkGatewayPropertiesFormat + err = json.Unmarshal(*v, &virtualNetworkGatewayPropertiesFormat) + if err != nil { + return err + } + vng.VirtualNetworkGatewayPropertiesFormat = &virtualNetworkGatewayPropertiesFormat + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + vng.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + vng.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + vng.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + vng.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + vng.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + vng.Tags = tags + } + } + } + + return nil +} + +// VirtualNetworkGatewayConnection a common class for general resource information +type VirtualNetworkGatewayConnection struct { + autorest.Response `json:"-"` + *VirtualNetworkGatewayConnectionPropertiesFormat `json:"properties,omitempty"` + // Etag - Gets a unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` + // Name - Resource name. + Name *string `json:"name,omitempty"` + // Type - Resource type. + Type *string `json:"type,omitempty"` + // Location - Resource location. + Location *string `json:"location,omitempty"` + // Tags - Resource tags. + Tags map[string]*string `json:"tags"` +} + +// MarshalJSON is the custom marshaler for VirtualNetworkGatewayConnection. +func (vngc VirtualNetworkGatewayConnection) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if vngc.VirtualNetworkGatewayConnectionPropertiesFormat != nil { + objectMap["properties"] = vngc.VirtualNetworkGatewayConnectionPropertiesFormat + } + if vngc.Etag != nil { + objectMap["etag"] = vngc.Etag + } + if vngc.ID != nil { + objectMap["id"] = vngc.ID + } + if vngc.Name != nil { + objectMap["name"] = vngc.Name + } + if vngc.Type != nil { + objectMap["type"] = vngc.Type + } + if vngc.Location != nil { + objectMap["location"] = vngc.Location + } + if vngc.Tags != nil { + objectMap["tags"] = vngc.Tags + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for VirtualNetworkGatewayConnection struct. +func (vngc *VirtualNetworkGatewayConnection) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var virtualNetworkGatewayConnectionPropertiesFormat VirtualNetworkGatewayConnectionPropertiesFormat + err = json.Unmarshal(*v, &virtualNetworkGatewayConnectionPropertiesFormat) + if err != nil { + return err + } + vngc.VirtualNetworkGatewayConnectionPropertiesFormat = &virtualNetworkGatewayConnectionPropertiesFormat + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + vngc.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + vngc.ID = &ID + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + vngc.Name = &name + } + case "type": + if v != nil { + var typeVar string + err = json.Unmarshal(*v, &typeVar) + if err != nil { + return err + } + vngc.Type = &typeVar + } + case "location": + if v != nil { + var location string + err = json.Unmarshal(*v, &location) + if err != nil { + return err + } + vngc.Location = &location + } + case "tags": + if v != nil { + var tags map[string]*string + err = json.Unmarshal(*v, &tags) + if err != nil { + return err + } + vngc.Tags = tags + } + } + } + + return nil +} + +// VirtualNetworkGatewayConnectionListResult response for the ListVirtualNetworkGatewayConnections API service call +type VirtualNetworkGatewayConnectionListResult struct { + autorest.Response `json:"-"` + // Value - Gets a list of VirtualNetworkGatewayConnection resources that exists in a resource group. + Value *[]VirtualNetworkGatewayConnection `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// VirtualNetworkGatewayConnectionListResultIterator provides access to a complete listing of +// VirtualNetworkGatewayConnection values. +type VirtualNetworkGatewayConnectionListResultIterator struct { + i int + page VirtualNetworkGatewayConnectionListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *VirtualNetworkGatewayConnectionListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter VirtualNetworkGatewayConnectionListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter VirtualNetworkGatewayConnectionListResultIterator) Response() VirtualNetworkGatewayConnectionListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter VirtualNetworkGatewayConnectionListResultIterator) Value() VirtualNetworkGatewayConnection { + if !iter.page.NotDone() { + return VirtualNetworkGatewayConnection{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (vngclr VirtualNetworkGatewayConnectionListResult) IsEmpty() bool { + return vngclr.Value == nil || len(*vngclr.Value) == 0 +} + +// virtualNetworkGatewayConnectionListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (vngclr VirtualNetworkGatewayConnectionListResult) virtualNetworkGatewayConnectionListResultPreparer() (*http.Request, error) { + if vngclr.NextLink == nil || len(to.String(vngclr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(vngclr.NextLink))) +} + +// VirtualNetworkGatewayConnectionListResultPage contains a page of VirtualNetworkGatewayConnection values. +type VirtualNetworkGatewayConnectionListResultPage struct { + fn func(VirtualNetworkGatewayConnectionListResult) (VirtualNetworkGatewayConnectionListResult, error) + vngclr VirtualNetworkGatewayConnectionListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *VirtualNetworkGatewayConnectionListResultPage) Next() error { + next, err := page.fn(page.vngclr) + if err != nil { + return err + } + page.vngclr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page VirtualNetworkGatewayConnectionListResultPage) NotDone() bool { + return !page.vngclr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page VirtualNetworkGatewayConnectionListResultPage) Response() VirtualNetworkGatewayConnectionListResult { + return page.vngclr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page VirtualNetworkGatewayConnectionListResultPage) Values() []VirtualNetworkGatewayConnection { + if page.vngclr.IsEmpty() { + return nil + } + return *page.vngclr.Value +} + +// VirtualNetworkGatewayConnectionPropertiesFormat virtualNetworkGatewayConnection properties +type VirtualNetworkGatewayConnectionPropertiesFormat struct { + // AuthorizationKey - The authorizationKey. + AuthorizationKey *string `json:"authorizationKey,omitempty"` + VirtualNetworkGateway1 *VirtualNetworkGateway `json:"virtualNetworkGateway1,omitempty"` + VirtualNetworkGateway2 *VirtualNetworkGateway `json:"virtualNetworkGateway2,omitempty"` + LocalNetworkGateway2 *LocalNetworkGateway `json:"localNetworkGateway2,omitempty"` + // ConnectionType - Gateway connection type. Possible values are: 'Ipsec','Vnet2Vnet','ExpressRoute', and 'VPNClient. Possible values include: 'IPsec', 'Vnet2Vnet', 'ExpressRoute', 'VPNClient' + ConnectionType VirtualNetworkGatewayConnectionType `json:"connectionType,omitempty"` + // RoutingWeight - The routing weight. + RoutingWeight *int32 `json:"routingWeight,omitempty"` + // SharedKey - The IPSec shared key. + SharedKey *string `json:"sharedKey,omitempty"` + // ConnectionStatus - Virtual network Gateway connection status. Possible values are 'Unknown', 'Connecting', 'Connected' and 'NotConnected'. Possible values include: 'Unknown', 'Connecting', 'Connected', 'NotConnected' + ConnectionStatus VirtualNetworkGatewayConnectionStatus `json:"connectionStatus,omitempty"` + // EgressBytesTransferred - The egress bytes transferred in this connection. + EgressBytesTransferred *int64 `json:"egressBytesTransferred,omitempty"` + // IngressBytesTransferred - The ingress bytes transferred in this connection. + IngressBytesTransferred *int64 `json:"ingressBytesTransferred,omitempty"` + // Peer - The reference to peerings resource. + Peer *SubResource `json:"peer,omitempty"` + // EnableBgp - EnableBgp flag + EnableBgp *bool `json:"enableBgp,omitempty"` + // ResourceGUID - The resource GUID property of the VirtualNetworkGatewayConnection resource. + ResourceGUID *string `json:"resourceGuid,omitempty"` + // ProvisioningState - The provisioning state of the VirtualNetworkGatewayConnection resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// VirtualNetworkGatewayConnectionsCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of +// a long-running operation. +type VirtualNetworkGatewayConnectionsCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future VirtualNetworkGatewayConnectionsCreateOrUpdateFuture) Result(client VirtualNetworkGatewayConnectionsClient) (vngc VirtualNetworkGatewayConnection, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return vngc, azure.NewAsyncOpIncompleteError("network.VirtualNetworkGatewayConnectionsCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + vngc, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + vngc, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// VirtualNetworkGatewayConnectionsDeleteFuture an abstraction for monitoring and retrieving the results of a +// long-running operation. +type VirtualNetworkGatewayConnectionsDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future VirtualNetworkGatewayConnectionsDeleteFuture) Result(client VirtualNetworkGatewayConnectionsClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.VirtualNetworkGatewayConnectionsDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// VirtualNetworkGatewayConnectionsResetSharedKeyFuture an abstraction for monitoring and retrieving the results of +// a long-running operation. +type VirtualNetworkGatewayConnectionsResetSharedKeyFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future VirtualNetworkGatewayConnectionsResetSharedKeyFuture) Result(client VirtualNetworkGatewayConnectionsClient) (crsk ConnectionResetSharedKey, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsResetSharedKeyFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return crsk, azure.NewAsyncOpIncompleteError("network.VirtualNetworkGatewayConnectionsResetSharedKeyFuture") + } + if future.PollingMethod() == azure.PollingLocation { + crsk, err = client.ResetSharedKeyResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsResetSharedKeyFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsResetSharedKeyFuture", "Result", resp, "Failure sending request") + return + } + crsk, err = client.ResetSharedKeyResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsResetSharedKeyFuture", "Result", resp, "Failure responding to request") + } + return +} + +// VirtualNetworkGatewayConnectionsSetSharedKeyFuture an abstraction for monitoring and retrieving the results of a +// long-running operation. +type VirtualNetworkGatewayConnectionsSetSharedKeyFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future VirtualNetworkGatewayConnectionsSetSharedKeyFuture) Result(client VirtualNetworkGatewayConnectionsClient) (csk ConnectionSharedKey, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsSetSharedKeyFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return csk, azure.NewAsyncOpIncompleteError("network.VirtualNetworkGatewayConnectionsSetSharedKeyFuture") + } + if future.PollingMethod() == azure.PollingLocation { + csk, err = client.SetSharedKeyResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsSetSharedKeyFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsSetSharedKeyFuture", "Result", resp, "Failure sending request") + return + } + csk, err = client.SetSharedKeyResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsSetSharedKeyFuture", "Result", resp, "Failure responding to request") + } + return +} + +// VirtualNetworkGatewayIPConfiguration IP configuration for virtual network gateway +type VirtualNetworkGatewayIPConfiguration struct { + *VirtualNetworkGatewayIPConfigurationPropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for VirtualNetworkGatewayIPConfiguration. +func (vngic VirtualNetworkGatewayIPConfiguration) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if vngic.VirtualNetworkGatewayIPConfigurationPropertiesFormat != nil { + objectMap["properties"] = vngic.VirtualNetworkGatewayIPConfigurationPropertiesFormat + } + if vngic.Name != nil { + objectMap["name"] = vngic.Name + } + if vngic.Etag != nil { + objectMap["etag"] = vngic.Etag + } + if vngic.ID != nil { + objectMap["id"] = vngic.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for VirtualNetworkGatewayIPConfiguration struct. +func (vngic *VirtualNetworkGatewayIPConfiguration) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var virtualNetworkGatewayIPConfigurationPropertiesFormat VirtualNetworkGatewayIPConfigurationPropertiesFormat + err = json.Unmarshal(*v, &virtualNetworkGatewayIPConfigurationPropertiesFormat) + if err != nil { + return err + } + vngic.VirtualNetworkGatewayIPConfigurationPropertiesFormat = &virtualNetworkGatewayIPConfigurationPropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + vngic.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + vngic.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + vngic.ID = &ID + } + } + } + + return nil +} + +// VirtualNetworkGatewayIPConfigurationPropertiesFormat properties of VirtualNetworkGatewayIPConfiguration +type VirtualNetworkGatewayIPConfigurationPropertiesFormat struct { + // PrivateIPAddress - Gets or sets the privateIPAddress of the IP Configuration + PrivateIPAddress *string `json:"privateIPAddress,omitempty"` + // PrivateIPAllocationMethod - The private IP allocation method. Possible values are: 'Static' and 'Dynamic'. Possible values include: 'Static', 'Dynamic' + PrivateIPAllocationMethod IPAllocationMethod `json:"privateIPAllocationMethod,omitempty"` + // Subnet - The reference of the subnet resource. + Subnet *SubResource `json:"subnet,omitempty"` + // PublicIPAddress - The reference of the public IP resource. + PublicIPAddress *SubResource `json:"publicIPAddress,omitempty"` + // ProvisioningState - The provisioning state of the public IP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// VirtualNetworkGatewayListResult response for the ListVirtualNetworkGateways API service call. +type VirtualNetworkGatewayListResult struct { + autorest.Response `json:"-"` + // Value - Gets a list of VirtualNetworkGateway resources that exists in a resource group. + Value *[]VirtualNetworkGateway `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// VirtualNetworkGatewayListResultIterator provides access to a complete listing of VirtualNetworkGateway values. +type VirtualNetworkGatewayListResultIterator struct { + i int + page VirtualNetworkGatewayListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *VirtualNetworkGatewayListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter VirtualNetworkGatewayListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter VirtualNetworkGatewayListResultIterator) Response() VirtualNetworkGatewayListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter VirtualNetworkGatewayListResultIterator) Value() VirtualNetworkGateway { + if !iter.page.NotDone() { + return VirtualNetworkGateway{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (vnglr VirtualNetworkGatewayListResult) IsEmpty() bool { + return vnglr.Value == nil || len(*vnglr.Value) == 0 +} + +// virtualNetworkGatewayListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (vnglr VirtualNetworkGatewayListResult) virtualNetworkGatewayListResultPreparer() (*http.Request, error) { + if vnglr.NextLink == nil || len(to.String(vnglr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(vnglr.NextLink))) +} + +// VirtualNetworkGatewayListResultPage contains a page of VirtualNetworkGateway values. +type VirtualNetworkGatewayListResultPage struct { + fn func(VirtualNetworkGatewayListResult) (VirtualNetworkGatewayListResult, error) + vnglr VirtualNetworkGatewayListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *VirtualNetworkGatewayListResultPage) Next() error { + next, err := page.fn(page.vnglr) + if err != nil { + return err + } + page.vnglr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page VirtualNetworkGatewayListResultPage) NotDone() bool { + return !page.vnglr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page VirtualNetworkGatewayListResultPage) Response() VirtualNetworkGatewayListResult { + return page.vnglr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page VirtualNetworkGatewayListResultPage) Values() []VirtualNetworkGateway { + if page.vnglr.IsEmpty() { + return nil + } + return *page.vnglr.Value +} + +// VirtualNetworkGatewayPropertiesFormat virtualNetworkGateway properties +type VirtualNetworkGatewayPropertiesFormat struct { + // IPConfigurations - IP configurations for virtual network gateway. + IPConfigurations *[]VirtualNetworkGatewayIPConfiguration `json:"ipConfigurations,omitempty"` + // GatewayType - The type of this virtual network gateway. Possible values are: 'Vpn' and 'ExpressRoute'. Possible values include: 'VirtualNetworkGatewayTypeVpn', 'VirtualNetworkGatewayTypeExpressRoute' + GatewayType VirtualNetworkGatewayType `json:"gatewayType,omitempty"` + // VpnType - The type of this virtual network gateway. Possible values are: 'PolicyBased' and 'RouteBased'. Possible values include: 'PolicyBased', 'RouteBased' + VpnType VpnType `json:"vpnType,omitempty"` + // EnableBgp - Whether BGP is enabled for this virtual network gateway or not. + EnableBgp *bool `json:"enableBgp,omitempty"` + // GatewayDefaultSite - The reference of the LocalNetworkGateway resource which represents local network site having default routes. Assign Null value in case of removing existing default site setting. + GatewayDefaultSite *SubResource `json:"gatewayDefaultSite,omitempty"` + // Sku - The reference of the VirtualNetworkGatewaySku resource which represents the SKU selected for Virtual network gateway. + Sku *VirtualNetworkGatewaySku `json:"sku,omitempty"` + // VpnClientConfiguration - The reference of the VpnClientConfiguration resource which represents the P2S VpnClient configurations. + VpnClientConfiguration *VpnClientConfiguration `json:"vpnClientConfiguration,omitempty"` + // BgpSettings - Virtual network gateway's BGP speaker settings. + BgpSettings *BgpSettings `json:"bgpSettings,omitempty"` + // ResourceGUID - The resource GUID property of the VirtualNetworkGateway resource. + ResourceGUID *string `json:"resourceGuid,omitempty"` + // ProvisioningState - The provisioning state of the VirtualNetworkGateway resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// VirtualNetworkGatewaysCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a +// long-running operation. +type VirtualNetworkGatewaysCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future VirtualNetworkGatewaysCreateOrUpdateFuture) Result(client VirtualNetworkGatewaysClient) (vng VirtualNetworkGateway, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return vng, azure.NewAsyncOpIncompleteError("network.VirtualNetworkGatewaysCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + vng, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + vng, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// VirtualNetworkGatewaysDeleteFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type VirtualNetworkGatewaysDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future VirtualNetworkGatewaysDeleteFuture) Result(client VirtualNetworkGatewaysClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.VirtualNetworkGatewaysDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// VirtualNetworkGatewaySku virtualNetworkGatewaySku details +type VirtualNetworkGatewaySku struct { + // Name - Gateway sku name -Basic/HighPerformance/Standard. Possible values include: 'VirtualNetworkGatewaySkuNameBasic', 'VirtualNetworkGatewaySkuNameHighPerformance', 'VirtualNetworkGatewaySkuNameStandard' + Name VirtualNetworkGatewaySkuName `json:"name,omitempty"` + // Tier - Gateway sku tier -Basic/HighPerformance/Standard. Possible values include: 'VirtualNetworkGatewaySkuTierBasic', 'VirtualNetworkGatewaySkuTierHighPerformance', 'VirtualNetworkGatewaySkuTierStandard' + Tier VirtualNetworkGatewaySkuTier `json:"tier,omitempty"` + // Capacity - The capacity + Capacity *int32 `json:"capacity,omitempty"` +} + +// VirtualNetworkGatewaysResetFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type VirtualNetworkGatewaysResetFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future VirtualNetworkGatewaysResetFuture) Result(client VirtualNetworkGatewaysClient) (vng VirtualNetworkGateway, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysResetFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return vng, azure.NewAsyncOpIncompleteError("network.VirtualNetworkGatewaysResetFuture") + } + if future.PollingMethod() == azure.PollingLocation { + vng, err = client.ResetResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysResetFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysResetFuture", "Result", resp, "Failure sending request") + return + } + vng, err = client.ResetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysResetFuture", "Result", resp, "Failure responding to request") + } + return +} + +// VirtualNetworkListResult response for the ListVirtualNetworks API service call. +type VirtualNetworkListResult struct { + autorest.Response `json:"-"` + // Value - Gets a list of VirtualNetwork resources in a resource group. + Value *[]VirtualNetwork `json:"value,omitempty"` + // NextLink - The URL to get the next set of results. + NextLink *string `json:"nextLink,omitempty"` +} + +// VirtualNetworkListResultIterator provides access to a complete listing of VirtualNetwork values. +type VirtualNetworkListResultIterator struct { + i int + page VirtualNetworkListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *VirtualNetworkListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter VirtualNetworkListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter VirtualNetworkListResultIterator) Response() VirtualNetworkListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter VirtualNetworkListResultIterator) Value() VirtualNetwork { + if !iter.page.NotDone() { + return VirtualNetwork{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (vnlr VirtualNetworkListResult) IsEmpty() bool { + return vnlr.Value == nil || len(*vnlr.Value) == 0 +} + +// virtualNetworkListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (vnlr VirtualNetworkListResult) virtualNetworkListResultPreparer() (*http.Request, error) { + if vnlr.NextLink == nil || len(to.String(vnlr.NextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(vnlr.NextLink))) +} + +// VirtualNetworkListResultPage contains a page of VirtualNetwork values. +type VirtualNetworkListResultPage struct { + fn func(VirtualNetworkListResult) (VirtualNetworkListResult, error) + vnlr VirtualNetworkListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *VirtualNetworkListResultPage) Next() error { + next, err := page.fn(page.vnlr) + if err != nil { + return err + } + page.vnlr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page VirtualNetworkListResultPage) NotDone() bool { + return !page.vnlr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page VirtualNetworkListResultPage) Response() VirtualNetworkListResult { + return page.vnlr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page VirtualNetworkListResultPage) Values() []VirtualNetwork { + if page.vnlr.IsEmpty() { + return nil + } + return *page.vnlr.Value +} + +// VirtualNetworkPropertiesFormat ... +type VirtualNetworkPropertiesFormat struct { + // AddressSpace - The AddressSpace that contains an array of IP address ranges that can be used by subnets. + AddressSpace *AddressSpace `json:"addressSpace,omitempty"` + // DhcpOptions - The dhcpOptions that contains an array of DNS servers available to VMs deployed in the virtual network. + DhcpOptions *DhcpOptions `json:"dhcpOptions,omitempty"` + // Subnets - A list of subnets in a Virtual Network. + Subnets *[]Subnet `json:"subnets,omitempty"` + // ResourceGUID - The resourceGuid property of the Virtual Network resource. + ResourceGUID *string `json:"resourceGuid,omitempty"` + // ProvisioningState - The provisioning state of the PublicIP resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// VirtualNetworksCreateOrUpdateFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type VirtualNetworksCreateOrUpdateFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future VirtualNetworksCreateOrUpdateFuture) Result(client VirtualNetworksClient) (vn VirtualNetwork, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksCreateOrUpdateFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return vn, azure.NewAsyncOpIncompleteError("network.VirtualNetworksCreateOrUpdateFuture") + } + if future.PollingMethod() == azure.PollingLocation { + vn, err = client.CreateOrUpdateResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksCreateOrUpdateFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksCreateOrUpdateFuture", "Result", resp, "Failure sending request") + return + } + vn, err = client.CreateOrUpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksCreateOrUpdateFuture", "Result", resp, "Failure responding to request") + } + return +} + +// VirtualNetworksDeleteFuture an abstraction for monitoring and retrieving the results of a long-running +// operation. +type VirtualNetworksDeleteFuture struct { + azure.Future + req *http.Request +} + +// Result returns the result of the asynchronous operation. +// If the operation has not completed it will return an error. +func (future VirtualNetworksDeleteFuture) Result(client VirtualNetworksClient) (ar autorest.Response, err error) { + var done bool + done, err = future.Done(client) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksDeleteFuture", "Result", future.Response(), "Polling failure") + return + } + if !done { + return ar, azure.NewAsyncOpIncompleteError("network.VirtualNetworksDeleteFuture") + } + if future.PollingMethod() == azure.PollingLocation { + ar, err = client.DeleteResponder(future.Response()) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksDeleteFuture", "Result", future.Response(), "Failure responding to request") + } + return + } + var req *http.Request + var resp *http.Response + if future.PollingURL() != "" { + req, err = http.NewRequest(http.MethodGet, future.PollingURL(), nil) + if err != nil { + return + } + } else { + req = autorest.ChangeToGet(future.req) + } + resp, err = autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksDeleteFuture", "Result", resp, "Failure sending request") + return + } + ar, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksDeleteFuture", "Result", resp, "Failure responding to request") + } + return +} + +// VpnClientConfiguration vpnClientConfiguration for P2S client +type VpnClientConfiguration struct { + // VpnClientAddressPool - Gets or sets the reference of the Address space resource which represents Address space for P2S VpnClient. + VpnClientAddressPool *AddressSpace `json:"vpnClientAddressPool,omitempty"` + // VpnClientRootCertificates - VpnClientRootCertificate for Virtual network gateway. + VpnClientRootCertificates *[]VpnClientRootCertificate `json:"vpnClientRootCertificates,omitempty"` + // VpnClientRevokedCertificates - VpnClientRevokedCertificate for Virtual network gateway. + VpnClientRevokedCertificates *[]VpnClientRevokedCertificate `json:"vpnClientRevokedCertificates,omitempty"` +} + +// VpnClientParameters vpnClientParameters +type VpnClientParameters struct { + // ProcessorArchitecture - VPN client Processor Architecture. Possible values are: 'AMD64' and 'X86'. Possible values include: 'Amd64', 'X86' + ProcessorArchitecture ProcessorArchitecture `json:"ProcessorArchitecture,omitempty"` +} + +// VpnClientRevokedCertificate VPN client revoked certificate of virtual network gateway. +type VpnClientRevokedCertificate struct { + *VpnClientRevokedCertificatePropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for VpnClientRevokedCertificate. +func (vcrc VpnClientRevokedCertificate) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if vcrc.VpnClientRevokedCertificatePropertiesFormat != nil { + objectMap["properties"] = vcrc.VpnClientRevokedCertificatePropertiesFormat + } + if vcrc.Name != nil { + objectMap["name"] = vcrc.Name + } + if vcrc.Etag != nil { + objectMap["etag"] = vcrc.Etag + } + if vcrc.ID != nil { + objectMap["id"] = vcrc.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for VpnClientRevokedCertificate struct. +func (vcrc *VpnClientRevokedCertificate) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var vpnClientRevokedCertificatePropertiesFormat VpnClientRevokedCertificatePropertiesFormat + err = json.Unmarshal(*v, &vpnClientRevokedCertificatePropertiesFormat) + if err != nil { + return err + } + vcrc.VpnClientRevokedCertificatePropertiesFormat = &vpnClientRevokedCertificatePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + vcrc.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + vcrc.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + vcrc.ID = &ID + } + } + } + + return nil +} + +// VpnClientRevokedCertificatePropertiesFormat properties of the revoked VPN client certificate of virtual network +// gateway. +type VpnClientRevokedCertificatePropertiesFormat struct { + // Thumbprint - The revoked VPN client certificate thumbprint. + Thumbprint *string `json:"thumbprint,omitempty"` + // ProvisioningState - The provisioning state of the VPN client revoked certificate resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} + +// VpnClientRootCertificate VPN client root certificate of virtual network gateway +type VpnClientRootCertificate struct { + *VpnClientRootCertificatePropertiesFormat `json:"properties,omitempty"` + // Name - The name of the resource that is unique within a resource group. This name can be used to access the resource. + Name *string `json:"name,omitempty"` + // Etag - A unique read-only string that changes whenever the resource is updated. + Etag *string `json:"etag,omitempty"` + // ID - Resource Identifier. + ID *string `json:"id,omitempty"` +} + +// MarshalJSON is the custom marshaler for VpnClientRootCertificate. +func (vcrc VpnClientRootCertificate) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]interface{}) + if vcrc.VpnClientRootCertificatePropertiesFormat != nil { + objectMap["properties"] = vcrc.VpnClientRootCertificatePropertiesFormat + } + if vcrc.Name != nil { + objectMap["name"] = vcrc.Name + } + if vcrc.Etag != nil { + objectMap["etag"] = vcrc.Etag + } + if vcrc.ID != nil { + objectMap["id"] = vcrc.ID + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON is the custom unmarshaler for VpnClientRootCertificate struct. +func (vcrc *VpnClientRootCertificate) UnmarshalJSON(body []byte) error { + var m map[string]*json.RawMessage + err := json.Unmarshal(body, &m) + if err != nil { + return err + } + for k, v := range m { + switch k { + case "properties": + if v != nil { + var vpnClientRootCertificatePropertiesFormat VpnClientRootCertificatePropertiesFormat + err = json.Unmarshal(*v, &vpnClientRootCertificatePropertiesFormat) + if err != nil { + return err + } + vcrc.VpnClientRootCertificatePropertiesFormat = &vpnClientRootCertificatePropertiesFormat + } + case "name": + if v != nil { + var name string + err = json.Unmarshal(*v, &name) + if err != nil { + return err + } + vcrc.Name = &name + } + case "etag": + if v != nil { + var etag string + err = json.Unmarshal(*v, &etag) + if err != nil { + return err + } + vcrc.Etag = &etag + } + case "id": + if v != nil { + var ID string + err = json.Unmarshal(*v, &ID) + if err != nil { + return err + } + vcrc.ID = &ID + } + } + } + + return nil +} + +// VpnClientRootCertificatePropertiesFormat properties of SSL certificates of application gateway +type VpnClientRootCertificatePropertiesFormat struct { + // PublicCertData - Gets or sets the certificate public data + PublicCertData *string `json:"publicCertData,omitempty"` + // ProvisioningState - The provisioning state of the VPN client root certificate resource. Possible values are: 'Updating', 'Deleting', and 'Failed'. + ProvisioningState *string `json:"provisioningState,omitempty"` +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/publicipaddresses.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/publicipaddresses.go new file mode 100644 index 000000000..f733de4f0 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/publicipaddresses.go @@ -0,0 +1,446 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/validation" + "net/http" +) + +// PublicIPAddressesClient is the network Client +type PublicIPAddressesClient struct { + BaseClient +} + +// NewPublicIPAddressesClient creates an instance of the PublicIPAddressesClient client. +func NewPublicIPAddressesClient(subscriptionID string) PublicIPAddressesClient { + return NewPublicIPAddressesClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewPublicIPAddressesClientWithBaseURI creates an instance of the PublicIPAddressesClient client. +func NewPublicIPAddressesClientWithBaseURI(baseURI string, subscriptionID string) PublicIPAddressesClient { + return PublicIPAddressesClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a static or dynamic public IP address. +// Parameters: +// resourceGroupName - the name of the resource group. +// publicIPAddressName - the name of the public IP address. +// parameters - parameters supplied to the create or update public IP address operation. +func (client PublicIPAddressesClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, publicIPAddressName string, parameters PublicIPAddress) (result PublicIPAddressesCreateOrUpdateFuture, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: parameters, + Constraints: []validation.Constraint{{Target: "parameters.PublicIPAddressPropertiesFormat", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "parameters.PublicIPAddressPropertiesFormat.IPConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "parameters.PublicIPAddressPropertiesFormat.IPConfiguration.IPConfigurationPropertiesFormat", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "parameters.PublicIPAddressPropertiesFormat.IPConfiguration.IPConfigurationPropertiesFormat.PublicIPAddress", Name: validation.Null, Rule: false, Chain: nil}}}, + }}, + }}}}}); err != nil { + return result, validation.NewError("network.PublicIPAddressesClient", "CreateOrUpdate", err.Error()) + } + + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, publicIPAddressName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client PublicIPAddressesClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, publicIPAddressName string, parameters PublicIPAddress) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "publicIpAddressName": autorest.Encode("path", publicIPAddressName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/publicIPAddresses/{publicIpAddressName}", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client PublicIPAddressesClient) CreateOrUpdateSender(req *http.Request) (future PublicIPAddressesCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client PublicIPAddressesClient) CreateOrUpdateResponder(resp *http.Response) (result PublicIPAddress, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified public IP address. +// Parameters: +// resourceGroupName - the name of the resource group. +// publicIPAddressName - the name of the subnet. +func (client PublicIPAddressesClient) Delete(ctx context.Context, resourceGroupName string, publicIPAddressName string) (result PublicIPAddressesDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, publicIPAddressName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client PublicIPAddressesClient) DeletePreparer(ctx context.Context, resourceGroupName string, publicIPAddressName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "publicIpAddressName": autorest.Encode("path", publicIPAddressName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/publicIPAddresses/{publicIpAddressName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client PublicIPAddressesClient) DeleteSender(req *http.Request) (future PublicIPAddressesDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client PublicIPAddressesClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified public IP address in a specified resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// publicIPAddressName - the name of the subnet. +// expand - expands referenced resources. +func (client PublicIPAddressesClient) Get(ctx context.Context, resourceGroupName string, publicIPAddressName string, expand string) (result PublicIPAddress, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, publicIPAddressName, expand) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client PublicIPAddressesClient) GetPreparer(ctx context.Context, resourceGroupName string, publicIPAddressName string, expand string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "publicIpAddressName": autorest.Encode("path", publicIPAddressName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/publicIPAddresses/{publicIpAddressName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client PublicIPAddressesClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client PublicIPAddressesClient) GetResponder(resp *http.Response) (result PublicIPAddress, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all public IP addresses in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +func (client PublicIPAddressesClient) List(ctx context.Context, resourceGroupName string) (result PublicIPAddressListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.pialr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "List", resp, "Failure sending request") + return + } + + result.pialr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client PublicIPAddressesClient) ListPreparer(ctx context.Context, resourceGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/publicIPAddresses", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client PublicIPAddressesClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client PublicIPAddressesClient) ListResponder(resp *http.Response) (result PublicIPAddressListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client PublicIPAddressesClient) listNextResults(lastResults PublicIPAddressListResult) (result PublicIPAddressListResult, err error) { + req, err := lastResults.publicIPAddressListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client PublicIPAddressesClient) ListComplete(ctx context.Context, resourceGroupName string) (result PublicIPAddressListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName) + return +} + +// ListAll gets all the public IP addresses in a subscription. +func (client PublicIPAddressesClient) ListAll(ctx context.Context) (result PublicIPAddressListResultPage, err error) { + result.fn = client.listAllNextResults + req, err := client.ListAllPreparer(ctx) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "ListAll", nil, "Failure preparing request") + return + } + + resp, err := client.ListAllSender(req) + if err != nil { + result.pialr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "ListAll", resp, "Failure sending request") + return + } + + result.pialr, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "ListAll", resp, "Failure responding to request") + } + + return +} + +// ListAllPreparer prepares the ListAll request. +func (client PublicIPAddressesClient) ListAllPreparer(ctx context.Context) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/Microsoft.Network/publicIPAddresses", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListAllSender sends the ListAll request. The method will close the +// http.Response Body if it receives an error. +func (client PublicIPAddressesClient) ListAllSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListAllResponder handles the response to the ListAll request. The method always +// closes the http.Response Body. +func (client PublicIPAddressesClient) ListAllResponder(resp *http.Response) (result PublicIPAddressListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listAllNextResults retrieves the next set of results, if any. +func (client PublicIPAddressesClient) listAllNextResults(lastResults PublicIPAddressListResult) (result PublicIPAddressListResult, err error) { + req, err := lastResults.publicIPAddressListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "listAllNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListAllSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "listAllNextResults", resp, "Failure sending next results request") + } + result, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.PublicIPAddressesClient", "listAllNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListAllComplete enumerates all values, automatically crossing page boundaries as required. +func (client PublicIPAddressesClient) ListAllComplete(ctx context.Context) (result PublicIPAddressListResultIterator, err error) { + result.page, err = client.ListAll(ctx) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/routes.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/routes.go new file mode 100644 index 000000000..c8c41ea1f --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/routes.go @@ -0,0 +1,348 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// RoutesClient is the network Client +type RoutesClient struct { + BaseClient +} + +// NewRoutesClient creates an instance of the RoutesClient client. +func NewRoutesClient(subscriptionID string) RoutesClient { + return NewRoutesClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewRoutesClientWithBaseURI creates an instance of the RoutesClient client. +func NewRoutesClientWithBaseURI(baseURI string, subscriptionID string) RoutesClient { + return RoutesClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a route in the specified route table. +// Parameters: +// resourceGroupName - the name of the resource group. +// routeTableName - the name of the route table. +// routeName - the name of the route. +// routeParameters - parameters supplied to the create or update route operation. +func (client RoutesClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, routeTableName string, routeName string, routeParameters Route) (result RoutesCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, routeTableName, routeName, routeParameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client RoutesClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, routeTableName string, routeName string, routeParameters Route) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "routeName": autorest.Encode("path", routeName), + "routeTableName": autorest.Encode("path", routeTableName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables/{routeTableName}/routes/{routeName}", pathParameters), + autorest.WithJSON(routeParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client RoutesClient) CreateOrUpdateSender(req *http.Request) (future RoutesCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client RoutesClient) CreateOrUpdateResponder(resp *http.Response) (result Route, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified route from a route table. +// Parameters: +// resourceGroupName - the name of the resource group. +// routeTableName - the name of the route table. +// routeName - the name of the route. +func (client RoutesClient) Delete(ctx context.Context, resourceGroupName string, routeTableName string, routeName string) (result RoutesDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, routeTableName, routeName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client RoutesClient) DeletePreparer(ctx context.Context, resourceGroupName string, routeTableName string, routeName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "routeName": autorest.Encode("path", routeName), + "routeTableName": autorest.Encode("path", routeTableName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables/{routeTableName}/routes/{routeName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client RoutesClient) DeleteSender(req *http.Request) (future RoutesDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client RoutesClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified route from a route table. +// Parameters: +// resourceGroupName - the name of the resource group. +// routeTableName - the name of the route table. +// routeName - the name of the route. +func (client RoutesClient) Get(ctx context.Context, resourceGroupName string, routeTableName string, routeName string) (result Route, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, routeTableName, routeName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.RoutesClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client RoutesClient) GetPreparer(ctx context.Context, resourceGroupName string, routeTableName string, routeName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "routeName": autorest.Encode("path", routeName), + "routeTableName": autorest.Encode("path", routeTableName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables/{routeTableName}/routes/{routeName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client RoutesClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client RoutesClient) GetResponder(resp *http.Response) (result Route, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all routes in a route table. +// Parameters: +// resourceGroupName - the name of the resource group. +// routeTableName - the name of the route table. +func (client RoutesClient) List(ctx context.Context, resourceGroupName string, routeTableName string) (result RouteListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName, routeTableName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.rlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.RoutesClient", "List", resp, "Failure sending request") + return + } + + result.rlr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client RoutesClient) ListPreparer(ctx context.Context, resourceGroupName string, routeTableName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "routeTableName": autorest.Encode("path", routeTableName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables/{routeTableName}/routes", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client RoutesClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client RoutesClient) ListResponder(resp *http.Response) (result RouteListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client RoutesClient) listNextResults(lastResults RouteListResult) (result RouteListResult, err error) { + req, err := lastResults.routeListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.RoutesClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.RoutesClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RoutesClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client RoutesClient) ListComplete(ctx context.Context, resourceGroupName string, routeTableName string) (result RouteListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName, routeTableName) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/routetables.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/routetables.go new file mode 100644 index 000000000..74097ca5b --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/routetables.go @@ -0,0 +1,434 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// RouteTablesClient is the network Client +type RouteTablesClient struct { + BaseClient +} + +// NewRouteTablesClient creates an instance of the RouteTablesClient client. +func NewRouteTablesClient(subscriptionID string) RouteTablesClient { + return NewRouteTablesClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewRouteTablesClientWithBaseURI creates an instance of the RouteTablesClient client. +func NewRouteTablesClientWithBaseURI(baseURI string, subscriptionID string) RouteTablesClient { + return RouteTablesClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate create or updates a route table in a specified resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// routeTableName - the name of the route table. +// parameters - parameters supplied to the create or update route table operation. +func (client RouteTablesClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, routeTableName string, parameters RouteTable) (result RouteTablesCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, routeTableName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client RouteTablesClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, routeTableName string, parameters RouteTable) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "routeTableName": autorest.Encode("path", routeTableName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables/{routeTableName}", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client RouteTablesClient) CreateOrUpdateSender(req *http.Request) (future RouteTablesCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client RouteTablesClient) CreateOrUpdateResponder(resp *http.Response) (result RouteTable, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified route table. +// Parameters: +// resourceGroupName - the name of the resource group. +// routeTableName - the name of the route table. +func (client RouteTablesClient) Delete(ctx context.Context, resourceGroupName string, routeTableName string) (result RouteTablesDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, routeTableName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client RouteTablesClient) DeletePreparer(ctx context.Context, resourceGroupName string, routeTableName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "routeTableName": autorest.Encode("path", routeTableName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables/{routeTableName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client RouteTablesClient) DeleteSender(req *http.Request) (future RouteTablesDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client RouteTablesClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified route table. +// Parameters: +// resourceGroupName - the name of the resource group. +// routeTableName - the name of the route table. +// expand - expands referenced resources. +func (client RouteTablesClient) Get(ctx context.Context, resourceGroupName string, routeTableName string, expand string) (result RouteTable, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, routeTableName, expand) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client RouteTablesClient) GetPreparer(ctx context.Context, resourceGroupName string, routeTableName string, expand string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "routeTableName": autorest.Encode("path", routeTableName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables/{routeTableName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client RouteTablesClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client RouteTablesClient) GetResponder(resp *http.Response) (result RouteTable, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all route tables in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +func (client RouteTablesClient) List(ctx context.Context, resourceGroupName string) (result RouteTableListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.rtlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "List", resp, "Failure sending request") + return + } + + result.rtlr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client RouteTablesClient) ListPreparer(ctx context.Context, resourceGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client RouteTablesClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client RouteTablesClient) ListResponder(resp *http.Response) (result RouteTableListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client RouteTablesClient) listNextResults(lastResults RouteTableListResult) (result RouteTableListResult, err error) { + req, err := lastResults.routeTableListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.RouteTablesClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.RouteTablesClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client RouteTablesClient) ListComplete(ctx context.Context, resourceGroupName string) (result RouteTableListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName) + return +} + +// ListAll gets all route tables in a subscription. +func (client RouteTablesClient) ListAll(ctx context.Context) (result RouteTableListResultPage, err error) { + result.fn = client.listAllNextResults + req, err := client.ListAllPreparer(ctx) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "ListAll", nil, "Failure preparing request") + return + } + + resp, err := client.ListAllSender(req) + if err != nil { + result.rtlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "ListAll", resp, "Failure sending request") + return + } + + result.rtlr, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "ListAll", resp, "Failure responding to request") + } + + return +} + +// ListAllPreparer prepares the ListAll request. +func (client RouteTablesClient) ListAllPreparer(ctx context.Context) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/Microsoft.Network/routeTables", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListAllSender sends the ListAll request. The method will close the +// http.Response Body if it receives an error. +func (client RouteTablesClient) ListAllSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListAllResponder handles the response to the ListAll request. The method always +// closes the http.Response Body. +func (client RouteTablesClient) ListAllResponder(resp *http.Response) (result RouteTableListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listAllNextResults retrieves the next set of results, if any. +func (client RouteTablesClient) listAllNextResults(lastResults RouteTableListResult) (result RouteTableListResult, err error) { + req, err := lastResults.routeTableListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.RouteTablesClient", "listAllNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListAllSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.RouteTablesClient", "listAllNextResults", resp, "Failure sending next results request") + } + result, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.RouteTablesClient", "listAllNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListAllComplete enumerates all values, automatically crossing page boundaries as required. +func (client RouteTablesClient) ListAllComplete(ctx context.Context) (result RouteTableListResultIterator, err error) { + result.page, err = client.ListAll(ctx) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/securitygroups.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/securitygroups.go new file mode 100644 index 000000000..42dd80275 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/securitygroups.go @@ -0,0 +1,434 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// SecurityGroupsClient is the network Client +type SecurityGroupsClient struct { + BaseClient +} + +// NewSecurityGroupsClient creates an instance of the SecurityGroupsClient client. +func NewSecurityGroupsClient(subscriptionID string) SecurityGroupsClient { + return NewSecurityGroupsClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewSecurityGroupsClientWithBaseURI creates an instance of the SecurityGroupsClient client. +func NewSecurityGroupsClientWithBaseURI(baseURI string, subscriptionID string) SecurityGroupsClient { + return SecurityGroupsClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a network security group in the specified resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// networkSecurityGroupName - the name of the network security group. +// parameters - parameters supplied to the create or update network security group operation. +func (client SecurityGroupsClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, networkSecurityGroupName string, parameters SecurityGroup) (result SecurityGroupsCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, networkSecurityGroupName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client SecurityGroupsClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, networkSecurityGroupName string, parameters SecurityGroup) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "networkSecurityGroupName": autorest.Encode("path", networkSecurityGroupName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkSecurityGroups/{networkSecurityGroupName}", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client SecurityGroupsClient) CreateOrUpdateSender(req *http.Request) (future SecurityGroupsCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client SecurityGroupsClient) CreateOrUpdateResponder(resp *http.Response) (result SecurityGroup, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified network security group. +// Parameters: +// resourceGroupName - the name of the resource group. +// networkSecurityGroupName - the name of the network security group. +func (client SecurityGroupsClient) Delete(ctx context.Context, resourceGroupName string, networkSecurityGroupName string) (result SecurityGroupsDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, networkSecurityGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client SecurityGroupsClient) DeletePreparer(ctx context.Context, resourceGroupName string, networkSecurityGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "networkSecurityGroupName": autorest.Encode("path", networkSecurityGroupName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkSecurityGroups/{networkSecurityGroupName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client SecurityGroupsClient) DeleteSender(req *http.Request) (future SecurityGroupsDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client SecurityGroupsClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified network security group. +// Parameters: +// resourceGroupName - the name of the resource group. +// networkSecurityGroupName - the name of the network security group. +// expand - expands referenced resources. +func (client SecurityGroupsClient) Get(ctx context.Context, resourceGroupName string, networkSecurityGroupName string, expand string) (result SecurityGroup, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, networkSecurityGroupName, expand) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client SecurityGroupsClient) GetPreparer(ctx context.Context, resourceGroupName string, networkSecurityGroupName string, expand string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "networkSecurityGroupName": autorest.Encode("path", networkSecurityGroupName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkSecurityGroups/{networkSecurityGroupName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client SecurityGroupsClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client SecurityGroupsClient) GetResponder(resp *http.Response) (result SecurityGroup, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all network security groups in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +func (client SecurityGroupsClient) List(ctx context.Context, resourceGroupName string) (result SecurityGroupListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.sglr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "List", resp, "Failure sending request") + return + } + + result.sglr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client SecurityGroupsClient) ListPreparer(ctx context.Context, resourceGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkSecurityGroups", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client SecurityGroupsClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client SecurityGroupsClient) ListResponder(resp *http.Response) (result SecurityGroupListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client SecurityGroupsClient) listNextResults(lastResults SecurityGroupListResult) (result SecurityGroupListResult, err error) { + req, err := lastResults.securityGroupListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client SecurityGroupsClient) ListComplete(ctx context.Context, resourceGroupName string) (result SecurityGroupListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName) + return +} + +// ListAll gets all network security groups in a subscription. +func (client SecurityGroupsClient) ListAll(ctx context.Context) (result SecurityGroupListResultPage, err error) { + result.fn = client.listAllNextResults + req, err := client.ListAllPreparer(ctx) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "ListAll", nil, "Failure preparing request") + return + } + + resp, err := client.ListAllSender(req) + if err != nil { + result.sglr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "ListAll", resp, "Failure sending request") + return + } + + result.sglr, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "ListAll", resp, "Failure responding to request") + } + + return +} + +// ListAllPreparer prepares the ListAll request. +func (client SecurityGroupsClient) ListAllPreparer(ctx context.Context) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/Microsoft.Network/networkSecurityGroups", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListAllSender sends the ListAll request. The method will close the +// http.Response Body if it receives an error. +func (client SecurityGroupsClient) ListAllSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListAllResponder handles the response to the ListAll request. The method always +// closes the http.Response Body. +func (client SecurityGroupsClient) ListAllResponder(resp *http.Response) (result SecurityGroupListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listAllNextResults retrieves the next set of results, if any. +func (client SecurityGroupsClient) listAllNextResults(lastResults SecurityGroupListResult) (result SecurityGroupListResult, err error) { + req, err := lastResults.securityGroupListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "listAllNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListAllSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "listAllNextResults", resp, "Failure sending next results request") + } + result, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityGroupsClient", "listAllNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListAllComplete enumerates all values, automatically crossing page boundaries as required. +func (client SecurityGroupsClient) ListAllComplete(ctx context.Context) (result SecurityGroupListResultIterator, err error) { + result.page, err = client.ListAll(ctx) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/securityrules.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/securityrules.go new file mode 100644 index 000000000..1040fe1a3 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/securityrules.go @@ -0,0 +1,358 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/validation" + "net/http" +) + +// SecurityRulesClient is the network Client +type SecurityRulesClient struct { + BaseClient +} + +// NewSecurityRulesClient creates an instance of the SecurityRulesClient client. +func NewSecurityRulesClient(subscriptionID string) SecurityRulesClient { + return NewSecurityRulesClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewSecurityRulesClientWithBaseURI creates an instance of the SecurityRulesClient client. +func NewSecurityRulesClientWithBaseURI(baseURI string, subscriptionID string) SecurityRulesClient { + return SecurityRulesClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a security rule in the specified network security group. +// Parameters: +// resourceGroupName - the name of the resource group. +// networkSecurityGroupName - the name of the network security group. +// securityRuleName - the name of the security rule. +// securityRuleParameters - parameters supplied to the create or update network security rule operation. +func (client SecurityRulesClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, networkSecurityGroupName string, securityRuleName string, securityRuleParameters SecurityRule) (result SecurityRulesCreateOrUpdateFuture, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: securityRuleParameters, + Constraints: []validation.Constraint{{Target: "securityRuleParameters.SecurityRulePropertiesFormat", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "securityRuleParameters.SecurityRulePropertiesFormat.SourceAddressPrefix", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "securityRuleParameters.SecurityRulePropertiesFormat.DestinationAddressPrefix", Name: validation.Null, Rule: true, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("network.SecurityRulesClient", "CreateOrUpdate", err.Error()) + } + + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, networkSecurityGroupName, securityRuleName, securityRuleParameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client SecurityRulesClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, networkSecurityGroupName string, securityRuleName string, securityRuleParameters SecurityRule) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "networkSecurityGroupName": autorest.Encode("path", networkSecurityGroupName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "securityRuleName": autorest.Encode("path", securityRuleName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkSecurityGroups/{networkSecurityGroupName}/securityRules/{securityRuleName}", pathParameters), + autorest.WithJSON(securityRuleParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client SecurityRulesClient) CreateOrUpdateSender(req *http.Request) (future SecurityRulesCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client SecurityRulesClient) CreateOrUpdateResponder(resp *http.Response) (result SecurityRule, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified network security rule. +// Parameters: +// resourceGroupName - the name of the resource group. +// networkSecurityGroupName - the name of the network security group. +// securityRuleName - the name of the security rule. +func (client SecurityRulesClient) Delete(ctx context.Context, resourceGroupName string, networkSecurityGroupName string, securityRuleName string) (result SecurityRulesDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, networkSecurityGroupName, securityRuleName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client SecurityRulesClient) DeletePreparer(ctx context.Context, resourceGroupName string, networkSecurityGroupName string, securityRuleName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "networkSecurityGroupName": autorest.Encode("path", networkSecurityGroupName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "securityRuleName": autorest.Encode("path", securityRuleName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkSecurityGroups/{networkSecurityGroupName}/securityRules/{securityRuleName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client SecurityRulesClient) DeleteSender(req *http.Request) (future SecurityRulesDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client SecurityRulesClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get get the specified network security rule. +// Parameters: +// resourceGroupName - the name of the resource group. +// networkSecurityGroupName - the name of the network security group. +// securityRuleName - the name of the security rule. +func (client SecurityRulesClient) Get(ctx context.Context, resourceGroupName string, networkSecurityGroupName string, securityRuleName string) (result SecurityRule, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, networkSecurityGroupName, securityRuleName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.SecurityRulesClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client SecurityRulesClient) GetPreparer(ctx context.Context, resourceGroupName string, networkSecurityGroupName string, securityRuleName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "networkSecurityGroupName": autorest.Encode("path", networkSecurityGroupName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "securityRuleName": autorest.Encode("path", securityRuleName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkSecurityGroups/{networkSecurityGroupName}/securityRules/{securityRuleName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client SecurityRulesClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client SecurityRulesClient) GetResponder(resp *http.Response) (result SecurityRule, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all security rules in a network security group. +// Parameters: +// resourceGroupName - the name of the resource group. +// networkSecurityGroupName - the name of the network security group. +func (client SecurityRulesClient) List(ctx context.Context, resourceGroupName string, networkSecurityGroupName string) (result SecurityRuleListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName, networkSecurityGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.srlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.SecurityRulesClient", "List", resp, "Failure sending request") + return + } + + result.srlr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client SecurityRulesClient) ListPreparer(ctx context.Context, resourceGroupName string, networkSecurityGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "networkSecurityGroupName": autorest.Encode("path", networkSecurityGroupName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/networkSecurityGroups/{networkSecurityGroupName}/securityRules", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client SecurityRulesClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client SecurityRulesClient) ListResponder(resp *http.Response) (result SecurityRuleListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client SecurityRulesClient) listNextResults(lastResults SecurityRuleListResult) (result SecurityRuleListResult, err error) { + req, err := lastResults.securityRuleListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.SecurityRulesClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.SecurityRulesClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SecurityRulesClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client SecurityRulesClient) ListComplete(ctx context.Context, resourceGroupName string, networkSecurityGroupName string) (result SecurityRuleListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName, networkSecurityGroupName) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/subnets.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/subnets.go new file mode 100644 index 000000000..886a65126 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/subnets.go @@ -0,0 +1,352 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// SubnetsClient is the network Client +type SubnetsClient struct { + BaseClient +} + +// NewSubnetsClient creates an instance of the SubnetsClient client. +func NewSubnetsClient(subscriptionID string) SubnetsClient { + return NewSubnetsClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewSubnetsClientWithBaseURI creates an instance of the SubnetsClient client. +func NewSubnetsClientWithBaseURI(baseURI string, subscriptionID string) SubnetsClient { + return SubnetsClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a subnet in the specified virtual network. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkName - the name of the virtual network. +// subnetName - the name of the subnet. +// subnetParameters - parameters supplied to the create or update subnet operation. +func (client SubnetsClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, virtualNetworkName string, subnetName string, subnetParameters Subnet) (result SubnetsCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, virtualNetworkName, subnetName, subnetParameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client SubnetsClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, virtualNetworkName string, subnetName string, subnetParameters Subnet) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subnetName": autorest.Encode("path", subnetName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkName": autorest.Encode("path", virtualNetworkName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualnetworks/{virtualNetworkName}/subnets/{subnetName}", pathParameters), + autorest.WithJSON(subnetParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client SubnetsClient) CreateOrUpdateSender(req *http.Request) (future SubnetsCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client SubnetsClient) CreateOrUpdateResponder(resp *http.Response) (result Subnet, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified subnet. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkName - the name of the virtual network. +// subnetName - the name of the subnet. +func (client SubnetsClient) Delete(ctx context.Context, resourceGroupName string, virtualNetworkName string, subnetName string) (result SubnetsDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, virtualNetworkName, subnetName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client SubnetsClient) DeletePreparer(ctx context.Context, resourceGroupName string, virtualNetworkName string, subnetName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subnetName": autorest.Encode("path", subnetName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkName": autorest.Encode("path", virtualNetworkName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualnetworks/{virtualNetworkName}/subnets/{subnetName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client SubnetsClient) DeleteSender(req *http.Request) (future SubnetsDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client SubnetsClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified subnet by virtual network and resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkName - the name of the virtual network. +// subnetName - the name of the subnet. +// expand - expands referenced resources. +func (client SubnetsClient) Get(ctx context.Context, resourceGroupName string, virtualNetworkName string, subnetName string, expand string) (result Subnet, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, virtualNetworkName, subnetName, expand) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.SubnetsClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client SubnetsClient) GetPreparer(ctx context.Context, resourceGroupName string, virtualNetworkName string, subnetName string, expand string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subnetName": autorest.Encode("path", subnetName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkName": autorest.Encode("path", virtualNetworkName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualnetworks/{virtualNetworkName}/subnets/{subnetName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client SubnetsClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client SubnetsClient) GetResponder(resp *http.Response) (result Subnet, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all subnets in a virtual network. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkName - the name of the virtual network. +func (client SubnetsClient) List(ctx context.Context, resourceGroupName string, virtualNetworkName string) (result SubnetListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName, virtualNetworkName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.slr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.SubnetsClient", "List", resp, "Failure sending request") + return + } + + result.slr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client SubnetsClient) ListPreparer(ctx context.Context, resourceGroupName string, virtualNetworkName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkName": autorest.Encode("path", virtualNetworkName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualnetworks/{virtualNetworkName}/subnets", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client SubnetsClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client SubnetsClient) ListResponder(resp *http.Response) (result SubnetListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client SubnetsClient) listNextResults(lastResults SubnetListResult) (result SubnetListResult, err error) { + req, err := lastResults.subnetListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.SubnetsClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.SubnetsClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.SubnetsClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client SubnetsClient) ListComplete(ctx context.Context, resourceGroupName string, virtualNetworkName string) (result SubnetListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName, virtualNetworkName) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/usages.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/usages.go new file mode 100644 index 000000000..0bbb27bbc --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/usages.go @@ -0,0 +1,140 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/validation" + "net/http" +) + +// UsagesClient is the network Client +type UsagesClient struct { + BaseClient +} + +// NewUsagesClient creates an instance of the UsagesClient client. +func NewUsagesClient(subscriptionID string) UsagesClient { + return NewUsagesClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewUsagesClientWithBaseURI creates an instance of the UsagesClient client. +func NewUsagesClientWithBaseURI(baseURI string, subscriptionID string) UsagesClient { + return UsagesClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// List lists compute usages for a subscription. +// Parameters: +// location - the location where resource usage is queried. +func (client UsagesClient) List(ctx context.Context, location string) (result UsagesListResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: location, + Constraints: []validation.Constraint{{Target: "location", Name: validation.Pattern, Rule: `^[-\w\._]+$`, Chain: nil}}}}); err != nil { + return result, validation.NewError("network.UsagesClient", "List", err.Error()) + } + + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, location) + if err != nil { + err = autorest.NewErrorWithError(err, "network.UsagesClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.ulr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.UsagesClient", "List", resp, "Failure sending request") + return + } + + result.ulr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.UsagesClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client UsagesClient) ListPreparer(ctx context.Context, location string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "location": autorest.Encode("path", location), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/Microsoft.Network/locations/{location}/usages", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client UsagesClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client UsagesClient) ListResponder(resp *http.Response) (result UsagesListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client UsagesClient) listNextResults(lastResults UsagesListResult) (result UsagesListResult, err error) { + req, err := lastResults.usagesListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.UsagesClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.UsagesClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.UsagesClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client UsagesClient) ListComplete(ctx context.Context, location string) (result UsagesListResultIterator, err error) { + result.page, err = client.List(ctx, location) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/version.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/version.go new file mode 100644 index 000000000..6890ea03f --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/version.go @@ -0,0 +1,30 @@ +package network + +import "github.com/Azure/azure-sdk-for-go/version" + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +// UserAgent returns the UserAgent string to use when sending http.Requests. +func UserAgent() string { + return "Azure-SDK-For-Go/" + version.Number + " network/2015-06-15" +} + +// Version returns the semantic version (see http://semver.org) of the client. +func Version() string { + return version.Number +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworkgatewayconnections.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworkgatewayconnections.go new file mode 100644 index 000000000..67456bf52 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworkgatewayconnections.go @@ -0,0 +1,560 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// VirtualNetworkGatewayConnectionsClient is the network Client +type VirtualNetworkGatewayConnectionsClient struct { + BaseClient +} + +// NewVirtualNetworkGatewayConnectionsClient creates an instance of the VirtualNetworkGatewayConnectionsClient client. +func NewVirtualNetworkGatewayConnectionsClient(subscriptionID string) VirtualNetworkGatewayConnectionsClient { + return NewVirtualNetworkGatewayConnectionsClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewVirtualNetworkGatewayConnectionsClientWithBaseURI creates an instance of the +// VirtualNetworkGatewayConnectionsClient client. +func NewVirtualNetworkGatewayConnectionsClientWithBaseURI(baseURI string, subscriptionID string) VirtualNetworkGatewayConnectionsClient { + return VirtualNetworkGatewayConnectionsClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a virtual network gateway connection in the specified resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkGatewayConnectionName - the name of the virtual network gateway connection. +// parameters - parameters supplied to the create or update virtual network gateway connection operation. +func (client VirtualNetworkGatewayConnectionsClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, virtualNetworkGatewayConnectionName string, parameters VirtualNetworkGatewayConnection) (result VirtualNetworkGatewayConnectionsCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, virtualNetworkGatewayConnectionName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client VirtualNetworkGatewayConnectionsClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, virtualNetworkGatewayConnectionName string, parameters VirtualNetworkGatewayConnection) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkGatewayConnectionName": autorest.Encode("path", virtualNetworkGatewayConnectionName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/connections/{virtualNetworkGatewayConnectionName}", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewayConnectionsClient) CreateOrUpdateSender(req *http.Request) (future VirtualNetworkGatewayConnectionsCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewayConnectionsClient) CreateOrUpdateResponder(resp *http.Response) (result VirtualNetworkGatewayConnection, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified virtual network Gateway connection. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkGatewayConnectionName - the name of the virtual network gateway connection. +func (client VirtualNetworkGatewayConnectionsClient) Delete(ctx context.Context, resourceGroupName string, virtualNetworkGatewayConnectionName string) (result VirtualNetworkGatewayConnectionsDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, virtualNetworkGatewayConnectionName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client VirtualNetworkGatewayConnectionsClient) DeletePreparer(ctx context.Context, resourceGroupName string, virtualNetworkGatewayConnectionName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkGatewayConnectionName": autorest.Encode("path", virtualNetworkGatewayConnectionName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/connections/{virtualNetworkGatewayConnectionName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewayConnectionsClient) DeleteSender(req *http.Request) (future VirtualNetworkGatewayConnectionsDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewayConnectionsClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified virtual network gateway connection by resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkGatewayConnectionName - the name of the virtual network gateway connection. +func (client VirtualNetworkGatewayConnectionsClient) Get(ctx context.Context, resourceGroupName string, virtualNetworkGatewayConnectionName string) (result VirtualNetworkGatewayConnection, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, virtualNetworkGatewayConnectionName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client VirtualNetworkGatewayConnectionsClient) GetPreparer(ctx context.Context, resourceGroupName string, virtualNetworkGatewayConnectionName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkGatewayConnectionName": autorest.Encode("path", virtualNetworkGatewayConnectionName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/connections/{virtualNetworkGatewayConnectionName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewayConnectionsClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewayConnectionsClient) GetResponder(resp *http.Response) (result VirtualNetworkGatewayConnection, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// GetSharedKey the Get VirtualNetworkGatewayConnectionSharedKey operation retrieves information about the specified +// virtual network gateway connection shared key through Network resource provider. +// Parameters: +// resourceGroupName - the name of the resource group. +// connectionSharedKeyName - the virtual network gateway connection shared key name. +func (client VirtualNetworkGatewayConnectionsClient) GetSharedKey(ctx context.Context, resourceGroupName string, connectionSharedKeyName string) (result ConnectionSharedKeyResult, err error) { + req, err := client.GetSharedKeyPreparer(ctx, resourceGroupName, connectionSharedKeyName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "GetSharedKey", nil, "Failure preparing request") + return + } + + resp, err := client.GetSharedKeySender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "GetSharedKey", resp, "Failure sending request") + return + } + + result, err = client.GetSharedKeyResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "GetSharedKey", resp, "Failure responding to request") + } + + return +} + +// GetSharedKeyPreparer prepares the GetSharedKey request. +func (client VirtualNetworkGatewayConnectionsClient) GetSharedKeyPreparer(ctx context.Context, resourceGroupName string, connectionSharedKeyName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "connectionSharedKeyName": autorest.Encode("path", connectionSharedKeyName), + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/connections/{connectionSharedKeyName}/sharedkey", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSharedKeySender sends the GetSharedKey request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewayConnectionsClient) GetSharedKeySender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetSharedKeyResponder handles the response to the GetSharedKey request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewayConnectionsClient) GetSharedKeyResponder(resp *http.Response) (result ConnectionSharedKeyResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List the List VirtualNetworkGatewayConnections operation retrieves all the virtual network gateways connections +// created. +// Parameters: +// resourceGroupName - the name of the resource group. +func (client VirtualNetworkGatewayConnectionsClient) List(ctx context.Context, resourceGroupName string) (result VirtualNetworkGatewayConnectionListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.vngclr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "List", resp, "Failure sending request") + return + } + + result.vngclr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client VirtualNetworkGatewayConnectionsClient) ListPreparer(ctx context.Context, resourceGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/connections", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewayConnectionsClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewayConnectionsClient) ListResponder(resp *http.Response) (result VirtualNetworkGatewayConnectionListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client VirtualNetworkGatewayConnectionsClient) listNextResults(lastResults VirtualNetworkGatewayConnectionListResult) (result VirtualNetworkGatewayConnectionListResult, err error) { + req, err := lastResults.virtualNetworkGatewayConnectionListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client VirtualNetworkGatewayConnectionsClient) ListComplete(ctx context.Context, resourceGroupName string) (result VirtualNetworkGatewayConnectionListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName) + return +} + +// ResetSharedKey the VirtualNetworkGatewayConnectionResetSharedKey operation resets the virtual network gateway +// connection shared key for passed virtual network gateway connection in the specified resource group through Network +// resource provider. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkGatewayConnectionName - the virtual network gateway connection reset shared key Name. +// parameters - parameters supplied to the begin reset virtual network gateway connection shared key operation +// through network resource provider. +func (client VirtualNetworkGatewayConnectionsClient) ResetSharedKey(ctx context.Context, resourceGroupName string, virtualNetworkGatewayConnectionName string, parameters ConnectionResetSharedKey) (result VirtualNetworkGatewayConnectionsResetSharedKeyFuture, err error) { + req, err := client.ResetSharedKeyPreparer(ctx, resourceGroupName, virtualNetworkGatewayConnectionName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "ResetSharedKey", nil, "Failure preparing request") + return + } + + result, err = client.ResetSharedKeySender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "ResetSharedKey", result.Response(), "Failure sending request") + return + } + + return +} + +// ResetSharedKeyPreparer prepares the ResetSharedKey request. +func (client VirtualNetworkGatewayConnectionsClient) ResetSharedKeyPreparer(ctx context.Context, resourceGroupName string, virtualNetworkGatewayConnectionName string, parameters ConnectionResetSharedKey) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkGatewayConnectionName": autorest.Encode("path", virtualNetworkGatewayConnectionName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/connections/{virtualNetworkGatewayConnectionName}/sharedkey/reset", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ResetSharedKeySender sends the ResetSharedKey request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewayConnectionsClient) ResetSharedKeySender(req *http.Request) (future VirtualNetworkGatewayConnectionsResetSharedKeyFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted)) + return +} + +// ResetSharedKeyResponder handles the response to the ResetSharedKey request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewayConnectionsClient) ResetSharedKeyResponder(resp *http.Response) (result ConnectionResetSharedKey, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// SetSharedKey the Put VirtualNetworkGatewayConnectionSharedKey operation sets the virtual network gateway connection +// shared key for passed virtual network gateway connection in the specified resource group through Network resource +// provider. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkGatewayConnectionName - the virtual network gateway connection name. +// parameters - parameters supplied to the Begin Set Virtual Network Gateway conection Shared key operation +// throughNetwork resource provider. +func (client VirtualNetworkGatewayConnectionsClient) SetSharedKey(ctx context.Context, resourceGroupName string, virtualNetworkGatewayConnectionName string, parameters ConnectionSharedKey) (result VirtualNetworkGatewayConnectionsSetSharedKeyFuture, err error) { + req, err := client.SetSharedKeyPreparer(ctx, resourceGroupName, virtualNetworkGatewayConnectionName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "SetSharedKey", nil, "Failure preparing request") + return + } + + result, err = client.SetSharedKeySender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewayConnectionsClient", "SetSharedKey", result.Response(), "Failure sending request") + return + } + + return +} + +// SetSharedKeyPreparer prepares the SetSharedKey request. +func (client VirtualNetworkGatewayConnectionsClient) SetSharedKeyPreparer(ctx context.Context, resourceGroupName string, virtualNetworkGatewayConnectionName string, parameters ConnectionSharedKey) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkGatewayConnectionName": autorest.Encode("path", virtualNetworkGatewayConnectionName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/connections/{virtualNetworkGatewayConnectionName}/sharedkey", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// SetSharedKeySender sends the SetSharedKey request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewayConnectionsClient) SetSharedKeySender(req *http.Request) (future VirtualNetworkGatewayConnectionsSetSharedKeyFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// SetSharedKeyResponder handles the response to the SetSharedKey request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewayConnectionsClient) SetSharedKeyResponder(resp *http.Response) (result ConnectionSharedKey, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworkgateways.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworkgateways.go new file mode 100644 index 000000000..191c7e675 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworkgateways.go @@ -0,0 +1,484 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// VirtualNetworkGatewaysClient is the network Client +type VirtualNetworkGatewaysClient struct { + BaseClient +} + +// NewVirtualNetworkGatewaysClient creates an instance of the VirtualNetworkGatewaysClient client. +func NewVirtualNetworkGatewaysClient(subscriptionID string) VirtualNetworkGatewaysClient { + return NewVirtualNetworkGatewaysClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewVirtualNetworkGatewaysClientWithBaseURI creates an instance of the VirtualNetworkGatewaysClient client. +func NewVirtualNetworkGatewaysClientWithBaseURI(baseURI string, subscriptionID string) VirtualNetworkGatewaysClient { + return VirtualNetworkGatewaysClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a virtual network gateway in the specified resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkGatewayName - the name of the virtual network gateway. +// parameters - parameters supplied to create or update virtual network gateway operation. +func (client VirtualNetworkGatewaysClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, virtualNetworkGatewayName string, parameters VirtualNetworkGateway) (result VirtualNetworkGatewaysCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, virtualNetworkGatewayName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client VirtualNetworkGatewaysClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, virtualNetworkGatewayName string, parameters VirtualNetworkGateway) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkGatewayName": autorest.Encode("path", virtualNetworkGatewayName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualNetworkGateways/{virtualNetworkGatewayName}", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewaysClient) CreateOrUpdateSender(req *http.Request) (future VirtualNetworkGatewaysCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewaysClient) CreateOrUpdateResponder(resp *http.Response) (result VirtualNetworkGateway, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified virtual network gateway. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkGatewayName - the name of the virtual network gateway. +func (client VirtualNetworkGatewaysClient) Delete(ctx context.Context, resourceGroupName string, virtualNetworkGatewayName string) (result VirtualNetworkGatewaysDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, virtualNetworkGatewayName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client VirtualNetworkGatewaysClient) DeletePreparer(ctx context.Context, resourceGroupName string, virtualNetworkGatewayName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkGatewayName": autorest.Encode("path", virtualNetworkGatewayName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualNetworkGateways/{virtualNetworkGatewayName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewaysClient) DeleteSender(req *http.Request) (future VirtualNetworkGatewaysDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewaysClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Generatevpnclientpackage generates VPN client package for P2S client of the virtual network gateway in the specified +// resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkGatewayName - the name of the virtual network gateway. +// parameters - parameters supplied to the generate virtual network gateway VPN client package operation. +func (client VirtualNetworkGatewaysClient) Generatevpnclientpackage(ctx context.Context, resourceGroupName string, virtualNetworkGatewayName string, parameters VpnClientParameters) (result String, err error) { + req, err := client.GeneratevpnclientpackagePreparer(ctx, resourceGroupName, virtualNetworkGatewayName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "Generatevpnclientpackage", nil, "Failure preparing request") + return + } + + resp, err := client.GeneratevpnclientpackageSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "Generatevpnclientpackage", resp, "Failure sending request") + return + } + + result, err = client.GeneratevpnclientpackageResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "Generatevpnclientpackage", resp, "Failure responding to request") + } + + return +} + +// GeneratevpnclientpackagePreparer prepares the Generatevpnclientpackage request. +func (client VirtualNetworkGatewaysClient) GeneratevpnclientpackagePreparer(ctx context.Context, resourceGroupName string, virtualNetworkGatewayName string, parameters VpnClientParameters) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkGatewayName": autorest.Encode("path", virtualNetworkGatewayName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualNetworkGateways/{virtualNetworkGatewayName}/generatevpnclientpackage", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GeneratevpnclientpackageSender sends the Generatevpnclientpackage request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewaysClient) GeneratevpnclientpackageSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GeneratevpnclientpackageResponder handles the response to the Generatevpnclientpackage request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewaysClient) GeneratevpnclientpackageResponder(resp *http.Response) (result String, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByUnmarshallingJSON(&result.Value), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Get gets the specified virtual network gateway by resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkGatewayName - the name of the virtual network gateway. +func (client VirtualNetworkGatewaysClient) Get(ctx context.Context, resourceGroupName string, virtualNetworkGatewayName string) (result VirtualNetworkGateway, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, virtualNetworkGatewayName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client VirtualNetworkGatewaysClient) GetPreparer(ctx context.Context, resourceGroupName string, virtualNetworkGatewayName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkGatewayName": autorest.Encode("path", virtualNetworkGatewayName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualNetworkGateways/{virtualNetworkGatewayName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewaysClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewaysClient) GetResponder(resp *http.Response) (result VirtualNetworkGateway, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all virtual network gateways by resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +func (client VirtualNetworkGatewaysClient) List(ctx context.Context, resourceGroupName string) (result VirtualNetworkGatewayListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.vnglr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "List", resp, "Failure sending request") + return + } + + result.vnglr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client VirtualNetworkGatewaysClient) ListPreparer(ctx context.Context, resourceGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualNetworkGateways", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewaysClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewaysClient) ListResponder(resp *http.Response) (result VirtualNetworkGatewayListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client VirtualNetworkGatewaysClient) listNextResults(lastResults VirtualNetworkGatewayListResult) (result VirtualNetworkGatewayListResult, err error) { + req, err := lastResults.virtualNetworkGatewayListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client VirtualNetworkGatewaysClient) ListComplete(ctx context.Context, resourceGroupName string) (result VirtualNetworkGatewayListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName) + return +} + +// Reset resets the primary of the virtual network gateway in the specified resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkGatewayName - the name of the virtual network gateway. +// parameters - virtual network gateway vip address supplied to the begin reset of the active-active feature +// enabled gateway. +func (client VirtualNetworkGatewaysClient) Reset(ctx context.Context, resourceGroupName string, virtualNetworkGatewayName string, parameters VirtualNetworkGateway) (result VirtualNetworkGatewaysResetFuture, err error) { + req, err := client.ResetPreparer(ctx, resourceGroupName, virtualNetworkGatewayName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "Reset", nil, "Failure preparing request") + return + } + + result, err = client.ResetSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworkGatewaysClient", "Reset", result.Response(), "Failure sending request") + return + } + + return +} + +// ResetPreparer prepares the Reset request. +func (client VirtualNetworkGatewaysClient) ResetPreparer(ctx context.Context, resourceGroupName string, virtualNetworkGatewayName string, parameters VirtualNetworkGateway) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkGatewayName": autorest.Encode("path", virtualNetworkGatewayName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualNetworkGateways/{virtualNetworkGatewayName}/reset", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ResetSender sends the Reset request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworkGatewaysClient) ResetSender(req *http.Request) (future VirtualNetworkGatewaysResetFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted)) + return +} + +// ResetResponder handles the response to the Reset request. The method always +// closes the http.Response Body. +func (client VirtualNetworkGatewaysClient) ResetResponder(resp *http.Response) (result VirtualNetworkGateway, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworks.go b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworks.go new file mode 100644 index 000000000..0ad02f2a0 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network/virtualnetworks.go @@ -0,0 +1,434 @@ +package network + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "net/http" +) + +// VirtualNetworksClient is the network Client +type VirtualNetworksClient struct { + BaseClient +} + +// NewVirtualNetworksClient creates an instance of the VirtualNetworksClient client. +func NewVirtualNetworksClient(subscriptionID string) VirtualNetworksClient { + return NewVirtualNetworksClientWithBaseURI(DefaultBaseURI, subscriptionID) +} + +// NewVirtualNetworksClientWithBaseURI creates an instance of the VirtualNetworksClient client. +func NewVirtualNetworksClientWithBaseURI(baseURI string, subscriptionID string) VirtualNetworksClient { + return VirtualNetworksClient{NewWithBaseURI(baseURI, subscriptionID)} +} + +// CreateOrUpdate creates or updates a virtual network in the specified resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkName - the name of the virtual network. +// parameters - parameters supplied to the create or update virtual network operation +func (client VirtualNetworksClient) CreateOrUpdate(ctx context.Context, resourceGroupName string, virtualNetworkName string, parameters VirtualNetwork) (result VirtualNetworksCreateOrUpdateFuture, err error) { + req, err := client.CreateOrUpdatePreparer(ctx, resourceGroupName, virtualNetworkName, parameters) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "CreateOrUpdate", nil, "Failure preparing request") + return + } + + result, err = client.CreateOrUpdateSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "CreateOrUpdate", result.Response(), "Failure sending request") + return + } + + return +} + +// CreateOrUpdatePreparer prepares the CreateOrUpdate request. +func (client VirtualNetworksClient) CreateOrUpdatePreparer(ctx context.Context, resourceGroupName string, virtualNetworkName string, parameters VirtualNetwork) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkName": autorest.Encode("path", virtualNetworkName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualNetworks/{virtualNetworkName}", pathParameters), + autorest.WithJSON(parameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CreateOrUpdateSender sends the CreateOrUpdate request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworksClient) CreateOrUpdateSender(req *http.Request) (future VirtualNetworksCreateOrUpdateFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated)) + return +} + +// CreateOrUpdateResponder handles the response to the CreateOrUpdate request. The method always +// closes the http.Response Body. +func (client VirtualNetworksClient) CreateOrUpdateResponder(resp *http.Response) (result VirtualNetwork, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete deletes the specified virtual network. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkName - the name of the virtual network. +func (client VirtualNetworksClient) Delete(ctx context.Context, resourceGroupName string, virtualNetworkName string) (result VirtualNetworksDeleteFuture, err error) { + req, err := client.DeletePreparer(ctx, resourceGroupName, virtualNetworkName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "Delete", nil, "Failure preparing request") + return + } + + result, err = client.DeleteSender(req) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "Delete", result.Response(), "Failure sending request") + return + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client VirtualNetworksClient) DeletePreparer(ctx context.Context, resourceGroupName string, virtualNetworkName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkName": autorest.Encode("path", virtualNetworkName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualNetworks/{virtualNetworkName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworksClient) DeleteSender(req *http.Request) (future VirtualNetworksDeleteFuture, err error) { + sender := autorest.DecorateSender(client, azure.DoRetryWithRegistration(client.Client)) + future.Future = azure.NewFuture(req) + future.req = req + _, err = future.Done(sender) + if err != nil { + return + } + err = autorest.Respond(future.Response(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent)) + return +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client VirtualNetworksClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets the specified virtual network by resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +// virtualNetworkName - the name of the virtual network. +// expand - expands referenced resources. +func (client VirtualNetworksClient) Get(ctx context.Context, resourceGroupName string, virtualNetworkName string, expand string) (result VirtualNetwork, err error) { + req, err := client.GetPreparer(ctx, resourceGroupName, virtualNetworkName, expand) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client VirtualNetworksClient) GetPreparer(ctx context.Context, resourceGroupName string, virtualNetworkName string, expand string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + "virtualNetworkName": autorest.Encode("path", virtualNetworkName), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualNetworks/{virtualNetworkName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworksClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client VirtualNetworksClient) GetResponder(resp *http.Response) (result VirtualNetwork, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List gets all virtual networks in a resource group. +// Parameters: +// resourceGroupName - the name of the resource group. +func (client VirtualNetworksClient) List(ctx context.Context, resourceGroupName string) (result VirtualNetworkListResultPage, err error) { + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, resourceGroupName) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.vnlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "List", resp, "Failure sending request") + return + } + + result.vnlr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client VirtualNetworksClient) ListPreparer(ctx context.Context, resourceGroupName string) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "resourceGroupName": autorest.Encode("path", resourceGroupName), + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualnetworks", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworksClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client VirtualNetworksClient) ListResponder(resp *http.Response) (result VirtualNetworkListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client VirtualNetworksClient) listNextResults(lastResults VirtualNetworkListResult) (result VirtualNetworkListResult, err error) { + req, err := lastResults.virtualNetworkListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client VirtualNetworksClient) ListComplete(ctx context.Context, resourceGroupName string) (result VirtualNetworkListResultIterator, err error) { + result.page, err = client.List(ctx, resourceGroupName) + return +} + +// ListAll gets all virtual networks in a subscription. +func (client VirtualNetworksClient) ListAll(ctx context.Context) (result VirtualNetworkListResultPage, err error) { + result.fn = client.listAllNextResults + req, err := client.ListAllPreparer(ctx) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "ListAll", nil, "Failure preparing request") + return + } + + resp, err := client.ListAllSender(req) + if err != nil { + result.vnlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "ListAll", resp, "Failure sending request") + return + } + + result.vnlr, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "ListAll", resp, "Failure responding to request") + } + + return +} + +// ListAllPreparer prepares the ListAll request. +func (client VirtualNetworksClient) ListAllPreparer(ctx context.Context) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "subscriptionId": autorest.Encode("path", client.SubscriptionID), + } + + const APIVersion = "2015-06-15" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/Microsoft.Network/virtualnetworks", pathParameters), + autorest.WithQueryParameters(queryParameters)) + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListAllSender sends the ListAll request. The method will close the +// http.Response Body if it receives an error. +func (client VirtualNetworksClient) ListAllSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + azure.DoRetryWithRegistration(client.Client)) +} + +// ListAllResponder handles the response to the ListAll request. The method always +// closes the http.Response Body. +func (client VirtualNetworksClient) ListAllResponder(resp *http.Response) (result VirtualNetworkListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listAllNextResults retrieves the next set of results, if any. +func (client VirtualNetworksClient) listAllNextResults(lastResults VirtualNetworkListResult) (result VirtualNetworkListResult, err error) { + req, err := lastResults.virtualNetworkListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "listAllNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListAllSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "listAllNextResults", resp, "Failure sending next results request") + } + result, err = client.ListAllResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "network.VirtualNetworksClient", "listAllNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListAllComplete enumerates all values, automatically crossing page boundaries as required. +func (client VirtualNetworksClient) ListAllComplete(ctx context.Context) (result VirtualNetworkListResultIterator, err error) { + result.page, err = client.ListAll(ctx) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/version/version.go b/vendor/github.com/Azure/azure-sdk-for-go/version/version.go new file mode 100644 index 000000000..3d022e441 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/version/version.go @@ -0,0 +1,21 @@ +package version + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +// Number contains the semantic version of this SDK. +const Number = "v16.1.0" diff --git a/vendor/github.com/Azure/go-ansiterm/csi_entry_state.go b/vendor/github.com/Azure/go-ansiterm/csi_entry_state.go index 1bd6057da..bcbe00d0c 100644 --- a/vendor/github.com/Azure/go-ansiterm/csi_entry_state.go +++ b/vendor/github.com/Azure/go-ansiterm/csi_entry_state.go @@ -5,7 +5,7 @@ type csiEntryState struct { } func (csiState csiEntryState) Handle(b byte) (s state, e error) { - logger.Infof("CsiEntry::Handle %#x", b) + csiState.parser.logf("CsiEntry::Handle %#x", b) nextState, err := csiState.baseState.Handle(b) if nextState != nil || err != nil { @@ -25,7 +25,7 @@ func (csiState csiEntryState) Handle(b byte) (s state, e error) { } func (csiState csiEntryState) Transition(s state) error { - logger.Infof("CsiEntry::Transition %s --> %s", csiState.Name(), s.Name()) + csiState.parser.logf("CsiEntry::Transition %s --> %s", csiState.Name(), s.Name()) csiState.baseState.Transition(s) switch s { diff --git a/vendor/github.com/Azure/go-ansiterm/csi_param_state.go b/vendor/github.com/Azure/go-ansiterm/csi_param_state.go index 4be35c5fd..7ed5e01c3 100644 --- a/vendor/github.com/Azure/go-ansiterm/csi_param_state.go +++ b/vendor/github.com/Azure/go-ansiterm/csi_param_state.go @@ -5,7 +5,7 @@ type csiParamState struct { } func (csiState csiParamState) Handle(b byte) (s state, e error) { - logger.Infof("CsiParam::Handle %#x", b) + csiState.parser.logf("CsiParam::Handle %#x", b) nextState, err := csiState.baseState.Handle(b) if nextState != nil || err != nil { @@ -26,7 +26,7 @@ func (csiState csiParamState) Handle(b byte) (s state, e error) { } func (csiState csiParamState) Transition(s state) error { - logger.Infof("CsiParam::Transition %s --> %s", csiState.Name(), s.Name()) + csiState.parser.logf("CsiParam::Transition %s --> %s", csiState.Name(), s.Name()) csiState.baseState.Transition(s) switch s { diff --git a/vendor/github.com/Azure/go-ansiterm/escape_intermediate_state.go b/vendor/github.com/Azure/go-ansiterm/escape_intermediate_state.go index 2189eb6b6..1c719db9e 100644 --- a/vendor/github.com/Azure/go-ansiterm/escape_intermediate_state.go +++ b/vendor/github.com/Azure/go-ansiterm/escape_intermediate_state.go @@ -5,7 +5,7 @@ type escapeIntermediateState struct { } func (escState escapeIntermediateState) Handle(b byte) (s state, e error) { - logger.Infof("escapeIntermediateState::Handle %#x", b) + escState.parser.logf("escapeIntermediateState::Handle %#x", b) nextState, err := escState.baseState.Handle(b) if nextState != nil || err != nil { return nextState, err @@ -24,7 +24,7 @@ func (escState escapeIntermediateState) Handle(b byte) (s state, e error) { } func (escState escapeIntermediateState) Transition(s state) error { - logger.Infof("escapeIntermediateState::Transition %s --> %s", escState.Name(), s.Name()) + escState.parser.logf("escapeIntermediateState::Transition %s --> %s", escState.Name(), s.Name()) escState.baseState.Transition(s) switch s { diff --git a/vendor/github.com/Azure/go-ansiterm/escape_state.go b/vendor/github.com/Azure/go-ansiterm/escape_state.go index 7b1b9ad3f..6390abd23 100644 --- a/vendor/github.com/Azure/go-ansiterm/escape_state.go +++ b/vendor/github.com/Azure/go-ansiterm/escape_state.go @@ -5,7 +5,7 @@ type escapeState struct { } func (escState escapeState) Handle(b byte) (s state, e error) { - logger.Infof("escapeState::Handle %#x", b) + escState.parser.logf("escapeState::Handle %#x", b) nextState, err := escState.baseState.Handle(b) if nextState != nil || err != nil { return nextState, err @@ -28,7 +28,7 @@ func (escState escapeState) Handle(b byte) (s state, e error) { } func (escState escapeState) Transition(s state) error { - logger.Infof("Escape::Transition %s --> %s", escState.Name(), s.Name()) + escState.parser.logf("Escape::Transition %s --> %s", escState.Name(), s.Name()) escState.baseState.Transition(s) switch s { diff --git a/vendor/github.com/Azure/go-ansiterm/osc_string_state.go b/vendor/github.com/Azure/go-ansiterm/osc_string_state.go index 24062d420..593b10ab6 100644 --- a/vendor/github.com/Azure/go-ansiterm/osc_string_state.go +++ b/vendor/github.com/Azure/go-ansiterm/osc_string_state.go @@ -5,7 +5,7 @@ type oscStringState struct { } func (oscState oscStringState) Handle(b byte) (s state, e error) { - logger.Infof("OscString::Handle %#x", b) + oscState.parser.logf("OscString::Handle %#x", b) nextState, err := oscState.baseState.Handle(b) if nextState != nil || err != nil { return nextState, err diff --git a/vendor/github.com/Azure/go-ansiterm/parser.go b/vendor/github.com/Azure/go-ansiterm/parser.go index 169f68dbe..03cec7ada 100644 --- a/vendor/github.com/Azure/go-ansiterm/parser.go +++ b/vendor/github.com/Azure/go-ansiterm/parser.go @@ -2,14 +2,10 @@ package ansiterm import ( "errors" - "io/ioutil" + "log" "os" - - "github.com/Sirupsen/logrus" ) -var logger *logrus.Logger - type AnsiParser struct { currState state eventHandler AnsiEventHandler @@ -23,50 +19,69 @@ type AnsiParser struct { ground state oscString state stateMap []state + + logf func(string, ...interface{}) } -func CreateParser(initialState string, evtHandler AnsiEventHandler) *AnsiParser { - logFile := ioutil.Discard +type Option func(*AnsiParser) - if isDebugEnv := os.Getenv(LogEnv); isDebugEnv == "1" { - logFile, _ = os.Create("ansiParser.log") +func WithLogf(f func(string, ...interface{})) Option { + return func(ap *AnsiParser) { + ap.logf = f } +} - logger = &logrus.Logger{ - Out: logFile, - Formatter: new(logrus.TextFormatter), - Level: logrus.InfoLevel, - } - - parser := &AnsiParser{ +func CreateParser(initialState string, evtHandler AnsiEventHandler, opts ...Option) *AnsiParser { + ap := &AnsiParser{ eventHandler: evtHandler, context: &ansiContext{}, } - - parser.csiEntry = csiEntryState{baseState{name: "CsiEntry", parser: parser}} - parser.csiParam = csiParamState{baseState{name: "CsiParam", parser: parser}} - parser.dcsEntry = dcsEntryState{baseState{name: "DcsEntry", parser: parser}} - parser.escape = escapeState{baseState{name: "Escape", parser: parser}} - parser.escapeIntermediate = escapeIntermediateState{baseState{name: "EscapeIntermediate", parser: parser}} - parser.error = errorState{baseState{name: "Error", parser: parser}} - parser.ground = groundState{baseState{name: "Ground", parser: parser}} - parser.oscString = oscStringState{baseState{name: "OscString", parser: parser}} - - parser.stateMap = []state{ - parser.csiEntry, - parser.csiParam, - parser.dcsEntry, - parser.escape, - parser.escapeIntermediate, - parser.error, - parser.ground, - parser.oscString, + for _, o := range opts { + o(ap) } - parser.currState = getState(initialState, parser.stateMap) + if isDebugEnv := os.Getenv(LogEnv); isDebugEnv == "1" { + logFile, _ := os.Create("ansiParser.log") + logger := log.New(logFile, "", log.LstdFlags) + if ap.logf != nil { + l := ap.logf + ap.logf = func(s string, v ...interface{}) { + l(s, v...) + logger.Printf(s, v...) + } + } else { + ap.logf = logger.Printf + } + } - logger.Infof("CreateParser: parser %p", parser) - return parser + if ap.logf == nil { + ap.logf = func(string, ...interface{}) {} + } + + ap.csiEntry = csiEntryState{baseState{name: "CsiEntry", parser: ap}} + ap.csiParam = csiParamState{baseState{name: "CsiParam", parser: ap}} + ap.dcsEntry = dcsEntryState{baseState{name: "DcsEntry", parser: ap}} + ap.escape = escapeState{baseState{name: "Escape", parser: ap}} + ap.escapeIntermediate = escapeIntermediateState{baseState{name: "EscapeIntermediate", parser: ap}} + ap.error = errorState{baseState{name: "Error", parser: ap}} + ap.ground = groundState{baseState{name: "Ground", parser: ap}} + ap.oscString = oscStringState{baseState{name: "OscString", parser: ap}} + + ap.stateMap = []state{ + ap.csiEntry, + ap.csiParam, + ap.dcsEntry, + ap.escape, + ap.escapeIntermediate, + ap.error, + ap.ground, + ap.oscString, + } + + ap.currState = getState(initialState, ap.stateMap) + + ap.logf("CreateParser: parser %p", ap) + return ap } func getState(name string, states []state) state { @@ -97,7 +112,7 @@ func (ap *AnsiParser) handle(b byte) error { } if newState == nil { - logger.Warning("newState is nil") + ap.logf("WARNING: newState is nil") return errors.New("New state of 'nil' is invalid.") } @@ -111,23 +126,23 @@ func (ap *AnsiParser) handle(b byte) error { } func (ap *AnsiParser) changeState(newState state) error { - logger.Infof("ChangeState %s --> %s", ap.currState.Name(), newState.Name()) + ap.logf("ChangeState %s --> %s", ap.currState.Name(), newState.Name()) // Exit old state if err := ap.currState.Exit(); err != nil { - logger.Infof("Exit state '%s' failed with : '%v'", ap.currState.Name(), err) + ap.logf("Exit state '%s' failed with : '%v'", ap.currState.Name(), err) return err } // Perform transition action if err := ap.currState.Transition(newState); err != nil { - logger.Infof("Transition from '%s' to '%s' failed with: '%v'", ap.currState.Name(), newState.Name, err) + ap.logf("Transition from '%s' to '%s' failed with: '%v'", ap.currState.Name(), newState.Name, err) return err } // Enter new state if err := newState.Enter(); err != nil { - logger.Infof("Enter state '%s' failed with: '%v'", newState.Name(), err) + ap.logf("Enter state '%s' failed with: '%v'", newState.Name(), err) return err } diff --git a/vendor/github.com/Azure/go-ansiterm/parser_action_helpers.go b/vendor/github.com/Azure/go-ansiterm/parser_action_helpers.go index 8b69a67a5..de0a1f9cd 100644 --- a/vendor/github.com/Azure/go-ansiterm/parser_action_helpers.go +++ b/vendor/github.com/Azure/go-ansiterm/parser_action_helpers.go @@ -27,7 +27,6 @@ func parseParams(bytes []byte) ([]string, error) { params = append(params, s) } - logger.Infof("Parsed params: %v with length: %d", params, len(params)) return params, nil } @@ -37,7 +36,6 @@ func parseCmd(context ansiContext) (string, error) { func getInt(params []string, dflt int) int { i := getInts(params, 1, dflt)[0] - logger.Infof("getInt: %v", i) return i } @@ -60,8 +58,6 @@ func getInts(params []string, minCount int, dflt int) []int { } } - logger.Infof("getInts: %v", ints) - return ints } diff --git a/vendor/github.com/Azure/go-ansiterm/parser_actions.go b/vendor/github.com/Azure/go-ansiterm/parser_actions.go index 58750a2d2..0bb5e51e9 100644 --- a/vendor/github.com/Azure/go-ansiterm/parser_actions.go +++ b/vendor/github.com/Azure/go-ansiterm/parser_actions.go @@ -1,19 +1,15 @@ package ansiterm -import ( - "fmt" -) - func (ap *AnsiParser) collectParam() error { currChar := ap.context.currentChar - logger.Infof("collectParam %#x", currChar) + ap.logf("collectParam %#x", currChar) ap.context.paramBuffer = append(ap.context.paramBuffer, currChar) return nil } func (ap *AnsiParser) collectInter() error { currChar := ap.context.currentChar - logger.Infof("collectInter %#x", currChar) + ap.logf("collectInter %#x", currChar) ap.context.paramBuffer = append(ap.context.interBuffer, currChar) return nil } @@ -21,8 +17,8 @@ func (ap *AnsiParser) collectInter() error { func (ap *AnsiParser) escDispatch() error { cmd, _ := parseCmd(*ap.context) intermeds := ap.context.interBuffer - logger.Infof("escDispatch currentChar: %#x", ap.context.currentChar) - logger.Infof("escDispatch: %v(%v)", cmd, intermeds) + ap.logf("escDispatch currentChar: %#x", ap.context.currentChar) + ap.logf("escDispatch: %v(%v)", cmd, intermeds) switch cmd { case "D": // IND @@ -43,8 +39,9 @@ func (ap *AnsiParser) escDispatch() error { func (ap *AnsiParser) csiDispatch() error { cmd, _ := parseCmd(*ap.context) params, _ := parseParams(ap.context.paramBuffer) + ap.logf("Parsed params: %v with length: %d", params, len(params)) - logger.Infof("csiDispatch: %v(%v)", cmd, params) + ap.logf("csiDispatch: %v(%v)", cmd, params) switch cmd { case "@": @@ -102,7 +99,7 @@ func (ap *AnsiParser) csiDispatch() error { top, bottom := ints[0], ints[1] return ap.eventHandler.DECSTBM(top, bottom) default: - logger.Errorf(fmt.Sprintf("Unsupported CSI command: '%s', with full context: %v", cmd, ap.context)) + ap.logf("ERROR: Unsupported CSI command: '%s', with full context: %v", cmd, ap.context) return nil } diff --git a/vendor/github.com/Azure/go-ansiterm/winterm/ansi.go b/vendor/github.com/Azure/go-ansiterm/winterm/ansi.go index daf2f0696..a67327972 100644 --- a/vendor/github.com/Azure/go-ansiterm/winterm/ansi.go +++ b/vendor/github.com/Azure/go-ansiterm/winterm/ansi.go @@ -175,7 +175,7 @@ func GetStdFile(nFile int) (*os.File, uintptr) { fd, err := syscall.GetStdHandle(nFile) if err != nil { - panic(fmt.Errorf("Invalid standard handle indentifier: %v -- %v", nFile, err)) + panic(fmt.Errorf("Invalid standard handle identifier: %v -- %v", nFile, err)) } return file, uintptr(fd) diff --git a/vendor/github.com/Azure/go-ansiterm/winterm/api.go b/vendor/github.com/Azure/go-ansiterm/winterm/api.go index 462d92f8e..6055e33b9 100644 --- a/vendor/github.com/Azure/go-ansiterm/winterm/api.go +++ b/vendor/github.com/Azure/go-ansiterm/winterm/api.go @@ -49,17 +49,22 @@ var ( const ( // Console modes // See https://msdn.microsoft.com/en-us/library/windows/desktop/ms686033(v=vs.85).aspx. - ENABLE_PROCESSED_INPUT = 0x0001 - ENABLE_LINE_INPUT = 0x0002 - ENABLE_ECHO_INPUT = 0x0004 - ENABLE_WINDOW_INPUT = 0x0008 - ENABLE_MOUSE_INPUT = 0x0010 - ENABLE_INSERT_MODE = 0x0020 - ENABLE_QUICK_EDIT_MODE = 0x0040 - ENABLE_EXTENDED_FLAGS = 0x0080 + ENABLE_PROCESSED_INPUT = 0x0001 + ENABLE_LINE_INPUT = 0x0002 + ENABLE_ECHO_INPUT = 0x0004 + ENABLE_WINDOW_INPUT = 0x0008 + ENABLE_MOUSE_INPUT = 0x0010 + ENABLE_INSERT_MODE = 0x0020 + ENABLE_QUICK_EDIT_MODE = 0x0040 + ENABLE_EXTENDED_FLAGS = 0x0080 + ENABLE_AUTO_POSITION = 0x0100 + ENABLE_VIRTUAL_TERMINAL_INPUT = 0x0200 - ENABLE_PROCESSED_OUTPUT = 0x0001 - ENABLE_WRAP_AT_EOL_OUTPUT = 0x0002 + ENABLE_PROCESSED_OUTPUT = 0x0001 + ENABLE_WRAP_AT_EOL_OUTPUT = 0x0002 + ENABLE_VIRTUAL_TERMINAL_PROCESSING = 0x0004 + DISABLE_NEWLINE_AUTO_RETURN = 0x0008 + ENABLE_LVB_GRID_WORLDWIDE = 0x0010 // Character attributes // Note: diff --git a/vendor/github.com/Azure/go-ansiterm/winterm/cursor_helpers.go b/vendor/github.com/Azure/go-ansiterm/winterm/cursor_helpers.go index f015723ad..3ee06ea72 100644 --- a/vendor/github.com/Azure/go-ansiterm/winterm/cursor_helpers.go +++ b/vendor/github.com/Azure/go-ansiterm/winterm/cursor_helpers.go @@ -34,7 +34,7 @@ func (h *windowsAnsiEventHandler) setCursorPosition(position COORD, window SMALL if err != nil { return err } - logger.Infof("Cursor position set: (%d, %d)", position.X, position.Y) + h.logf("Cursor position set: (%d, %d)", position.X, position.Y) return err } diff --git a/vendor/github.com/Azure/go-ansiterm/winterm/scroll_helper.go b/vendor/github.com/Azure/go-ansiterm/winterm/scroll_helper.go index 706d27057..2d27fa1d0 100644 --- a/vendor/github.com/Azure/go-ansiterm/winterm/scroll_helper.go +++ b/vendor/github.com/Azure/go-ansiterm/winterm/scroll_helper.go @@ -50,8 +50,8 @@ func (h *windowsAnsiEventHandler) insertLines(param int) error { // scroll scrolls the provided scroll region by param lines. The scroll region is in buffer coordinates. func (h *windowsAnsiEventHandler) scroll(param int, sr scrollRegion, info *CONSOLE_SCREEN_BUFFER_INFO) error { - logger.Infof("scroll: scrollTop: %d, scrollBottom: %d", sr.top, sr.bottom) - logger.Infof("scroll: windowTop: %d, windowBottom: %d", info.Window.Top, info.Window.Bottom) + h.logf("scroll: scrollTop: %d, scrollBottom: %d", sr.top, sr.bottom) + h.logf("scroll: windowTop: %d, windowBottom: %d", info.Window.Top, info.Window.Bottom) // Copy from and clip to the scroll region (full buffer width) scrollRect := SMALL_RECT{ diff --git a/vendor/github.com/Azure/go-ansiterm/winterm/win_event_handler.go b/vendor/github.com/Azure/go-ansiterm/winterm/win_event_handler.go index 4d858ed61..2d40fb75a 100644 --- a/vendor/github.com/Azure/go-ansiterm/winterm/win_event_handler.go +++ b/vendor/github.com/Azure/go-ansiterm/winterm/win_event_handler.go @@ -4,16 +4,13 @@ package winterm import ( "bytes" - "io/ioutil" + "log" "os" "strconv" "github.com/Azure/go-ansiterm" - "github.com/Sirupsen/logrus" ) -var logger *logrus.Logger - type windowsAnsiEventHandler struct { fd uintptr file *os.File @@ -28,32 +25,52 @@ type windowsAnsiEventHandler struct { marginByte byte curInfo *CONSOLE_SCREEN_BUFFER_INFO curPos COORD + logf func(string, ...interface{}) } -func CreateWinEventHandler(fd uintptr, file *os.File) ansiterm.AnsiEventHandler { - logFile := ioutil.Discard +type Option func(*windowsAnsiEventHandler) - if isDebugEnv := os.Getenv(ansiterm.LogEnv); isDebugEnv == "1" { - logFile, _ = os.Create("winEventHandler.log") - } - - logger = &logrus.Logger{ - Out: logFile, - Formatter: new(logrus.TextFormatter), - Level: logrus.DebugLevel, +func WithLogf(f func(string, ...interface{})) Option { + return func(w *windowsAnsiEventHandler) { + w.logf = f } +} +func CreateWinEventHandler(fd uintptr, file *os.File, opts ...Option) ansiterm.AnsiEventHandler { infoReset, err := GetConsoleScreenBufferInfo(fd) if err != nil { return nil } - return &windowsAnsiEventHandler{ + h := &windowsAnsiEventHandler{ fd: fd, file: file, infoReset: infoReset, attributes: infoReset.Attributes, } + for _, o := range opts { + o(h) + } + + if isDebugEnv := os.Getenv(ansiterm.LogEnv); isDebugEnv == "1" { + logFile, _ := os.Create("winEventHandler.log") + logger := log.New(logFile, "", log.LstdFlags) + if h.logf != nil { + l := h.logf + h.logf = func(s string, v ...interface{}) { + l(s, v...) + logger.Printf(s, v...) + } + } else { + h.logf = logger.Printf + } + } + + if h.logf == nil { + h.logf = func(string, ...interface{}) {} + } + + return h } type scrollRegion struct { @@ -96,7 +113,7 @@ func (h *windowsAnsiEventHandler) simulateLF(includeCR bool) (bool, error) { if err := h.Flush(); err != nil { return false, err } - logger.Info("Simulating LF inside scroll region") + h.logf("Simulating LF inside scroll region") if err := h.scrollUp(1); err != nil { return false, err } @@ -119,7 +136,7 @@ func (h *windowsAnsiEventHandler) simulateLF(includeCR bool) (bool, error) { } else { // The cursor is at the bottom of the screen but outside the scroll // region. Skip the LF. - logger.Info("Simulating LF outside scroll region") + h.logf("Simulating LF outside scroll region") if includeCR { if err := h.Flush(); err != nil { return false, err @@ -151,7 +168,7 @@ func (h *windowsAnsiEventHandler) executeLF() error { if err := h.Flush(); err != nil { return err } - logger.Info("Resetting cursor position for LF without CR") + h.logf("Resetting cursor position for LF without CR") if err := SetConsoleCursorPosition(h.fd, pos); err != nil { return err } @@ -186,7 +203,7 @@ func (h *windowsAnsiEventHandler) Print(b byte) error { func (h *windowsAnsiEventHandler) Execute(b byte) error { switch b { case ansiterm.ANSI_TAB: - logger.Info("Execute(TAB)") + h.logf("Execute(TAB)") // Move to the next tab stop, but preserve auto-wrap if already set. if !h.wrapNext { pos, info, err := h.getCurrentInfo() @@ -269,7 +286,7 @@ func (h *windowsAnsiEventHandler) CUU(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("CUU: [%v]", []string{strconv.Itoa(param)}) + h.logf("CUU: [%v]", []string{strconv.Itoa(param)}) h.clearWrap() return h.moveCursorVertical(-param) } @@ -278,7 +295,7 @@ func (h *windowsAnsiEventHandler) CUD(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("CUD: [%v]", []string{strconv.Itoa(param)}) + h.logf("CUD: [%v]", []string{strconv.Itoa(param)}) h.clearWrap() return h.moveCursorVertical(param) } @@ -287,7 +304,7 @@ func (h *windowsAnsiEventHandler) CUF(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("CUF: [%v]", []string{strconv.Itoa(param)}) + h.logf("CUF: [%v]", []string{strconv.Itoa(param)}) h.clearWrap() return h.moveCursorHorizontal(param) } @@ -296,7 +313,7 @@ func (h *windowsAnsiEventHandler) CUB(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("CUB: [%v]", []string{strconv.Itoa(param)}) + h.logf("CUB: [%v]", []string{strconv.Itoa(param)}) h.clearWrap() return h.moveCursorHorizontal(-param) } @@ -305,7 +322,7 @@ func (h *windowsAnsiEventHandler) CNL(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("CNL: [%v]", []string{strconv.Itoa(param)}) + h.logf("CNL: [%v]", []string{strconv.Itoa(param)}) h.clearWrap() return h.moveCursorLine(param) } @@ -314,7 +331,7 @@ func (h *windowsAnsiEventHandler) CPL(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("CPL: [%v]", []string{strconv.Itoa(param)}) + h.logf("CPL: [%v]", []string{strconv.Itoa(param)}) h.clearWrap() return h.moveCursorLine(-param) } @@ -323,7 +340,7 @@ func (h *windowsAnsiEventHandler) CHA(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("CHA: [%v]", []string{strconv.Itoa(param)}) + h.logf("CHA: [%v]", []string{strconv.Itoa(param)}) h.clearWrap() return h.moveCursorColumn(param) } @@ -332,7 +349,7 @@ func (h *windowsAnsiEventHandler) VPA(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("VPA: [[%d]]", param) + h.logf("VPA: [[%d]]", param) h.clearWrap() info, err := GetConsoleScreenBufferInfo(h.fd) if err != nil { @@ -348,7 +365,7 @@ func (h *windowsAnsiEventHandler) CUP(row int, col int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("CUP: [[%d %d]]", row, col) + h.logf("CUP: [[%d %d]]", row, col) h.clearWrap() info, err := GetConsoleScreenBufferInfo(h.fd) if err != nil { @@ -364,7 +381,7 @@ func (h *windowsAnsiEventHandler) HVP(row int, col int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("HVP: [[%d %d]]", row, col) + h.logf("HVP: [[%d %d]]", row, col) h.clearWrap() return h.CUP(row, col) } @@ -373,7 +390,7 @@ func (h *windowsAnsiEventHandler) DECTCEM(visible bool) error { if err := h.Flush(); err != nil { return err } - logger.Infof("DECTCEM: [%v]", []string{strconv.FormatBool(visible)}) + h.logf("DECTCEM: [%v]", []string{strconv.FormatBool(visible)}) h.clearWrap() return nil } @@ -382,7 +399,7 @@ func (h *windowsAnsiEventHandler) DECOM(enable bool) error { if err := h.Flush(); err != nil { return err } - logger.Infof("DECOM: [%v]", []string{strconv.FormatBool(enable)}) + h.logf("DECOM: [%v]", []string{strconv.FormatBool(enable)}) h.clearWrap() h.originMode = enable return h.CUP(1, 1) @@ -392,7 +409,7 @@ func (h *windowsAnsiEventHandler) DECCOLM(use132 bool) error { if err := h.Flush(); err != nil { return err } - logger.Infof("DECCOLM: [%v]", []string{strconv.FormatBool(use132)}) + h.logf("DECCOLM: [%v]", []string{strconv.FormatBool(use132)}) h.clearWrap() if err := h.ED(2); err != nil { return err @@ -407,7 +424,7 @@ func (h *windowsAnsiEventHandler) DECCOLM(use132 bool) error { } if info.Size.X < targetWidth { if err := SetConsoleScreenBufferSize(h.fd, COORD{targetWidth, info.Size.Y}); err != nil { - logger.Info("set buffer failed:", err) + h.logf("set buffer failed: %v", err) return err } } @@ -415,12 +432,12 @@ func (h *windowsAnsiEventHandler) DECCOLM(use132 bool) error { window.Left = 0 window.Right = targetWidth - 1 if err := SetConsoleWindowInfo(h.fd, true, window); err != nil { - logger.Info("set window failed:", err) + h.logf("set window failed: %v", err) return err } if info.Size.X > targetWidth { if err := SetConsoleScreenBufferSize(h.fd, COORD{targetWidth, info.Size.Y}); err != nil { - logger.Info("set buffer failed:", err) + h.logf("set buffer failed: %v", err) return err } } @@ -431,7 +448,7 @@ func (h *windowsAnsiEventHandler) ED(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("ED: [%v]", []string{strconv.Itoa(param)}) + h.logf("ED: [%v]", []string{strconv.Itoa(param)}) h.clearWrap() // [J -- Erases from the cursor to the end of the screen, including the cursor position. @@ -490,7 +507,7 @@ func (h *windowsAnsiEventHandler) EL(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("EL: [%v]", strconv.Itoa(param)) + h.logf("EL: [%v]", strconv.Itoa(param)) h.clearWrap() // [K -- Erases from the cursor to the end of the line, including the cursor position. @@ -531,7 +548,7 @@ func (h *windowsAnsiEventHandler) IL(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("IL: [%v]", strconv.Itoa(param)) + h.logf("IL: [%v]", strconv.Itoa(param)) h.clearWrap() return h.insertLines(param) } @@ -540,7 +557,7 @@ func (h *windowsAnsiEventHandler) DL(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("DL: [%v]", strconv.Itoa(param)) + h.logf("DL: [%v]", strconv.Itoa(param)) h.clearWrap() return h.deleteLines(param) } @@ -549,7 +566,7 @@ func (h *windowsAnsiEventHandler) ICH(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("ICH: [%v]", strconv.Itoa(param)) + h.logf("ICH: [%v]", strconv.Itoa(param)) h.clearWrap() return h.insertCharacters(param) } @@ -558,7 +575,7 @@ func (h *windowsAnsiEventHandler) DCH(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("DCH: [%v]", strconv.Itoa(param)) + h.logf("DCH: [%v]", strconv.Itoa(param)) h.clearWrap() return h.deleteCharacters(param) } @@ -572,7 +589,7 @@ func (h *windowsAnsiEventHandler) SGR(params []int) error { strings = append(strings, strconv.Itoa(v)) } - logger.Infof("SGR: [%v]", strings) + h.logf("SGR: [%v]", strings) if len(params) <= 0 { h.attributes = h.infoReset.Attributes @@ -606,7 +623,7 @@ func (h *windowsAnsiEventHandler) SU(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("SU: [%v]", []string{strconv.Itoa(param)}) + h.logf("SU: [%v]", []string{strconv.Itoa(param)}) h.clearWrap() return h.scrollUp(param) } @@ -615,13 +632,13 @@ func (h *windowsAnsiEventHandler) SD(param int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("SD: [%v]", []string{strconv.Itoa(param)}) + h.logf("SD: [%v]", []string{strconv.Itoa(param)}) h.clearWrap() return h.scrollDown(param) } func (h *windowsAnsiEventHandler) DA(params []string) error { - logger.Infof("DA: [%v]", params) + h.logf("DA: [%v]", params) // DA cannot be implemented because it must send data on the VT100 input stream, // which is not available to go-ansiterm. return nil @@ -631,7 +648,7 @@ func (h *windowsAnsiEventHandler) DECSTBM(top int, bottom int) error { if err := h.Flush(); err != nil { return err } - logger.Infof("DECSTBM: [%d, %d]", top, bottom) + h.logf("DECSTBM: [%d, %d]", top, bottom) // Windows is 0 indexed, Linux is 1 indexed h.sr.top = int16(top - 1) @@ -646,7 +663,7 @@ func (h *windowsAnsiEventHandler) RI() error { if err := h.Flush(); err != nil { return err } - logger.Info("RI: []") + h.logf("RI: []") h.clearWrap() info, err := GetConsoleScreenBufferInfo(h.fd) @@ -663,21 +680,21 @@ func (h *windowsAnsiEventHandler) RI() error { } func (h *windowsAnsiEventHandler) IND() error { - logger.Info("IND: []") + h.logf("IND: []") return h.executeLF() } func (h *windowsAnsiEventHandler) Flush() error { h.curInfo = nil if h.buffer.Len() > 0 { - logger.Infof("Flush: [%s]", h.buffer.Bytes()) + h.logf("Flush: [%s]", h.buffer.Bytes()) if _, err := h.buffer.WriteTo(h.file); err != nil { return err } } if h.wrapNext && !h.drewMarginByte { - logger.Infof("Flush: drawing margin byte '%c'", h.marginByte) + h.logf("Flush: drawing margin byte '%c'", h.marginByte) info, err := GetConsoleScreenBufferInfo(h.fd) if err != nil { diff --git a/vendor/github.com/Sirupsen/logrus/.gitignore b/vendor/github.com/Sirupsen/logrus/.gitignore deleted file mode 100644 index 66be63a00..000000000 --- a/vendor/github.com/Sirupsen/logrus/.gitignore +++ /dev/null @@ -1 +0,0 @@ -logrus diff --git a/vendor/github.com/Sirupsen/logrus/.travis.yml b/vendor/github.com/Sirupsen/logrus/.travis.yml deleted file mode 100644 index ff23150dc..000000000 --- a/vendor/github.com/Sirupsen/logrus/.travis.yml +++ /dev/null @@ -1,9 +0,0 @@ -language: go -go: - - 1.3 - - 1.4 - - 1.5 - - tip -install: - - go get -t ./... -script: GOMAXPROCS=4 GORACE="halt_on_error=1" go test -race -v ./... diff --git a/vendor/github.com/Sirupsen/logrus/json_formatter.go b/vendor/github.com/Sirupsen/logrus/json_formatter.go deleted file mode 100644 index 2ad6dc5cf..000000000 --- a/vendor/github.com/Sirupsen/logrus/json_formatter.go +++ /dev/null @@ -1,41 +0,0 @@ -package logrus - -import ( - "encoding/json" - "fmt" -) - -type JSONFormatter struct { - // TimestampFormat sets the format used for marshaling timestamps. - TimestampFormat string -} - -func (f *JSONFormatter) Format(entry *Entry) ([]byte, error) { - data := make(Fields, len(entry.Data)+3) - for k, v := range entry.Data { - switch v := v.(type) { - case error: - // Otherwise errors are ignored by `encoding/json` - // https://github.com/Sirupsen/logrus/issues/137 - data[k] = v.Error() - default: - data[k] = v - } - } - prefixFieldClashes(data) - - timestampFormat := f.TimestampFormat - if timestampFormat == "" { - timestampFormat = DefaultTimestampFormat - } - - data["time"] = entry.Time.Format(timestampFormat) - data["msg"] = entry.Message - data["level"] = entry.Level.String() - - serialized, err := json.Marshal(data) - if err != nil { - return nil, fmt.Errorf("Failed to marshal fields to JSON, %v", err) - } - return append(serialized, '\n'), nil -} diff --git a/vendor/github.com/Sirupsen/logrus/logger.go b/vendor/github.com/Sirupsen/logrus/logger.go deleted file mode 100644 index 2fdb23176..000000000 --- a/vendor/github.com/Sirupsen/logrus/logger.go +++ /dev/null @@ -1,212 +0,0 @@ -package logrus - -import ( - "io" - "os" - "sync" -) - -type Logger struct { - // The logs are `io.Copy`'d to this in a mutex. It's common to set this to a - // file, or leave it default which is `os.Stderr`. You can also set this to - // something more adventorous, such as logging to Kafka. - Out io.Writer - // Hooks for the logger instance. These allow firing events based on logging - // levels and log entries. For example, to send errors to an error tracking - // service, log to StatsD or dump the core on fatal errors. - Hooks LevelHooks - // All log entries pass through the formatter before logged to Out. The - // included formatters are `TextFormatter` and `JSONFormatter` for which - // TextFormatter is the default. In development (when a TTY is attached) it - // logs with colors, but to a file it wouldn't. You can easily implement your - // own that implements the `Formatter` interface, see the `README` or included - // formatters for examples. - Formatter Formatter - // The logging level the logger should log at. This is typically (and defaults - // to) `logrus.Info`, which allows Info(), Warn(), Error() and Fatal() to be - // logged. `logrus.Debug` is useful in - Level Level - // Used to sync writing to the log. - mu sync.Mutex -} - -// Creates a new logger. Configuration should be set by changing `Formatter`, -// `Out` and `Hooks` directly on the default logger instance. You can also just -// instantiate your own: -// -// var log = &Logger{ -// Out: os.Stderr, -// Formatter: new(JSONFormatter), -// Hooks: make(LevelHooks), -// Level: logrus.DebugLevel, -// } -// -// It's recommended to make this a global instance called `log`. -func New() *Logger { - return &Logger{ - Out: os.Stderr, - Formatter: new(TextFormatter), - Hooks: make(LevelHooks), - Level: InfoLevel, - } -} - -// Adds a field to the log entry, note that you it doesn't log until you call -// Debug, Print, Info, Warn, Fatal or Panic. It only creates a log entry. -// If you want multiple fields, use `WithFields`. -func (logger *Logger) WithField(key string, value interface{}) *Entry { - return NewEntry(logger).WithField(key, value) -} - -// Adds a struct of fields to the log entry. All it does is call `WithField` for -// each `Field`. -func (logger *Logger) WithFields(fields Fields) *Entry { - return NewEntry(logger).WithFields(fields) -} - -// Add an error as single field to the log entry. All it does is call -// `WithError` for the given `error`. -func (logger *Logger) WithError(err error) *Entry { - return NewEntry(logger).WithError(err) -} - -func (logger *Logger) Debugf(format string, args ...interface{}) { - if logger.Level >= DebugLevel { - NewEntry(logger).Debugf(format, args...) - } -} - -func (logger *Logger) Infof(format string, args ...interface{}) { - if logger.Level >= InfoLevel { - NewEntry(logger).Infof(format, args...) - } -} - -func (logger *Logger) Printf(format string, args ...interface{}) { - NewEntry(logger).Printf(format, args...) -} - -func (logger *Logger) Warnf(format string, args ...interface{}) { - if logger.Level >= WarnLevel { - NewEntry(logger).Warnf(format, args...) - } -} - -func (logger *Logger) Warningf(format string, args ...interface{}) { - if logger.Level >= WarnLevel { - NewEntry(logger).Warnf(format, args...) - } -} - -func (logger *Logger) Errorf(format string, args ...interface{}) { - if logger.Level >= ErrorLevel { - NewEntry(logger).Errorf(format, args...) - } -} - -func (logger *Logger) Fatalf(format string, args ...interface{}) { - if logger.Level >= FatalLevel { - NewEntry(logger).Fatalf(format, args...) - } - os.Exit(1) -} - -func (logger *Logger) Panicf(format string, args ...interface{}) { - if logger.Level >= PanicLevel { - NewEntry(logger).Panicf(format, args...) - } -} - -func (logger *Logger) Debug(args ...interface{}) { - if logger.Level >= DebugLevel { - NewEntry(logger).Debug(args...) - } -} - -func (logger *Logger) Info(args ...interface{}) { - if logger.Level >= InfoLevel { - NewEntry(logger).Info(args...) - } -} - -func (logger *Logger) Print(args ...interface{}) { - NewEntry(logger).Info(args...) -} - -func (logger *Logger) Warn(args ...interface{}) { - if logger.Level >= WarnLevel { - NewEntry(logger).Warn(args...) - } -} - -func (logger *Logger) Warning(args ...interface{}) { - if logger.Level >= WarnLevel { - NewEntry(logger).Warn(args...) - } -} - -func (logger *Logger) Error(args ...interface{}) { - if logger.Level >= ErrorLevel { - NewEntry(logger).Error(args...) - } -} - -func (logger *Logger) Fatal(args ...interface{}) { - if logger.Level >= FatalLevel { - NewEntry(logger).Fatal(args...) - } - os.Exit(1) -} - -func (logger *Logger) Panic(args ...interface{}) { - if logger.Level >= PanicLevel { - NewEntry(logger).Panic(args...) - } -} - -func (logger *Logger) Debugln(args ...interface{}) { - if logger.Level >= DebugLevel { - NewEntry(logger).Debugln(args...) - } -} - -func (logger *Logger) Infoln(args ...interface{}) { - if logger.Level >= InfoLevel { - NewEntry(logger).Infoln(args...) - } -} - -func (logger *Logger) Println(args ...interface{}) { - NewEntry(logger).Println(args...) -} - -func (logger *Logger) Warnln(args ...interface{}) { - if logger.Level >= WarnLevel { - NewEntry(logger).Warnln(args...) - } -} - -func (logger *Logger) Warningln(args ...interface{}) { - if logger.Level >= WarnLevel { - NewEntry(logger).Warnln(args...) - } -} - -func (logger *Logger) Errorln(args ...interface{}) { - if logger.Level >= ErrorLevel { - NewEntry(logger).Errorln(args...) - } -} - -func (logger *Logger) Fatalln(args ...interface{}) { - if logger.Level >= FatalLevel { - NewEntry(logger).Fatalln(args...) - } - os.Exit(1) -} - -func (logger *Logger) Panicln(args ...interface{}) { - if logger.Level >= PanicLevel { - NewEntry(logger).Panicln(args...) - } -} diff --git a/vendor/github.com/Sirupsen/logrus/terminal_bsd.go b/vendor/github.com/Sirupsen/logrus/terminal_bsd.go deleted file mode 100644 index 71f8d67a5..000000000 --- a/vendor/github.com/Sirupsen/logrus/terminal_bsd.go +++ /dev/null @@ -1,9 +0,0 @@ -// +build darwin freebsd openbsd netbsd dragonfly - -package logrus - -import "syscall" - -const ioctlReadTermios = syscall.TIOCGETA - -type Termios syscall.Termios diff --git a/vendor/github.com/Sirupsen/logrus/terminal_notwindows.go b/vendor/github.com/Sirupsen/logrus/terminal_notwindows.go deleted file mode 100644 index b343b3a37..000000000 --- a/vendor/github.com/Sirupsen/logrus/terminal_notwindows.go +++ /dev/null @@ -1,21 +0,0 @@ -// Based on ssh/terminal: -// Copyright 2011 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build linux darwin freebsd openbsd netbsd dragonfly - -package logrus - -import ( - "syscall" - "unsafe" -) - -// IsTerminal returns true if stderr's file descriptor is a terminal. -func IsTerminal() bool { - fd := syscall.Stderr - var termios Termios - _, _, err := syscall.Syscall6(syscall.SYS_IOCTL, uintptr(fd), ioctlReadTermios, uintptr(unsafe.Pointer(&termios)), 0, 0, 0) - return err == 0 -} diff --git a/vendor/github.com/Sirupsen/logrus/terminal_solaris.go b/vendor/github.com/Sirupsen/logrus/terminal_solaris.go deleted file mode 100644 index 3e70bf7bf..000000000 --- a/vendor/github.com/Sirupsen/logrus/terminal_solaris.go +++ /dev/null @@ -1,15 +0,0 @@ -// +build solaris - -package logrus - -import ( - "os" - - "golang.org/x/sys/unix" -) - -// IsTerminal returns true if the given file descriptor is a terminal. -func IsTerminal() bool { - _, err := unix.IoctlGetTermios(int(os.Stdout.Fd()), unix.TCGETA) - return err == nil -} diff --git a/vendor/github.com/Sirupsen/logrus/terminal_windows.go b/vendor/github.com/Sirupsen/logrus/terminal_windows.go deleted file mode 100644 index 0146845d1..000000000 --- a/vendor/github.com/Sirupsen/logrus/terminal_windows.go +++ /dev/null @@ -1,27 +0,0 @@ -// Based on ssh/terminal: -// Copyright 2011 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build windows - -package logrus - -import ( - "syscall" - "unsafe" -) - -var kernel32 = syscall.NewLazyDLL("kernel32.dll") - -var ( - procGetConsoleMode = kernel32.NewProc("GetConsoleMode") -) - -// IsTerminal returns true if stderr's file descriptor is a terminal. -func IsTerminal() bool { - fd := syscall.Stderr - var st uint32 - r, _, e := syscall.Syscall(procGetConsoleMode.Addr(), 2, uintptr(fd), uintptr(unsafe.Pointer(&st)), 0) - return r != 0 && e == 0 -} diff --git a/vendor/github.com/Sirupsen/logrus/text_formatter.go b/vendor/github.com/Sirupsen/logrus/text_formatter.go deleted file mode 100644 index 06ef20233..000000000 --- a/vendor/github.com/Sirupsen/logrus/text_formatter.go +++ /dev/null @@ -1,161 +0,0 @@ -package logrus - -import ( - "bytes" - "fmt" - "runtime" - "sort" - "strings" - "time" -) - -const ( - nocolor = 0 - red = 31 - green = 32 - yellow = 33 - blue = 34 - gray = 37 -) - -var ( - baseTimestamp time.Time - isTerminal bool -) - -func init() { - baseTimestamp = time.Now() - isTerminal = IsTerminal() -} - -func miniTS() int { - return int(time.Since(baseTimestamp) / time.Second) -} - -type TextFormatter struct { - // Set to true to bypass checking for a TTY before outputting colors. - ForceColors bool - - // Force disabling colors. - DisableColors bool - - // Disable timestamp logging. useful when output is redirected to logging - // system that already adds timestamps. - DisableTimestamp bool - - // Enable logging the full timestamp when a TTY is attached instead of just - // the time passed since beginning of execution. - FullTimestamp bool - - // TimestampFormat to use for display when a full timestamp is printed - TimestampFormat string - - // The fields are sorted by default for a consistent output. For applications - // that log extremely frequently and don't use the JSON formatter this may not - // be desired. - DisableSorting bool -} - -func (f *TextFormatter) Format(entry *Entry) ([]byte, error) { - var keys []string = make([]string, 0, len(entry.Data)) - for k := range entry.Data { - keys = append(keys, k) - } - - if !f.DisableSorting { - sort.Strings(keys) - } - - b := &bytes.Buffer{} - - prefixFieldClashes(entry.Data) - - isColorTerminal := isTerminal && (runtime.GOOS != "windows") - isColored := (f.ForceColors || isColorTerminal) && !f.DisableColors - - timestampFormat := f.TimestampFormat - if timestampFormat == "" { - timestampFormat = DefaultTimestampFormat - } - if isColored { - f.printColored(b, entry, keys, timestampFormat) - } else { - if !f.DisableTimestamp { - f.appendKeyValue(b, "time", entry.Time.Format(timestampFormat)) - } - f.appendKeyValue(b, "level", entry.Level.String()) - if entry.Message != "" { - f.appendKeyValue(b, "msg", entry.Message) - } - for _, key := range keys { - f.appendKeyValue(b, key, entry.Data[key]) - } - } - - b.WriteByte('\n') - return b.Bytes(), nil -} - -func (f *TextFormatter) printColored(b *bytes.Buffer, entry *Entry, keys []string, timestampFormat string) { - var levelColor int - switch entry.Level { - case DebugLevel: - levelColor = gray - case WarnLevel: - levelColor = yellow - case ErrorLevel, FatalLevel, PanicLevel: - levelColor = red - default: - levelColor = blue - } - - levelText := strings.ToUpper(entry.Level.String())[0:4] - - if !f.FullTimestamp { - fmt.Fprintf(b, "\x1b[%dm%s\x1b[0m[%04d] %-44s ", levelColor, levelText, miniTS(), entry.Message) - } else { - fmt.Fprintf(b, "\x1b[%dm%s\x1b[0m[%s] %-44s ", levelColor, levelText, entry.Time.Format(timestampFormat), entry.Message) - } - for _, k := range keys { - v := entry.Data[k] - fmt.Fprintf(b, " \x1b[%dm%s\x1b[0m=%+v", levelColor, k, v) - } -} - -func needsQuoting(text string) bool { - for _, ch := range text { - if !((ch >= 'a' && ch <= 'z') || - (ch >= 'A' && ch <= 'Z') || - (ch >= '0' && ch <= '9') || - ch == '-' || ch == '.') { - return false - } - } - return true -} - -func (f *TextFormatter) appendKeyValue(b *bytes.Buffer, key string, value interface{}) { - - b.WriteString(key) - b.WriteByte('=') - - switch value := value.(type) { - case string: - if needsQuoting(value) { - b.WriteString(value) - } else { - fmt.Fprintf(b, "%q", value) - } - case error: - errmsg := value.Error() - if needsQuoting(errmsg) { - b.WriteString(errmsg) - } else { - fmt.Fprintf(b, "%q", value) - } - default: - fmt.Fprint(b, value) - } - - b.WriteByte(' ') -} diff --git a/vendor/github.com/Sirupsen/logrus/writer.go b/vendor/github.com/Sirupsen/logrus/writer.go deleted file mode 100644 index 1e30b1c75..000000000 --- a/vendor/github.com/Sirupsen/logrus/writer.go +++ /dev/null @@ -1,31 +0,0 @@ -package logrus - -import ( - "bufio" - "io" - "runtime" -) - -func (logger *Logger) Writer() *io.PipeWriter { - reader, writer := io.Pipe() - - go logger.writerScanner(reader) - runtime.SetFinalizer(writer, writerFinalizer) - - return writer -} - -func (logger *Logger) writerScanner(reader *io.PipeReader) { - scanner := bufio.NewScanner(reader) - for scanner.Scan() { - logger.Print(scanner.Text()) - } - if err := scanner.Err(); err != nil { - logger.Errorf("Error while reading from Writer: %s", err) - } - reader.Close() -} - -func writerFinalizer(writer *io.PipeWriter) { - writer.Close() -} diff --git a/vendor/github.com/armon/circbuf/.gitignore b/vendor/github.com/armon/circbuf/.gitignore deleted file mode 100644 index 00268614f..000000000 --- a/vendor/github.com/armon/circbuf/.gitignore +++ /dev/null @@ -1,22 +0,0 @@ -# Compiled Object files, Static and Dynamic libs (Shared Objects) -*.o -*.a -*.so - -# Folders -_obj -_test - -# Architecture specific extensions/prefixes -*.[568vq] -[568vq].out - -*.cgo1.go -*.cgo2.c -_cgo_defun.c -_cgo_gotypes.go -_cgo_export.* - -_testmain.go - -*.exe diff --git a/vendor/github.com/armon/circbuf/LICENSE b/vendor/github.com/armon/circbuf/LICENSE old mode 100644 new mode 100755 diff --git a/vendor/github.com/bgentry/speakeasy/.gitignore b/vendor/github.com/bgentry/speakeasy/.gitignore deleted file mode 100644 index 9e1311461..000000000 --- a/vendor/github.com/bgentry/speakeasy/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -example/example -example/example.exe diff --git a/vendor/github.com/containerd/console/LICENSE b/vendor/github.com/containerd/console/LICENSE new file mode 100644 index 000000000..261eeb9e9 --- /dev/null +++ b/vendor/github.com/containerd/console/LICENSE @@ -0,0 +1,201 @@ + Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "[]" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright [yyyy] [name of copyright owner] + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/vendor/github.com/containerd/console/README.md b/vendor/github.com/containerd/console/README.md new file mode 100644 index 000000000..4c56d9d13 --- /dev/null +++ b/vendor/github.com/containerd/console/README.md @@ -0,0 +1,17 @@ +# console + +[![Build Status](https://travis-ci.org/containerd/console.svg?branch=master)](https://travis-ci.org/containerd/console) + +Golang package for dealing with consoles. Light on deps and a simple API. + +## Modifying the current process + +```go +current := console.Current() +defer current.Reset() + +if err := current.SetRaw(); err != nil { +} +ws, err := current.Size() +current.Resize(ws) +``` diff --git a/vendor/github.com/containerd/console/console.go b/vendor/github.com/containerd/console/console.go new file mode 100644 index 000000000..bf2798fda --- /dev/null +++ b/vendor/github.com/containerd/console/console.go @@ -0,0 +1,62 @@ +package console + +import ( + "errors" + "io" + "os" +) + +var ErrNotAConsole = errors.New("provided file is not a console") + +type Console interface { + io.Reader + io.Writer + io.Closer + + // Resize resizes the console to the provided window size + Resize(WinSize) error + // ResizeFrom resizes the calling console to the size of the + // provided console + ResizeFrom(Console) error + // SetRaw sets the console in raw mode + SetRaw() error + // DisableEcho disables echo on the console + DisableEcho() error + // Reset restores the console to its orignal state + Reset() error + // Size returns the window size of the console + Size() (WinSize, error) + // Fd returns the console's file descriptor + Fd() uintptr + // Name returns the console's file name + Name() string +} + +// WinSize specifies the window size of the console +type WinSize struct { + // Height of the console + Height uint16 + // Width of the console + Width uint16 + x uint16 + y uint16 +} + +// Current returns the current processes console +func Current() Console { + c, err := ConsoleFromFile(os.Stdin) + if err != nil { + // stdin should always be a console for the design + // of this function + panic(err) + } + return c +} + +// ConsoleFromFile returns a console using the provided file +func ConsoleFromFile(f *os.File) (Console, error) { + if err := checkConsole(f); err != nil { + return nil, err + } + return newMaster(f) +} diff --git a/vendor/github.com/containerd/console/console_linux.go b/vendor/github.com/containerd/console/console_linux.go new file mode 100644 index 000000000..c96372929 --- /dev/null +++ b/vendor/github.com/containerd/console/console_linux.go @@ -0,0 +1,255 @@ +// +build linux + +package console + +import ( + "io" + "os" + "sync" + + "golang.org/x/sys/unix" +) + +const ( + maxEvents = 128 +) + +// Epoller manages multiple epoll consoles using edge-triggered epoll api so we +// dont have to deal with repeated wake-up of EPOLLER or EPOLLHUP. +// For more details, see: +// - https://github.com/systemd/systemd/pull/4262 +// - https://github.com/moby/moby/issues/27202 +// +// Example usage of Epoller and EpollConsole can be as follow: +// +// epoller, _ := NewEpoller() +// epollConsole, _ := epoller.Add(console) +// go epoller.Wait() +// var ( +// b bytes.Buffer +// wg sync.WaitGroup +// ) +// wg.Add(1) +// go func() { +// io.Copy(&b, epollConsole) +// wg.Done() +// }() +// // perform I/O on the console +// epollConsole.Shutdown(epoller.CloseConsole) +// wg.Wait() +// epollConsole.Close() +type Epoller struct { + efd int + mu sync.Mutex + fdMapping map[int]*EpollConsole +} + +// NewEpoller returns an instance of epoller with a valid epoll fd. +func NewEpoller() (*Epoller, error) { + efd, err := unix.EpollCreate1(unix.EPOLL_CLOEXEC) + if err != nil { + return nil, err + } + return &Epoller{ + efd: efd, + fdMapping: make(map[int]*EpollConsole), + }, nil +} + +// Add creates a epoll console based on the provided console. The console will +// be registered with EPOLLET (i.e. using edge-triggered notification) and its +// file descriptor will be set to non-blocking mode. After this, user should use +// the return console to perform I/O. +func (e *Epoller) Add(console Console) (*EpollConsole, error) { + sysfd := int(console.Fd()) + // Set sysfd to non-blocking mode + if err := unix.SetNonblock(sysfd, true); err != nil { + return nil, err + } + + ev := unix.EpollEvent{ + Events: unix.EPOLLIN | unix.EPOLLOUT | unix.EPOLLRDHUP | unix.EPOLLET, + Fd: int32(sysfd), + } + if err := unix.EpollCtl(e.efd, unix.EPOLL_CTL_ADD, sysfd, &ev); err != nil { + return nil, err + } + ef := &EpollConsole{ + Console: console, + sysfd: sysfd, + readc: sync.NewCond(&sync.Mutex{}), + writec: sync.NewCond(&sync.Mutex{}), + } + e.mu.Lock() + e.fdMapping[sysfd] = ef + e.mu.Unlock() + return ef, nil +} + +// Wait starts the loop to wait for its consoles' notifications and signal +// appropriate console that it can perform I/O. +func (e *Epoller) Wait() error { + events := make([]unix.EpollEvent, maxEvents) + for { + n, err := unix.EpollWait(e.efd, events, -1) + if err != nil { + // EINTR: The call was interrupted by a signal handler before either + // any of the requested events occurred or the timeout expired + if err == unix.EINTR { + continue + } + return err + } + for i := 0; i < n; i++ { + ev := &events[i] + // the console is ready to be read from + if ev.Events&(unix.EPOLLIN|unix.EPOLLHUP|unix.EPOLLERR) != 0 { + if epfile := e.getConsole(int(ev.Fd)); epfile != nil { + epfile.signalRead() + } + } + // the console is ready to be written to + if ev.Events&(unix.EPOLLOUT|unix.EPOLLHUP|unix.EPOLLERR) != 0 { + if epfile := e.getConsole(int(ev.Fd)); epfile != nil { + epfile.signalWrite() + } + } + } + } +} + +// Close unregister the console's file descriptor from epoll interface +func (e *Epoller) CloseConsole(fd int) error { + e.mu.Lock() + defer e.mu.Unlock() + delete(e.fdMapping, fd) + return unix.EpollCtl(e.efd, unix.EPOLL_CTL_DEL, fd, &unix.EpollEvent{}) +} + +func (e *Epoller) getConsole(sysfd int) *EpollConsole { + e.mu.Lock() + f := e.fdMapping[sysfd] + e.mu.Unlock() + return f +} + +// Close the epoll fd +func (e *Epoller) Close() error { + return unix.Close(e.efd) +} + +// EpollConsole acts like a console but register its file descriptor with a +// epoll fd and uses epoll API to perform I/O. +type EpollConsole struct { + Console + readc *sync.Cond + writec *sync.Cond + sysfd int + closed bool +} + +// Read reads up to len(p) bytes into p. It returns the number of bytes read +// (0 <= n <= len(p)) and any error encountered. +// +// If the console's read returns EAGAIN or EIO, we assumes that its a +// temporary error because the other side went away and wait for the signal +// generated by epoll event to continue. +func (ec *EpollConsole) Read(p []byte) (n int, err error) { + var read int + ec.readc.L.Lock() + defer ec.readc.L.Unlock() + for { + read, err = ec.Console.Read(p[n:]) + n += read + if err != nil { + var hangup bool + if perr, ok := err.(*os.PathError); ok { + hangup = (perr.Err == unix.EAGAIN || perr.Err == unix.EIO) + } else { + hangup = (err == unix.EAGAIN || err == unix.EIO) + } + // if the other end disappear, assume this is temporary and wait for the + // signal to continue again. Unless we didnt read anything and the + // console is already marked as closed then we should exit + if hangup && !(n == 0 && len(p) > 0 && ec.closed) { + ec.readc.Wait() + continue + } + } + break + } + // if we didnt read anything then return io.EOF to end gracefully + if n == 0 && len(p) > 0 && err == nil { + err = io.EOF + } + // signal for others that we finished the read + ec.readc.Signal() + return n, err +} + +// Writes len(p) bytes from p to the console. It returns the number of bytes +// written from p (0 <= n <= len(p)) and any error encountered that caused +// the write to stop early. +// +// If writes to the console returns EAGAIN or EIO, we assumes that its a +// temporary error because the other side went away and wait for the signal +// generated by epoll event to continue. +func (ec *EpollConsole) Write(p []byte) (n int, err error) { + var written int + ec.writec.L.Lock() + defer ec.writec.L.Unlock() + for { + written, err = ec.Console.Write(p[n:]) + n += written + if err != nil { + var hangup bool + if perr, ok := err.(*os.PathError); ok { + hangup = (perr.Err == unix.EAGAIN || perr.Err == unix.EIO) + } else { + hangup = (err == unix.EAGAIN || err == unix.EIO) + } + // if the other end disappear, assume this is temporary and wait for the + // signal to continue again. + if hangup { + ec.writec.Wait() + continue + } + } + // unrecoverable error, break the loop and return the error + break + } + if n < len(p) && err == nil { + err = io.ErrShortWrite + } + // signal for others that we finished the write + ec.writec.Signal() + return n, err +} + +// Close closed the file descriptor and signal call waiters for this fd. +// It accepts a callback which will be called with the console's fd. The +// callback typically will be used to do further cleanup such as unregister the +// console's fd from the epoll interface. +// User should call Shutdown and wait for all I/O operation to be finished +// before closing the console. +func (ec *EpollConsole) Shutdown(close func(int) error) error { + ec.readc.L.Lock() + defer ec.readc.L.Unlock() + ec.writec.L.Lock() + defer ec.writec.L.Unlock() + + ec.readc.Broadcast() + ec.writec.Broadcast() + ec.closed = true + return close(ec.sysfd) +} + +// signalRead signals that the console is readable. +func (ec *EpollConsole) signalRead() { + ec.readc.Signal() +} + +// signalWrite signals that the console is writable. +func (ec *EpollConsole) signalWrite() { + ec.writec.Signal() +} diff --git a/vendor/github.com/containerd/console/console_unix.go b/vendor/github.com/containerd/console/console_unix.go new file mode 100644 index 000000000..118c8c3ab --- /dev/null +++ b/vendor/github.com/containerd/console/console_unix.go @@ -0,0 +1,142 @@ +// +build darwin freebsd linux solaris + +package console + +import ( + "os" + + "golang.org/x/sys/unix" +) + +// NewPty creates a new pty pair +// The master is returned as the first console and a string +// with the path to the pty slave is returned as the second +func NewPty() (Console, string, error) { + f, err := os.OpenFile("/dev/ptmx", unix.O_RDWR|unix.O_NOCTTY|unix.O_CLOEXEC, 0) + if err != nil { + return nil, "", err + } + slave, err := ptsname(f) + if err != nil { + return nil, "", err + } + if err := unlockpt(f); err != nil { + return nil, "", err + } + m, err := newMaster(f) + if err != nil { + return nil, "", err + } + return m, slave, nil +} + +type master struct { + f *os.File + original *unix.Termios +} + +func (m *master) Read(b []byte) (int, error) { + return m.f.Read(b) +} + +func (m *master) Write(b []byte) (int, error) { + return m.f.Write(b) +} + +func (m *master) Close() error { + return m.f.Close() +} + +func (m *master) Resize(ws WinSize) error { + return tcswinsz(m.f.Fd(), ws) +} + +func (m *master) ResizeFrom(c Console) error { + ws, err := c.Size() + if err != nil { + return err + } + return m.Resize(ws) +} + +func (m *master) Reset() error { + if m.original == nil { + return nil + } + return tcset(m.f.Fd(), m.original) +} + +func (m *master) getCurrent() (unix.Termios, error) { + var termios unix.Termios + if err := tcget(m.f.Fd(), &termios); err != nil { + return unix.Termios{}, err + } + return termios, nil +} + +func (m *master) SetRaw() error { + rawState, err := m.getCurrent() + if err != nil { + return err + } + rawState = cfmakeraw(rawState) + rawState.Oflag = rawState.Oflag | unix.OPOST + return tcset(m.f.Fd(), &rawState) +} + +func (m *master) DisableEcho() error { + rawState, err := m.getCurrent() + if err != nil { + return err + } + rawState.Lflag = rawState.Lflag &^ unix.ECHO + return tcset(m.f.Fd(), &rawState) +} + +func (m *master) Size() (WinSize, error) { + return tcgwinsz(m.f.Fd()) +} + +func (m *master) Fd() uintptr { + return m.f.Fd() +} + +func (m *master) Name() string { + return m.f.Name() +} + +// checkConsole checks if the provided file is a console +func checkConsole(f *os.File) error { + var termios unix.Termios + if tcget(f.Fd(), &termios) != nil { + return ErrNotAConsole + } + return nil +} + +func newMaster(f *os.File) (Console, error) { + m := &master{ + f: f, + } + t, err := m.getCurrent() + if err != nil { + return nil, err + } + m.original = &t + return m, nil +} + +// ClearONLCR sets the necessary tty_ioctl(4)s to ensure that a pty pair +// created by us acts normally. In particular, a not-very-well-known default of +// Linux unix98 ptys is that they have +onlcr by default. While this isn't a +// problem for terminal emulators, because we relay data from the terminal we +// also relay that funky line discipline. +func ClearONLCR(fd uintptr) error { + return setONLCR(fd, false) +} + +// SetONLCR sets the necessary tty_ioctl(4)s to ensure that a pty pair +// created by us acts as intended for a terminal emulator. +func SetONLCR(fd uintptr) error { + return setONLCR(fd, true) +} diff --git a/vendor/github.com/containerd/console/console_windows.go b/vendor/github.com/containerd/console/console_windows.go new file mode 100644 index 000000000..d78a0b841 --- /dev/null +++ b/vendor/github.com/containerd/console/console_windows.go @@ -0,0 +1,200 @@ +package console + +import ( + "fmt" + "os" + + "github.com/pkg/errors" + "golang.org/x/sys/windows" +) + +var ( + vtInputSupported bool + ErrNotImplemented = errors.New("not implemented") +) + +func (m *master) initStdios() { + m.in = windows.Handle(os.Stdin.Fd()) + if err := windows.GetConsoleMode(m.in, &m.inMode); err == nil { + // Validate that windows.ENABLE_VIRTUAL_TERMINAL_INPUT is supported, but do not set it. + if err = windows.SetConsoleMode(m.in, m.inMode|windows.ENABLE_VIRTUAL_TERMINAL_INPUT); err == nil { + vtInputSupported = true + } + // Unconditionally set the console mode back even on failure because SetConsoleMode + // remembers invalid bits on input handles. + windows.SetConsoleMode(m.in, m.inMode) + } else { + fmt.Printf("failed to get console mode for stdin: %v\n", err) + } + + m.out = windows.Handle(os.Stdout.Fd()) + if err := windows.GetConsoleMode(m.out, &m.outMode); err == nil { + if err := windows.SetConsoleMode(m.out, m.outMode|windows.ENABLE_VIRTUAL_TERMINAL_PROCESSING); err == nil { + m.outMode |= windows.ENABLE_VIRTUAL_TERMINAL_PROCESSING + } else { + windows.SetConsoleMode(m.out, m.outMode) + } + } else { + fmt.Printf("failed to get console mode for stdout: %v\n", err) + } + + m.err = windows.Handle(os.Stderr.Fd()) + if err := windows.GetConsoleMode(m.err, &m.errMode); err == nil { + if err := windows.SetConsoleMode(m.err, m.errMode|windows.ENABLE_VIRTUAL_TERMINAL_PROCESSING); err == nil { + m.errMode |= windows.ENABLE_VIRTUAL_TERMINAL_PROCESSING + } else { + windows.SetConsoleMode(m.err, m.errMode) + } + } else { + fmt.Printf("failed to get console mode for stderr: %v\n", err) + } +} + +type master struct { + in windows.Handle + inMode uint32 + + out windows.Handle + outMode uint32 + + err windows.Handle + errMode uint32 +} + +func (m *master) SetRaw() error { + if err := makeInputRaw(m.in, m.inMode); err != nil { + return err + } + + // Set StdOut and StdErr to raw mode, we ignore failures since + // windows.DISABLE_NEWLINE_AUTO_RETURN might not be supported on this version of + // Windows. + + windows.SetConsoleMode(m.out, m.outMode|windows.DISABLE_NEWLINE_AUTO_RETURN) + + windows.SetConsoleMode(m.err, m.errMode|windows.DISABLE_NEWLINE_AUTO_RETURN) + + return nil +} + +func (m *master) Reset() error { + for _, s := range []struct { + fd windows.Handle + mode uint32 + }{ + {m.in, m.inMode}, + {m.out, m.outMode}, + {m.err, m.errMode}, + } { + if err := windows.SetConsoleMode(s.fd, s.mode); err != nil { + return errors.Wrap(err, "unable to restore console mode") + } + } + + return nil +} + +func (m *master) Size() (WinSize, error) { + var info windows.ConsoleScreenBufferInfo + err := windows.GetConsoleScreenBufferInfo(m.out, &info) + if err != nil { + return WinSize{}, errors.Wrap(err, "unable to get console info") + } + + winsize := WinSize{ + Width: uint16(info.Window.Right - info.Window.Left + 1), + Height: uint16(info.Window.Bottom - info.Window.Top + 1), + } + + return winsize, nil +} + +func (m *master) Resize(ws WinSize) error { + return ErrNotImplemented +} + +func (m *master) ResizeFrom(c Console) error { + return ErrNotImplemented +} + +func (m *master) DisableEcho() error { + mode := m.inMode &^ windows.ENABLE_ECHO_INPUT + mode |= windows.ENABLE_PROCESSED_INPUT + mode |= windows.ENABLE_LINE_INPUT + + if err := windows.SetConsoleMode(m.in, mode); err != nil { + return errors.Wrap(err, "unable to set console to disable echo") + } + + return nil +} + +func (m *master) Close() error { + return nil +} + +func (m *master) Read(b []byte) (int, error) { + panic("not implemented on windows") +} + +func (m *master) Write(b []byte) (int, error) { + panic("not implemented on windows") +} + +func (m *master) Fd() uintptr { + return uintptr(m.in) +} + +// on windows, console can only be made from os.Std{in,out,err}, hence there +// isnt a single name here we can use. Return a dummy "console" value in this +// case should be sufficient. +func (m *master) Name() string { + return "console" +} + +// makeInputRaw puts the terminal (Windows Console) connected to the given +// file descriptor into raw mode +func makeInputRaw(fd windows.Handle, mode uint32) error { + // See + // -- https://msdn.microsoft.com/en-us/library/windows/desktop/ms686033(v=vs.85).aspx + // -- https://msdn.microsoft.com/en-us/library/windows/desktop/ms683462(v=vs.85).aspx + + // Disable these modes + mode &^= windows.ENABLE_ECHO_INPUT + mode &^= windows.ENABLE_LINE_INPUT + mode &^= windows.ENABLE_MOUSE_INPUT + mode &^= windows.ENABLE_WINDOW_INPUT + mode &^= windows.ENABLE_PROCESSED_INPUT + + // Enable these modes + mode |= windows.ENABLE_EXTENDED_FLAGS + mode |= windows.ENABLE_INSERT_MODE + mode |= windows.ENABLE_QUICK_EDIT_MODE + + if vtInputSupported { + mode |= windows.ENABLE_VIRTUAL_TERMINAL_INPUT + } + + if err := windows.SetConsoleMode(fd, mode); err != nil { + return errors.Wrap(err, "unable to set console to raw mode") + } + + return nil +} + +func checkConsole(f *os.File) error { + var mode uint32 + if err := windows.GetConsoleMode(windows.Handle(f.Fd()), &mode); err != nil { + return err + } + return nil +} + +func newMaster(f *os.File) (Console, error) { + if f != os.Stdin && f != os.Stdout && f != os.Stderr { + return nil, errors.New("creating a console from a file is not supported on windows") + } + m := &master{} + m.initStdios() + return m, nil +} diff --git a/vendor/github.com/containerd/console/tc_darwin.go b/vendor/github.com/containerd/console/tc_darwin.go new file mode 100644 index 000000000..b102bad74 --- /dev/null +++ b/vendor/github.com/containerd/console/tc_darwin.go @@ -0,0 +1,37 @@ +package console + +import ( + "fmt" + "os" + "unsafe" + + "golang.org/x/sys/unix" +) + +const ( + cmdTcGet = unix.TIOCGETA + cmdTcSet = unix.TIOCSETA +) + +func ioctl(fd, flag, data uintptr) error { + if _, _, err := unix.Syscall(unix.SYS_IOCTL, fd, flag, data); err != 0 { + return err + } + return nil +} + +// unlockpt unlocks the slave pseudoterminal device corresponding to the master pseudoterminal referred to by f. +// unlockpt should be called before opening the slave side of a pty. +func unlockpt(f *os.File) error { + var u int32 + return ioctl(f.Fd(), unix.TIOCPTYUNLK, uintptr(unsafe.Pointer(&u))) +} + +// ptsname retrieves the name of the first available pts for the given master. +func ptsname(f *os.File) (string, error) { + n, err := unix.IoctlGetInt(int(f.Fd()), unix.TIOCPTYGNAME) + if err != nil { + return "", err + } + return fmt.Sprintf("/dev/pts/%d", n), nil +} diff --git a/vendor/github.com/containerd/console/tc_freebsd.go b/vendor/github.com/containerd/console/tc_freebsd.go new file mode 100644 index 000000000..e2a10e441 --- /dev/null +++ b/vendor/github.com/containerd/console/tc_freebsd.go @@ -0,0 +1,29 @@ +package console + +import ( + "fmt" + "os" + + "golang.org/x/sys/unix" +) + +const ( + cmdTcGet = unix.TIOCGETA + cmdTcSet = unix.TIOCSETA +) + +// unlockpt unlocks the slave pseudoterminal device corresponding to the master pseudoterminal referred to by f. +// unlockpt should be called before opening the slave side of a pty. +// This does not exist on FreeBSD, it does not allocate controlling terminals on open +func unlockpt(f *os.File) error { + return nil +} + +// ptsname retrieves the name of the first available pts for the given master. +func ptsname(f *os.File) (string, error) { + n, err := unix.IoctlGetInt(int(f.Fd()), unix.TIOCGPTN) + if err != nil { + return "", err + } + return fmt.Sprintf("/dev/pts/%d", n), nil +} diff --git a/vendor/github.com/containerd/console/tc_linux.go b/vendor/github.com/containerd/console/tc_linux.go new file mode 100644 index 000000000..73330eb71 --- /dev/null +++ b/vendor/github.com/containerd/console/tc_linux.go @@ -0,0 +1,33 @@ +package console + +import ( + "fmt" + "os" + "unsafe" + + "golang.org/x/sys/unix" +) + +const ( + cmdTcGet = unix.TCGETS + cmdTcSet = unix.TCSETS +) + +// unlockpt unlocks the slave pseudoterminal device corresponding to the master pseudoterminal referred to by f. +// unlockpt should be called before opening the slave side of a pty. +func unlockpt(f *os.File) error { + var u int32 + if _, _, err := unix.Syscall(unix.SYS_IOCTL, f.Fd(), unix.TIOCSPTLCK, uintptr(unsafe.Pointer(&u))); err != 0 { + return err + } + return nil +} + +// ptsname retrieves the name of the first available pts for the given master. +func ptsname(f *os.File) (string, error) { + var u uint32 + if _, _, err := unix.Syscall(unix.SYS_IOCTL, f.Fd(), unix.TIOCGPTN, uintptr(unsafe.Pointer(&u))); err != 0 { + return "", err + } + return fmt.Sprintf("/dev/pts/%d", u), nil +} diff --git a/vendor/github.com/containerd/console/tc_solaris_cgo.go b/vendor/github.com/containerd/console/tc_solaris_cgo.go new file mode 100644 index 000000000..f8066d8e3 --- /dev/null +++ b/vendor/github.com/containerd/console/tc_solaris_cgo.go @@ -0,0 +1,35 @@ +// +build solaris,cgo + +package console + +import ( + "os" + + "golang.org/x/sys/unix" +) + +//#include +import "C" + +const ( + cmdTcGet = unix.TCGETS + cmdTcSet = unix.TCSETS +) + +// ptsname retrieves the name of the first available pts for the given master. +func ptsname(f *os.File) (string, error) { + ptspath, err := C.ptsname(C.int(f.Fd())) + if err != nil { + return "", err + } + return C.GoString(ptspath), nil +} + +// unlockpt unlocks the slave pseudoterminal device corresponding to the master pseudoterminal referred to by f. +// unlockpt should be called before opening the slave side of a pty. +func unlockpt(f *os.File) error { + if _, err := C.grantpt(C.int(f.Fd())); err != nil { + return err + } + return nil +} diff --git a/vendor/github.com/containerd/console/tc_solaris_nocgo.go b/vendor/github.com/containerd/console/tc_solaris_nocgo.go new file mode 100644 index 000000000..0aefa0d2b --- /dev/null +++ b/vendor/github.com/containerd/console/tc_solaris_nocgo.go @@ -0,0 +1,31 @@ +// +build solaris,!cgo + +// +// Implementing the functions below requires cgo support. Non-cgo stubs +// versions are defined below to enable cross-compilation of source code +// that depends on these functions, but the resultant cross-compiled +// binaries cannot actually be used. If the stub function(s) below are +// actually invoked they will display an error message and cause the +// calling process to exit. +// + +package console + +import ( + "os" + + "golang.org/x/sys/unix" +) + +const ( + cmdTcGet = unix.TCGETS + cmdTcSet = unix.TCSETS +) + +func ptsname(f *os.File) (string, error) { + panic("ptsname() support requires cgo.") +} + +func unlockpt(f *os.File) error { + panic("unlockpt() support requires cgo.") +} diff --git a/vendor/github.com/containerd/console/tc_unix.go b/vendor/github.com/containerd/console/tc_unix.go new file mode 100644 index 000000000..df7dcb933 --- /dev/null +++ b/vendor/github.com/containerd/console/tc_unix.go @@ -0,0 +1,75 @@ +// +build darwin freebsd linux solaris + +package console + +import ( + "golang.org/x/sys/unix" +) + +func tcget(fd uintptr, p *unix.Termios) error { + termios, err := unix.IoctlGetTermios(int(fd), cmdTcGet) + if err != nil { + return err + } + *p = *termios + return nil +} + +func tcset(fd uintptr, p *unix.Termios) error { + return unix.IoctlSetTermios(int(fd), cmdTcSet, p) +} + +func tcgwinsz(fd uintptr) (WinSize, error) { + var ws WinSize + + uws, err := unix.IoctlGetWinsize(int(fd), unix.TIOCGWINSZ) + if err != nil { + return ws, err + } + + // Translate from unix.Winsize to console.WinSize + ws.Height = uws.Row + ws.Width = uws.Col + ws.x = uws.Xpixel + ws.y = uws.Ypixel + return ws, nil +} + +func tcswinsz(fd uintptr, ws WinSize) error { + // Translate from console.WinSize to unix.Winsize + + var uws unix.Winsize + uws.Row = ws.Height + uws.Col = ws.Width + uws.Xpixel = ws.x + uws.Ypixel = ws.y + + return unix.IoctlSetWinsize(int(fd), unix.TIOCSWINSZ, &uws) +} + +func setONLCR(fd uintptr, enable bool) error { + var termios unix.Termios + if err := tcget(fd, &termios); err != nil { + return err + } + if enable { + // Set +onlcr so we can act like a real terminal + termios.Oflag |= unix.ONLCR + } else { + // Set -onlcr so we don't have to deal with \r. + termios.Oflag &^= unix.ONLCR + } + return tcset(fd, &termios) +} + +func cfmakeraw(t unix.Termios) unix.Termios { + t.Iflag &^= (unix.IGNBRK | unix.BRKINT | unix.PARMRK | unix.ISTRIP | unix.INLCR | unix.IGNCR | unix.ICRNL | unix.IXON) + t.Oflag &^= unix.OPOST + t.Lflag &^= (unix.ECHO | unix.ECHONL | unix.ICANON | unix.ISIG | unix.IEXTEN) + t.Cflag &^= (unix.CSIZE | unix.PARENB) + t.Cflag &^= unix.CS8 + t.Cc[unix.VMIN] = 1 + t.Cc[unix.VTIME] = 0 + + return t +} diff --git a/vendor/github.com/containerd/continuity/LICENSE b/vendor/github.com/containerd/continuity/LICENSE new file mode 100644 index 000000000..8dada3eda --- /dev/null +++ b/vendor/github.com/containerd/continuity/LICENSE @@ -0,0 +1,201 @@ + Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "{}" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright {yyyy} {name of copyright owner} + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/vendor/github.com/containerd/continuity/pathdriver/path_driver.go b/vendor/github.com/containerd/continuity/pathdriver/path_driver.go new file mode 100644 index 000000000..b43d55fe9 --- /dev/null +++ b/vendor/github.com/containerd/continuity/pathdriver/path_driver.go @@ -0,0 +1,85 @@ +package pathdriver + +import ( + "path/filepath" +) + +// PathDriver provides all of the path manipulation functions in a common +// interface. The context should call these and never use the `filepath` +// package or any other package to manipulate paths. +type PathDriver interface { + Join(paths ...string) string + IsAbs(path string) bool + Rel(base, target string) (string, error) + Base(path string) string + Dir(path string) string + Clean(path string) string + Split(path string) (dir, file string) + Separator() byte + Abs(path string) (string, error) + Walk(string, filepath.WalkFunc) error + FromSlash(path string) string + ToSlash(path string) string + Match(pattern, name string) (matched bool, err error) +} + +// pathDriver is a simple default implementation calls the filepath package. +type pathDriver struct{} + +// LocalPathDriver is the exported pathDriver struct for convenience. +var LocalPathDriver PathDriver = &pathDriver{} + +func (*pathDriver) Join(paths ...string) string { + return filepath.Join(paths...) +} + +func (*pathDriver) IsAbs(path string) bool { + return filepath.IsAbs(path) +} + +func (*pathDriver) Rel(base, target string) (string, error) { + return filepath.Rel(base, target) +} + +func (*pathDriver) Base(path string) string { + return filepath.Base(path) +} + +func (*pathDriver) Dir(path string) string { + return filepath.Dir(path) +} + +func (*pathDriver) Clean(path string) string { + return filepath.Clean(path) +} + +func (*pathDriver) Split(path string) (dir, file string) { + return filepath.Split(path) +} + +func (*pathDriver) Separator() byte { + return filepath.Separator +} + +func (*pathDriver) Abs(path string) (string, error) { + return filepath.Abs(path) +} + +// Note that filepath.Walk calls os.Stat, so if the context wants to +// to call Driver.Stat() for Walk, they need to create a new struct that +// overrides this method. +func (*pathDriver) Walk(root string, walkFn filepath.WalkFunc) error { + return filepath.Walk(root, walkFn) +} + +func (*pathDriver) FromSlash(path string) string { + return filepath.FromSlash(path) +} + +func (*pathDriver) ToSlash(path string) string { + return filepath.ToSlash(path) +} + +func (*pathDriver) Match(pattern, name string) (bool, error) { + return filepath.Match(pattern, name) +} diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/coreos/go-systemd/LICENSE b/vendor/github.com/coreos/go-systemd/LICENSE similarity index 100% rename from vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/coreos/go-systemd/LICENSE rename to vendor/github.com/coreos/go-systemd/LICENSE diff --git a/vendor/github.com/coreos/go-systemd/dbus/dbus.go b/vendor/github.com/coreos/go-systemd/dbus/dbus.go new file mode 100644 index 000000000..c1694fb52 --- /dev/null +++ b/vendor/github.com/coreos/go-systemd/dbus/dbus.go @@ -0,0 +1,213 @@ +// Copyright 2015 CoreOS, Inc. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// Integration with the systemd D-Bus API. See http://www.freedesktop.org/wiki/Software/systemd/dbus/ +package dbus + +import ( + "fmt" + "os" + "strconv" + "strings" + "sync" + + "github.com/godbus/dbus" +) + +const ( + alpha = `abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ` + num = `0123456789` + alphanum = alpha + num + signalBuffer = 100 +) + +// needsEscape checks whether a byte in a potential dbus ObjectPath needs to be escaped +func needsEscape(i int, b byte) bool { + // Escape everything that is not a-z-A-Z-0-9 + // Also escape 0-9 if it's the first character + return strings.IndexByte(alphanum, b) == -1 || + (i == 0 && strings.IndexByte(num, b) != -1) +} + +// PathBusEscape sanitizes a constituent string of a dbus ObjectPath using the +// rules that systemd uses for serializing special characters. +func PathBusEscape(path string) string { + // Special case the empty string + if len(path) == 0 { + return "_" + } + n := []byte{} + for i := 0; i < len(path); i++ { + c := path[i] + if needsEscape(i, c) { + e := fmt.Sprintf("_%x", c) + n = append(n, []byte(e)...) + } else { + n = append(n, c) + } + } + return string(n) +} + +// Conn is a connection to systemd's dbus endpoint. +type Conn struct { + // sysconn/sysobj are only used to call dbus methods + sysconn *dbus.Conn + sysobj dbus.BusObject + + // sigconn/sigobj are only used to receive dbus signals + sigconn *dbus.Conn + sigobj dbus.BusObject + + jobListener struct { + jobs map[dbus.ObjectPath]chan<- string + sync.Mutex + } + subscriber struct { + updateCh chan<- *SubStateUpdate + errCh chan<- error + sync.Mutex + ignore map[dbus.ObjectPath]int64 + cleanIgnore int64 + } +} + +// New establishes a connection to any available bus and authenticates. +// Callers should call Close() when done with the connection. +func New() (*Conn, error) { + conn, err := NewSystemConnection() + if err != nil && os.Geteuid() == 0 { + return NewSystemdConnection() + } + return conn, err +} + +// NewSystemConnection establishes a connection to the system bus and authenticates. +// Callers should call Close() when done with the connection +func NewSystemConnection() (*Conn, error) { + return NewConnection(func() (*dbus.Conn, error) { + return dbusAuthHelloConnection(dbus.SystemBusPrivate) + }) +} + +// NewUserConnection establishes a connection to the session bus and +// authenticates. This can be used to connect to systemd user instances. +// Callers should call Close() when done with the connection. +func NewUserConnection() (*Conn, error) { + return NewConnection(func() (*dbus.Conn, error) { + return dbusAuthHelloConnection(dbus.SessionBusPrivate) + }) +} + +// NewSystemdConnection establishes a private, direct connection to systemd. +// This can be used for communicating with systemd without a dbus daemon. +// Callers should call Close() when done with the connection. +func NewSystemdConnection() (*Conn, error) { + return NewConnection(func() (*dbus.Conn, error) { + // We skip Hello when talking directly to systemd. + return dbusAuthConnection(func() (*dbus.Conn, error) { + return dbus.Dial("unix:path=/run/systemd/private") + }) + }) +} + +// Close closes an established connection +func (c *Conn) Close() { + c.sysconn.Close() + c.sigconn.Close() +} + +// NewConnection establishes a connection to a bus using a caller-supplied function. +// This allows connecting to remote buses through a user-supplied mechanism. +// The supplied function may be called multiple times, and should return independent connections. +// The returned connection must be fully initialised: the org.freedesktop.DBus.Hello call must have succeeded, +// and any authentication should be handled by the function. +func NewConnection(dialBus func() (*dbus.Conn, error)) (*Conn, error) { + sysconn, err := dialBus() + if err != nil { + return nil, err + } + + sigconn, err := dialBus() + if err != nil { + sysconn.Close() + return nil, err + } + + c := &Conn{ + sysconn: sysconn, + sysobj: systemdObject(sysconn), + sigconn: sigconn, + sigobj: systemdObject(sigconn), + } + + c.subscriber.ignore = make(map[dbus.ObjectPath]int64) + c.jobListener.jobs = make(map[dbus.ObjectPath]chan<- string) + + // Setup the listeners on jobs so that we can get completions + c.sigconn.BusObject().Call("org.freedesktop.DBus.AddMatch", 0, + "type='signal', interface='org.freedesktop.systemd1.Manager', member='JobRemoved'") + + c.dispatch() + return c, nil +} + +// GetManagerProperty returns the value of a property on the org.freedesktop.systemd1.Manager +// interface. The value is returned in its string representation, as defined at +// https://developer.gnome.org/glib/unstable/gvariant-text.html +func (c *Conn) GetManagerProperty(prop string) (string, error) { + variant, err := c.sysobj.GetProperty("org.freedesktop.systemd1.Manager." + prop) + if err != nil { + return "", err + } + return variant.String(), nil +} + +func dbusAuthConnection(createBus func() (*dbus.Conn, error)) (*dbus.Conn, error) { + conn, err := createBus() + if err != nil { + return nil, err + } + + // Only use EXTERNAL method, and hardcode the uid (not username) + // to avoid a username lookup (which requires a dynamically linked + // libc) + methods := []dbus.Auth{dbus.AuthExternal(strconv.Itoa(os.Getuid()))} + + err = conn.Auth(methods) + if err != nil { + conn.Close() + return nil, err + } + + return conn, nil +} + +func dbusAuthHelloConnection(createBus func() (*dbus.Conn, error)) (*dbus.Conn, error) { + conn, err := dbusAuthConnection(createBus) + if err != nil { + return nil, err + } + + if err = conn.Hello(); err != nil { + conn.Close() + return nil, err + } + + return conn, nil +} + +func systemdObject(conn *dbus.Conn) dbus.BusObject { + return conn.Object("org.freedesktop.systemd1", dbus.ObjectPath("/org/freedesktop/systemd1")) +} diff --git a/vendor/github.com/coreos/go-systemd/dbus/methods.go b/vendor/github.com/coreos/go-systemd/dbus/methods.go new file mode 100644 index 000000000..ab17f7cc7 --- /dev/null +++ b/vendor/github.com/coreos/go-systemd/dbus/methods.go @@ -0,0 +1,565 @@ +// Copyright 2015 CoreOS, Inc. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package dbus + +import ( + "errors" + "path" + "strconv" + + "github.com/godbus/dbus" +) + +func (c *Conn) jobComplete(signal *dbus.Signal) { + var id uint32 + var job dbus.ObjectPath + var unit string + var result string + dbus.Store(signal.Body, &id, &job, &unit, &result) + c.jobListener.Lock() + out, ok := c.jobListener.jobs[job] + if ok { + out <- result + delete(c.jobListener.jobs, job) + } + c.jobListener.Unlock() +} + +func (c *Conn) startJob(ch chan<- string, job string, args ...interface{}) (int, error) { + if ch != nil { + c.jobListener.Lock() + defer c.jobListener.Unlock() + } + + var p dbus.ObjectPath + err := c.sysobj.Call(job, 0, args...).Store(&p) + if err != nil { + return 0, err + } + + if ch != nil { + c.jobListener.jobs[p] = ch + } + + // ignore error since 0 is fine if conversion fails + jobID, _ := strconv.Atoi(path.Base(string(p))) + + return jobID, nil +} + +// StartUnit enqueues a start job and depending jobs, if any (unless otherwise +// specified by the mode string). +// +// Takes the unit to activate, plus a mode string. The mode needs to be one of +// replace, fail, isolate, ignore-dependencies, ignore-requirements. If +// "replace" the call will start the unit and its dependencies, possibly +// replacing already queued jobs that conflict with this. If "fail" the call +// will start the unit and its dependencies, but will fail if this would change +// an already queued job. If "isolate" the call will start the unit in question +// and terminate all units that aren't dependencies of it. If +// "ignore-dependencies" it will start a unit but ignore all its dependencies. +// If "ignore-requirements" it will start a unit but only ignore the +// requirement dependencies. It is not recommended to make use of the latter +// two options. +// +// If the provided channel is non-nil, a result string will be sent to it upon +// job completion: one of done, canceled, timeout, failed, dependency, skipped. +// done indicates successful execution of a job. canceled indicates that a job +// has been canceled before it finished execution. timeout indicates that the +// job timeout was reached. failed indicates that the job failed. dependency +// indicates that a job this job has been depending on failed and the job hence +// has been removed too. skipped indicates that a job was skipped because it +// didn't apply to the units current state. +// +// If no error occurs, the ID of the underlying systemd job will be returned. There +// does exist the possibility for no error to be returned, but for the returned job +// ID to be 0. In this case, the actual underlying ID is not 0 and this datapoint +// should not be considered authoritative. +// +// If an error does occur, it will be returned to the user alongside a job ID of 0. +func (c *Conn) StartUnit(name string, mode string, ch chan<- string) (int, error) { + return c.startJob(ch, "org.freedesktop.systemd1.Manager.StartUnit", name, mode) +} + +// StopUnit is similar to StartUnit but stops the specified unit rather +// than starting it. +func (c *Conn) StopUnit(name string, mode string, ch chan<- string) (int, error) { + return c.startJob(ch, "org.freedesktop.systemd1.Manager.StopUnit", name, mode) +} + +// ReloadUnit reloads a unit. Reloading is done only if the unit is already running and fails otherwise. +func (c *Conn) ReloadUnit(name string, mode string, ch chan<- string) (int, error) { + return c.startJob(ch, "org.freedesktop.systemd1.Manager.ReloadUnit", name, mode) +} + +// RestartUnit restarts a service. If a service is restarted that isn't +// running it will be started. +func (c *Conn) RestartUnit(name string, mode string, ch chan<- string) (int, error) { + return c.startJob(ch, "org.freedesktop.systemd1.Manager.RestartUnit", name, mode) +} + +// TryRestartUnit is like RestartUnit, except that a service that isn't running +// is not affected by the restart. +func (c *Conn) TryRestartUnit(name string, mode string, ch chan<- string) (int, error) { + return c.startJob(ch, "org.freedesktop.systemd1.Manager.TryRestartUnit", name, mode) +} + +// ReloadOrRestart attempts a reload if the unit supports it and use a restart +// otherwise. +func (c *Conn) ReloadOrRestartUnit(name string, mode string, ch chan<- string) (int, error) { + return c.startJob(ch, "org.freedesktop.systemd1.Manager.ReloadOrRestartUnit", name, mode) +} + +// ReloadOrTryRestart attempts a reload if the unit supports it and use a "Try" +// flavored restart otherwise. +func (c *Conn) ReloadOrTryRestartUnit(name string, mode string, ch chan<- string) (int, error) { + return c.startJob(ch, "org.freedesktop.systemd1.Manager.ReloadOrTryRestartUnit", name, mode) +} + +// StartTransientUnit() may be used to create and start a transient unit, which +// will be released as soon as it is not running or referenced anymore or the +// system is rebooted. name is the unit name including suffix, and must be +// unique. mode is the same as in StartUnit(), properties contains properties +// of the unit. +func (c *Conn) StartTransientUnit(name string, mode string, properties []Property, ch chan<- string) (int, error) { + return c.startJob(ch, "org.freedesktop.systemd1.Manager.StartTransientUnit", name, mode, properties, make([]PropertyCollection, 0)) +} + +// KillUnit takes the unit name and a UNIX signal number to send. All of the unit's +// processes are killed. +func (c *Conn) KillUnit(name string, signal int32) { + c.sysobj.Call("org.freedesktop.systemd1.Manager.KillUnit", 0, name, "all", signal).Store() +} + +// ResetFailedUnit resets the "failed" state of a specific unit. +func (c *Conn) ResetFailedUnit(name string) error { + return c.sysobj.Call("org.freedesktop.systemd1.Manager.ResetFailedUnit", 0, name).Store() +} + +// getProperties takes the unit name and returns all of its dbus object properties, for the given dbus interface +func (c *Conn) getProperties(unit string, dbusInterface string) (map[string]interface{}, error) { + var err error + var props map[string]dbus.Variant + + path := unitPath(unit) + if !path.IsValid() { + return nil, errors.New("invalid unit name: " + unit) + } + + obj := c.sysconn.Object("org.freedesktop.systemd1", path) + err = obj.Call("org.freedesktop.DBus.Properties.GetAll", 0, dbusInterface).Store(&props) + if err != nil { + return nil, err + } + + out := make(map[string]interface{}, len(props)) + for k, v := range props { + out[k] = v.Value() + } + + return out, nil +} + +// GetUnitProperties takes the unit name and returns all of its dbus object properties. +func (c *Conn) GetUnitProperties(unit string) (map[string]interface{}, error) { + return c.getProperties(unit, "org.freedesktop.systemd1.Unit") +} + +func (c *Conn) getProperty(unit string, dbusInterface string, propertyName string) (*Property, error) { + var err error + var prop dbus.Variant + + path := unitPath(unit) + if !path.IsValid() { + return nil, errors.New("invalid unit name: " + unit) + } + + obj := c.sysconn.Object("org.freedesktop.systemd1", path) + err = obj.Call("org.freedesktop.DBus.Properties.Get", 0, dbusInterface, propertyName).Store(&prop) + if err != nil { + return nil, err + } + + return &Property{Name: propertyName, Value: prop}, nil +} + +func (c *Conn) GetUnitProperty(unit string, propertyName string) (*Property, error) { + return c.getProperty(unit, "org.freedesktop.systemd1.Unit", propertyName) +} + +// GetServiceProperty returns property for given service name and property name +func (c *Conn) GetServiceProperty(service string, propertyName string) (*Property, error) { + return c.getProperty(service, "org.freedesktop.systemd1.Service", propertyName) +} + +// GetUnitTypeProperties returns the extra properties for a unit, specific to the unit type. +// Valid values for unitType: Service, Socket, Target, Device, Mount, Automount, Snapshot, Timer, Swap, Path, Slice, Scope +// return "dbus.Error: Unknown interface" if the unitType is not the correct type of the unit +func (c *Conn) GetUnitTypeProperties(unit string, unitType string) (map[string]interface{}, error) { + return c.getProperties(unit, "org.freedesktop.systemd1."+unitType) +} + +// SetUnitProperties() may be used to modify certain unit properties at runtime. +// Not all properties may be changed at runtime, but many resource management +// settings (primarily those in systemd.cgroup(5)) may. The changes are applied +// instantly, and stored on disk for future boots, unless runtime is true, in which +// case the settings only apply until the next reboot. name is the name of the unit +// to modify. properties are the settings to set, encoded as an array of property +// name and value pairs. +func (c *Conn) SetUnitProperties(name string, runtime bool, properties ...Property) error { + return c.sysobj.Call("org.freedesktop.systemd1.Manager.SetUnitProperties", 0, name, runtime, properties).Store() +} + +func (c *Conn) GetUnitTypeProperty(unit string, unitType string, propertyName string) (*Property, error) { + return c.getProperty(unit, "org.freedesktop.systemd1."+unitType, propertyName) +} + +type UnitStatus struct { + Name string // The primary unit name as string + Description string // The human readable description string + LoadState string // The load state (i.e. whether the unit file has been loaded successfully) + ActiveState string // The active state (i.e. whether the unit is currently started or not) + SubState string // The sub state (a more fine-grained version of the active state that is specific to the unit type, which the active state is not) + Followed string // A unit that is being followed in its state by this unit, if there is any, otherwise the empty string. + Path dbus.ObjectPath // The unit object path + JobId uint32 // If there is a job queued for the job unit the numeric job id, 0 otherwise + JobType string // The job type as string + JobPath dbus.ObjectPath // The job object path +} + +type storeFunc func(retvalues ...interface{}) error + +func (c *Conn) listUnitsInternal(f storeFunc) ([]UnitStatus, error) { + result := make([][]interface{}, 0) + err := f(&result) + if err != nil { + return nil, err + } + + resultInterface := make([]interface{}, len(result)) + for i := range result { + resultInterface[i] = result[i] + } + + status := make([]UnitStatus, len(result)) + statusInterface := make([]interface{}, len(status)) + for i := range status { + statusInterface[i] = &status[i] + } + + err = dbus.Store(resultInterface, statusInterface...) + if err != nil { + return nil, err + } + + return status, nil +} + +// ListUnits returns an array with all currently loaded units. Note that +// units may be known by multiple names at the same time, and hence there might +// be more unit names loaded than actual units behind them. +func (c *Conn) ListUnits() ([]UnitStatus, error) { + return c.listUnitsInternal(c.sysobj.Call("org.freedesktop.systemd1.Manager.ListUnits", 0).Store) +} + +// ListUnitsFiltered returns an array with units filtered by state. +// It takes a list of units' statuses to filter. +func (c *Conn) ListUnitsFiltered(states []string) ([]UnitStatus, error) { + return c.listUnitsInternal(c.sysobj.Call("org.freedesktop.systemd1.Manager.ListUnitsFiltered", 0, states).Store) +} + +// ListUnitsByPatterns returns an array with units. +// It takes a list of units' statuses and names to filter. +// Note that units may be known by multiple names at the same time, +// and hence there might be more unit names loaded than actual units behind them. +func (c *Conn) ListUnitsByPatterns(states []string, patterns []string) ([]UnitStatus, error) { + return c.listUnitsInternal(c.sysobj.Call("org.freedesktop.systemd1.Manager.ListUnitsByPatterns", 0, states, patterns).Store) +} + +// ListUnitsByNames returns an array with units. It takes a list of units' +// names and returns an UnitStatus array. Comparing to ListUnitsByPatterns +// method, this method returns statuses even for inactive or non-existing +// units. Input array should contain exact unit names, but not patterns. +func (c *Conn) ListUnitsByNames(units []string) ([]UnitStatus, error) { + return c.listUnitsInternal(c.sysobj.Call("org.freedesktop.systemd1.Manager.ListUnitsByNames", 0, units).Store) +} + +type UnitFile struct { + Path string + Type string +} + +func (c *Conn) listUnitFilesInternal(f storeFunc) ([]UnitFile, error) { + result := make([][]interface{}, 0) + err := f(&result) + if err != nil { + return nil, err + } + + resultInterface := make([]interface{}, len(result)) + for i := range result { + resultInterface[i] = result[i] + } + + files := make([]UnitFile, len(result)) + fileInterface := make([]interface{}, len(files)) + for i := range files { + fileInterface[i] = &files[i] + } + + err = dbus.Store(resultInterface, fileInterface...) + if err != nil { + return nil, err + } + + return files, nil +} + +// ListUnitFiles returns an array of all available units on disk. +func (c *Conn) ListUnitFiles() ([]UnitFile, error) { + return c.listUnitFilesInternal(c.sysobj.Call("org.freedesktop.systemd1.Manager.ListUnitFiles", 0).Store) +} + +// ListUnitFilesByPatterns returns an array of all available units on disk matched the patterns. +func (c *Conn) ListUnitFilesByPatterns(states []string, patterns []string) ([]UnitFile, error) { + return c.listUnitFilesInternal(c.sysobj.Call("org.freedesktop.systemd1.Manager.ListUnitFilesByPatterns", 0, states, patterns).Store) +} + +type LinkUnitFileChange EnableUnitFileChange + +// LinkUnitFiles() links unit files (that are located outside of the +// usual unit search paths) into the unit search path. +// +// It takes a list of absolute paths to unit files to link and two +// booleans. The first boolean controls whether the unit shall be +// enabled for runtime only (true, /run), or persistently (false, +// /etc). +// The second controls whether symlinks pointing to other units shall +// be replaced if necessary. +// +// This call returns a list of the changes made. The list consists of +// structures with three strings: the type of the change (one of symlink +// or unlink), the file name of the symlink and the destination of the +// symlink. +func (c *Conn) LinkUnitFiles(files []string, runtime bool, force bool) ([]LinkUnitFileChange, error) { + result := make([][]interface{}, 0) + err := c.sysobj.Call("org.freedesktop.systemd1.Manager.LinkUnitFiles", 0, files, runtime, force).Store(&result) + if err != nil { + return nil, err + } + + resultInterface := make([]interface{}, len(result)) + for i := range result { + resultInterface[i] = result[i] + } + + changes := make([]LinkUnitFileChange, len(result)) + changesInterface := make([]interface{}, len(changes)) + for i := range changes { + changesInterface[i] = &changes[i] + } + + err = dbus.Store(resultInterface, changesInterface...) + if err != nil { + return nil, err + } + + return changes, nil +} + +// EnableUnitFiles() may be used to enable one or more units in the system (by +// creating symlinks to them in /etc or /run). +// +// It takes a list of unit files to enable (either just file names or full +// absolute paths if the unit files are residing outside the usual unit +// search paths), and two booleans: the first controls whether the unit shall +// be enabled for runtime only (true, /run), or persistently (false, /etc). +// The second one controls whether symlinks pointing to other units shall +// be replaced if necessary. +// +// This call returns one boolean and an array with the changes made. The +// boolean signals whether the unit files contained any enablement +// information (i.e. an [Install]) section. The changes list consists of +// structures with three strings: the type of the change (one of symlink +// or unlink), the file name of the symlink and the destination of the +// symlink. +func (c *Conn) EnableUnitFiles(files []string, runtime bool, force bool) (bool, []EnableUnitFileChange, error) { + var carries_install_info bool + + result := make([][]interface{}, 0) + err := c.sysobj.Call("org.freedesktop.systemd1.Manager.EnableUnitFiles", 0, files, runtime, force).Store(&carries_install_info, &result) + if err != nil { + return false, nil, err + } + + resultInterface := make([]interface{}, len(result)) + for i := range result { + resultInterface[i] = result[i] + } + + changes := make([]EnableUnitFileChange, len(result)) + changesInterface := make([]interface{}, len(changes)) + for i := range changes { + changesInterface[i] = &changes[i] + } + + err = dbus.Store(resultInterface, changesInterface...) + if err != nil { + return false, nil, err + } + + return carries_install_info, changes, nil +} + +type EnableUnitFileChange struct { + Type string // Type of the change (one of symlink or unlink) + Filename string // File name of the symlink + Destination string // Destination of the symlink +} + +// DisableUnitFiles() may be used to disable one or more units in the system (by +// removing symlinks to them from /etc or /run). +// +// It takes a list of unit files to disable (either just file names or full +// absolute paths if the unit files are residing outside the usual unit +// search paths), and one boolean: whether the unit was enabled for runtime +// only (true, /run), or persistently (false, /etc). +// +// This call returns an array with the changes made. The changes list +// consists of structures with three strings: the type of the change (one of +// symlink or unlink), the file name of the symlink and the destination of the +// symlink. +func (c *Conn) DisableUnitFiles(files []string, runtime bool) ([]DisableUnitFileChange, error) { + result := make([][]interface{}, 0) + err := c.sysobj.Call("org.freedesktop.systemd1.Manager.DisableUnitFiles", 0, files, runtime).Store(&result) + if err != nil { + return nil, err + } + + resultInterface := make([]interface{}, len(result)) + for i := range result { + resultInterface[i] = result[i] + } + + changes := make([]DisableUnitFileChange, len(result)) + changesInterface := make([]interface{}, len(changes)) + for i := range changes { + changesInterface[i] = &changes[i] + } + + err = dbus.Store(resultInterface, changesInterface...) + if err != nil { + return nil, err + } + + return changes, nil +} + +type DisableUnitFileChange struct { + Type string // Type of the change (one of symlink or unlink) + Filename string // File name of the symlink + Destination string // Destination of the symlink +} + +// MaskUnitFiles masks one or more units in the system +// +// It takes three arguments: +// * list of units to mask (either just file names or full +// absolute paths if the unit files are residing outside +// the usual unit search paths) +// * runtime to specify whether the unit was enabled for runtime +// only (true, /run/systemd/..), or persistently (false, /etc/systemd/..) +// * force flag +func (c *Conn) MaskUnitFiles(files []string, runtime bool, force bool) ([]MaskUnitFileChange, error) { + result := make([][]interface{}, 0) + err := c.sysobj.Call("org.freedesktop.systemd1.Manager.MaskUnitFiles", 0, files, runtime, force).Store(&result) + if err != nil { + return nil, err + } + + resultInterface := make([]interface{}, len(result)) + for i := range result { + resultInterface[i] = result[i] + } + + changes := make([]MaskUnitFileChange, len(result)) + changesInterface := make([]interface{}, len(changes)) + for i := range changes { + changesInterface[i] = &changes[i] + } + + err = dbus.Store(resultInterface, changesInterface...) + if err != nil { + return nil, err + } + + return changes, nil +} + +type MaskUnitFileChange struct { + Type string // Type of the change (one of symlink or unlink) + Filename string // File name of the symlink + Destination string // Destination of the symlink +} + +// UnmaskUnitFiles unmasks one or more units in the system +// +// It takes two arguments: +// * list of unit files to mask (either just file names or full +// absolute paths if the unit files are residing outside +// the usual unit search paths) +// * runtime to specify whether the unit was enabled for runtime +// only (true, /run/systemd/..), or persistently (false, /etc/systemd/..) +func (c *Conn) UnmaskUnitFiles(files []string, runtime bool) ([]UnmaskUnitFileChange, error) { + result := make([][]interface{}, 0) + err := c.sysobj.Call("org.freedesktop.systemd1.Manager.UnmaskUnitFiles", 0, files, runtime).Store(&result) + if err != nil { + return nil, err + } + + resultInterface := make([]interface{}, len(result)) + for i := range result { + resultInterface[i] = result[i] + } + + changes := make([]UnmaskUnitFileChange, len(result)) + changesInterface := make([]interface{}, len(changes)) + for i := range changes { + changesInterface[i] = &changes[i] + } + + err = dbus.Store(resultInterface, changesInterface...) + if err != nil { + return nil, err + } + + return changes, nil +} + +type UnmaskUnitFileChange struct { + Type string // Type of the change (one of symlink or unlink) + Filename string // File name of the symlink + Destination string // Destination of the symlink +} + +// Reload instructs systemd to scan for and reload unit files. This is +// equivalent to a 'systemctl daemon-reload'. +func (c *Conn) Reload() error { + return c.sysobj.Call("org.freedesktop.systemd1.Manager.Reload", 0).Store() +} + +func unitPath(name string) dbus.ObjectPath { + return dbus.ObjectPath("/org/freedesktop/systemd1/unit/" + PathBusEscape(name)) +} diff --git a/vendor/github.com/coreos/go-systemd/dbus/properties.go b/vendor/github.com/coreos/go-systemd/dbus/properties.go new file mode 100644 index 000000000..6c8189587 --- /dev/null +++ b/vendor/github.com/coreos/go-systemd/dbus/properties.go @@ -0,0 +1,237 @@ +// Copyright 2015 CoreOS, Inc. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package dbus + +import ( + "github.com/godbus/dbus" +) + +// From the systemd docs: +// +// The properties array of StartTransientUnit() may take many of the settings +// that may also be configured in unit files. Not all parameters are currently +// accepted though, but we plan to cover more properties with future release. +// Currently you may set the Description, Slice and all dependency types of +// units, as well as RemainAfterExit, ExecStart for service units, +// TimeoutStopUSec and PIDs for scope units, and CPUAccounting, CPUShares, +// BlockIOAccounting, BlockIOWeight, BlockIOReadBandwidth, +// BlockIOWriteBandwidth, BlockIODeviceWeight, MemoryAccounting, MemoryLimit, +// DevicePolicy, DeviceAllow for services/scopes/slices. These fields map +// directly to their counterparts in unit files and as normal D-Bus object +// properties. The exception here is the PIDs field of scope units which is +// used for construction of the scope only and specifies the initial PIDs to +// add to the scope object. + +type Property struct { + Name string + Value dbus.Variant +} + +type PropertyCollection struct { + Name string + Properties []Property +} + +type execStart struct { + Path string // the binary path to execute + Args []string // an array with all arguments to pass to the executed command, starting with argument 0 + UncleanIsFailure bool // a boolean whether it should be considered a failure if the process exits uncleanly +} + +// PropExecStart sets the ExecStart service property. The first argument is a +// slice with the binary path to execute followed by the arguments to pass to +// the executed command. See +// http://www.freedesktop.org/software/systemd/man/systemd.service.html#ExecStart= +func PropExecStart(command []string, uncleanIsFailure bool) Property { + execStarts := []execStart{ + execStart{ + Path: command[0], + Args: command, + UncleanIsFailure: uncleanIsFailure, + }, + } + + return Property{ + Name: "ExecStart", + Value: dbus.MakeVariant(execStarts), + } +} + +// PropRemainAfterExit sets the RemainAfterExit service property. See +// http://www.freedesktop.org/software/systemd/man/systemd.service.html#RemainAfterExit= +func PropRemainAfterExit(b bool) Property { + return Property{ + Name: "RemainAfterExit", + Value: dbus.MakeVariant(b), + } +} + +// PropType sets the Type service property. See +// http://www.freedesktop.org/software/systemd/man/systemd.service.html#Type= +func PropType(t string) Property { + return Property{ + Name: "Type", + Value: dbus.MakeVariant(t), + } +} + +// PropDescription sets the Description unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit#Description= +func PropDescription(desc string) Property { + return Property{ + Name: "Description", + Value: dbus.MakeVariant(desc), + } +} + +func propDependency(name string, units []string) Property { + return Property{ + Name: name, + Value: dbus.MakeVariant(units), + } +} + +// PropRequires sets the Requires unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#Requires= +func PropRequires(units ...string) Property { + return propDependency("Requires", units) +} + +// PropRequiresOverridable sets the RequiresOverridable unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#RequiresOverridable= +func PropRequiresOverridable(units ...string) Property { + return propDependency("RequiresOverridable", units) +} + +// PropRequisite sets the Requisite unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#Requisite= +func PropRequisite(units ...string) Property { + return propDependency("Requisite", units) +} + +// PropRequisiteOverridable sets the RequisiteOverridable unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#RequisiteOverridable= +func PropRequisiteOverridable(units ...string) Property { + return propDependency("RequisiteOverridable", units) +} + +// PropWants sets the Wants unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#Wants= +func PropWants(units ...string) Property { + return propDependency("Wants", units) +} + +// PropBindsTo sets the BindsTo unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#BindsTo= +func PropBindsTo(units ...string) Property { + return propDependency("BindsTo", units) +} + +// PropRequiredBy sets the RequiredBy unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#RequiredBy= +func PropRequiredBy(units ...string) Property { + return propDependency("RequiredBy", units) +} + +// PropRequiredByOverridable sets the RequiredByOverridable unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#RequiredByOverridable= +func PropRequiredByOverridable(units ...string) Property { + return propDependency("RequiredByOverridable", units) +} + +// PropWantedBy sets the WantedBy unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#WantedBy= +func PropWantedBy(units ...string) Property { + return propDependency("WantedBy", units) +} + +// PropBoundBy sets the BoundBy unit property. See +// http://www.freedesktop.org/software/systemd/main/systemd.unit.html#BoundBy= +func PropBoundBy(units ...string) Property { + return propDependency("BoundBy", units) +} + +// PropConflicts sets the Conflicts unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#Conflicts= +func PropConflicts(units ...string) Property { + return propDependency("Conflicts", units) +} + +// PropConflictedBy sets the ConflictedBy unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#ConflictedBy= +func PropConflictedBy(units ...string) Property { + return propDependency("ConflictedBy", units) +} + +// PropBefore sets the Before unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#Before= +func PropBefore(units ...string) Property { + return propDependency("Before", units) +} + +// PropAfter sets the After unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#After= +func PropAfter(units ...string) Property { + return propDependency("After", units) +} + +// PropOnFailure sets the OnFailure unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#OnFailure= +func PropOnFailure(units ...string) Property { + return propDependency("OnFailure", units) +} + +// PropTriggers sets the Triggers unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#Triggers= +func PropTriggers(units ...string) Property { + return propDependency("Triggers", units) +} + +// PropTriggeredBy sets the TriggeredBy unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#TriggeredBy= +func PropTriggeredBy(units ...string) Property { + return propDependency("TriggeredBy", units) +} + +// PropPropagatesReloadTo sets the PropagatesReloadTo unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#PropagatesReloadTo= +func PropPropagatesReloadTo(units ...string) Property { + return propDependency("PropagatesReloadTo", units) +} + +// PropRequiresMountsFor sets the RequiresMountsFor unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.unit.html#RequiresMountsFor= +func PropRequiresMountsFor(units ...string) Property { + return propDependency("RequiresMountsFor", units) +} + +// PropSlice sets the Slice unit property. See +// http://www.freedesktop.org/software/systemd/man/systemd.resource-control.html#Slice= +func PropSlice(slice string) Property { + return Property{ + Name: "Slice", + Value: dbus.MakeVariant(slice), + } +} + +// PropPids sets the PIDs field of scope units used in the initial construction +// of the scope only and specifies the initial PIDs to add to the scope object. +// See https://www.freedesktop.org/wiki/Software/systemd/ControlGroupInterface/#properties +func PropPids(pids ...uint32) Property { + return Property{ + Name: "PIDs", + Value: dbus.MakeVariant(pids), + } +} diff --git a/vendor/github.com/coreos/go-systemd/dbus/set.go b/vendor/github.com/coreos/go-systemd/dbus/set.go new file mode 100644 index 000000000..f92e6fbed --- /dev/null +++ b/vendor/github.com/coreos/go-systemd/dbus/set.go @@ -0,0 +1,47 @@ +// Copyright 2015 CoreOS, Inc. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package dbus + +type set struct { + data map[string]bool +} + +func (s *set) Add(value string) { + s.data[value] = true +} + +func (s *set) Remove(value string) { + delete(s.data, value) +} + +func (s *set) Contains(value string) (exists bool) { + _, exists = s.data[value] + return +} + +func (s *set) Length() int { + return len(s.data) +} + +func (s *set) Values() (values []string) { + for val, _ := range s.data { + values = append(values, val) + } + return +} + +func newSet() *set { + return &set{make(map[string]bool)} +} diff --git a/vendor/github.com/coreos/go-systemd/dbus/subscription.go b/vendor/github.com/coreos/go-systemd/dbus/subscription.go new file mode 100644 index 000000000..996451445 --- /dev/null +++ b/vendor/github.com/coreos/go-systemd/dbus/subscription.go @@ -0,0 +1,250 @@ +// Copyright 2015 CoreOS, Inc. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package dbus + +import ( + "errors" + "time" + + "github.com/godbus/dbus" +) + +const ( + cleanIgnoreInterval = int64(10 * time.Second) + ignoreInterval = int64(30 * time.Millisecond) +) + +// Subscribe sets up this connection to subscribe to all systemd dbus events. +// This is required before calling SubscribeUnits. When the connection closes +// systemd will automatically stop sending signals so there is no need to +// explicitly call Unsubscribe(). +func (c *Conn) Subscribe() error { + c.sigconn.BusObject().Call("org.freedesktop.DBus.AddMatch", 0, + "type='signal',interface='org.freedesktop.systemd1.Manager',member='UnitNew'") + c.sigconn.BusObject().Call("org.freedesktop.DBus.AddMatch", 0, + "type='signal',interface='org.freedesktop.DBus.Properties',member='PropertiesChanged'") + + err := c.sigobj.Call("org.freedesktop.systemd1.Manager.Subscribe", 0).Store() + if err != nil { + return err + } + + return nil +} + +// Unsubscribe this connection from systemd dbus events. +func (c *Conn) Unsubscribe() error { + err := c.sigobj.Call("org.freedesktop.systemd1.Manager.Unsubscribe", 0).Store() + if err != nil { + return err + } + + return nil +} + +func (c *Conn) dispatch() { + ch := make(chan *dbus.Signal, signalBuffer) + + c.sigconn.Signal(ch) + + go func() { + for { + signal, ok := <-ch + if !ok { + return + } + + if signal.Name == "org.freedesktop.systemd1.Manager.JobRemoved" { + c.jobComplete(signal) + } + + if c.subscriber.updateCh == nil { + continue + } + + var unitPath dbus.ObjectPath + switch signal.Name { + case "org.freedesktop.systemd1.Manager.JobRemoved": + unitName := signal.Body[2].(string) + c.sysobj.Call("org.freedesktop.systemd1.Manager.GetUnit", 0, unitName).Store(&unitPath) + case "org.freedesktop.systemd1.Manager.UnitNew": + unitPath = signal.Body[1].(dbus.ObjectPath) + case "org.freedesktop.DBus.Properties.PropertiesChanged": + if signal.Body[0].(string) == "org.freedesktop.systemd1.Unit" { + unitPath = signal.Path + } + } + + if unitPath == dbus.ObjectPath("") { + continue + } + + c.sendSubStateUpdate(unitPath) + } + }() +} + +// Returns two unbuffered channels which will receive all changed units every +// interval. Deleted units are sent as nil. +func (c *Conn) SubscribeUnits(interval time.Duration) (<-chan map[string]*UnitStatus, <-chan error) { + return c.SubscribeUnitsCustom(interval, 0, func(u1, u2 *UnitStatus) bool { return *u1 != *u2 }, nil) +} + +// SubscribeUnitsCustom is like SubscribeUnits but lets you specify the buffer +// size of the channels, the comparison function for detecting changes and a filter +// function for cutting down on the noise that your channel receives. +func (c *Conn) SubscribeUnitsCustom(interval time.Duration, buffer int, isChanged func(*UnitStatus, *UnitStatus) bool, filterUnit func(string) bool) (<-chan map[string]*UnitStatus, <-chan error) { + old := make(map[string]*UnitStatus) + statusChan := make(chan map[string]*UnitStatus, buffer) + errChan := make(chan error, buffer) + + go func() { + for { + timerChan := time.After(interval) + + units, err := c.ListUnits() + if err == nil { + cur := make(map[string]*UnitStatus) + for i := range units { + if filterUnit != nil && filterUnit(units[i].Name) { + continue + } + cur[units[i].Name] = &units[i] + } + + // add all new or changed units + changed := make(map[string]*UnitStatus) + for n, u := range cur { + if oldU, ok := old[n]; !ok || isChanged(oldU, u) { + changed[n] = u + } + delete(old, n) + } + + // add all deleted units + for oldN := range old { + changed[oldN] = nil + } + + old = cur + + if len(changed) != 0 { + statusChan <- changed + } + } else { + errChan <- err + } + + <-timerChan + } + }() + + return statusChan, errChan +} + +type SubStateUpdate struct { + UnitName string + SubState string +} + +// SetSubStateSubscriber writes to updateCh when any unit's substate changes. +// Although this writes to updateCh on every state change, the reported state +// may be more recent than the change that generated it (due to an unavoidable +// race in the systemd dbus interface). That is, this method provides a good +// way to keep a current view of all units' states, but is not guaranteed to +// show every state transition they go through. Furthermore, state changes +// will only be written to the channel with non-blocking writes. If updateCh +// is full, it attempts to write an error to errCh; if errCh is full, the error +// passes silently. +func (c *Conn) SetSubStateSubscriber(updateCh chan<- *SubStateUpdate, errCh chan<- error) { + c.subscriber.Lock() + defer c.subscriber.Unlock() + c.subscriber.updateCh = updateCh + c.subscriber.errCh = errCh +} + +func (c *Conn) sendSubStateUpdate(path dbus.ObjectPath) { + c.subscriber.Lock() + defer c.subscriber.Unlock() + + if c.shouldIgnore(path) { + return + } + + info, err := c.GetUnitProperties(string(path)) + if err != nil { + select { + case c.subscriber.errCh <- err: + default: + } + } + + name := info["Id"].(string) + substate := info["SubState"].(string) + + update := &SubStateUpdate{name, substate} + select { + case c.subscriber.updateCh <- update: + default: + select { + case c.subscriber.errCh <- errors.New("update channel full!"): + default: + } + } + + c.updateIgnore(path, info) +} + +// The ignore functions work around a wart in the systemd dbus interface. +// Requesting the properties of an unloaded unit will cause systemd to send a +// pair of UnitNew/UnitRemoved signals. Because we need to get a unit's +// properties on UnitNew (as that's the only indication of a new unit coming up +// for the first time), we would enter an infinite loop if we did not attempt +// to detect and ignore these spurious signals. The signal themselves are +// indistinguishable from relevant ones, so we (somewhat hackishly) ignore an +// unloaded unit's signals for a short time after requesting its properties. +// This means that we will miss e.g. a transient unit being restarted +// *immediately* upon failure and also a transient unit being started +// immediately after requesting its status (with systemctl status, for example, +// because this causes a UnitNew signal to be sent which then causes us to fetch +// the properties). + +func (c *Conn) shouldIgnore(path dbus.ObjectPath) bool { + t, ok := c.subscriber.ignore[path] + return ok && t >= time.Now().UnixNano() +} + +func (c *Conn) updateIgnore(path dbus.ObjectPath, info map[string]interface{}) { + c.cleanIgnore() + + // unit is unloaded - it will trigger bad systemd dbus behavior + if info["LoadState"].(string) == "not-found" { + c.subscriber.ignore[path] = time.Now().UnixNano() + ignoreInterval + } +} + +// without this, ignore would grow unboundedly over time +func (c *Conn) cleanIgnore() { + now := time.Now().UnixNano() + if c.subscriber.cleanIgnore < now { + c.subscriber.cleanIgnore = now + cleanIgnoreInterval + + for p, t := range c.subscriber.ignore { + if t < now { + delete(c.subscriber.ignore, p) + } + } + } +} diff --git a/vendor/github.com/coreos/go-systemd/dbus/subscription_set.go b/vendor/github.com/coreos/go-systemd/dbus/subscription_set.go new file mode 100644 index 000000000..5b408d584 --- /dev/null +++ b/vendor/github.com/coreos/go-systemd/dbus/subscription_set.go @@ -0,0 +1,57 @@ +// Copyright 2015 CoreOS, Inc. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package dbus + +import ( + "time" +) + +// SubscriptionSet returns a subscription set which is like conn.Subscribe but +// can filter to only return events for a set of units. +type SubscriptionSet struct { + *set + conn *Conn +} + +func (s *SubscriptionSet) filter(unit string) bool { + return !s.Contains(unit) +} + +// Subscribe starts listening for dbus events for all of the units in the set. +// Returns channels identical to conn.SubscribeUnits. +func (s *SubscriptionSet) Subscribe() (<-chan map[string]*UnitStatus, <-chan error) { + // TODO: Make fully evented by using systemd 209 with properties changed values + return s.conn.SubscribeUnitsCustom(time.Second, 0, + mismatchUnitStatus, + func(unit string) bool { return s.filter(unit) }, + ) +} + +// NewSubscriptionSet returns a new subscription set. +func (conn *Conn) NewSubscriptionSet() *SubscriptionSet { + return &SubscriptionSet{newSet(), conn} +} + +// mismatchUnitStatus returns true if the provided UnitStatus objects +// are not equivalent. false is returned if the objects are equivalent. +// Only the Name, Description and state-related fields are used in +// the comparison. +func mismatchUnitStatus(u1, u2 *UnitStatus) bool { + return u1.Name != u2.Name || + u1.Description != u2.Description || + u1.LoadState != u2.LoadState || + u1.ActiveState != u2.ActiveState || + u1.SubState != u2.SubState +} diff --git a/vendor/github.com/coreos/go-systemd/util/util.go b/vendor/github.com/coreos/go-systemd/util/util.go new file mode 100644 index 000000000..7828ce6f0 --- /dev/null +++ b/vendor/github.com/coreos/go-systemd/util/util.go @@ -0,0 +1,90 @@ +// Copyright 2015 CoreOS, Inc. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// Package util contains utility functions related to systemd that applications +// can use to check things like whether systemd is running. Note that some of +// these functions attempt to manually load systemd libraries at runtime rather +// than linking against them. +package util + +import ( + "fmt" + "io/ioutil" + "os" + "strings" +) + +var ( + ErrNoCGO = fmt.Errorf("go-systemd built with CGO disabled") +) + +// GetRunningSlice attempts to retrieve the name of the systemd slice in which +// the current process is running. +// This function is a wrapper around the libsystemd C library; if it cannot be +// opened, an error is returned. +func GetRunningSlice() (string, error) { + return getRunningSlice() +} + +// RunningFromSystemService tries to detect whether the current process has +// been invoked from a system service. The condition for this is whether the +// process is _not_ a user process. User processes are those running in session +// scopes or under per-user `systemd --user` instances. +// +// To avoid false positives on systems without `pam_systemd` (which is +// responsible for creating user sessions), this function also uses a heuristic +// to detect whether it's being invoked from a session leader process. This is +// the case if the current process is executed directly from a service file +// (e.g. with `ExecStart=/this/cmd`). Note that this heuristic will fail if the +// command is instead launched in a subshell or similar so that it is not +// session leader (e.g. `ExecStart=/bin/bash -c "/this/cmd"`) +// +// This function is a wrapper around the libsystemd C library; if this is +// unable to successfully open a handle to the library for any reason (e.g. it +// cannot be found), an error will be returned. +func RunningFromSystemService() (bool, error) { + return runningFromSystemService() +} + +// CurrentUnitName attempts to retrieve the name of the systemd system unit +// from which the calling process has been invoked. It wraps the systemd +// `sd_pid_get_unit` call, with the same caveat: for processes not part of a +// systemd system unit, this function will return an error. +func CurrentUnitName() (string, error) { + return currentUnitName() +} + +// IsRunningSystemd checks whether the host was booted with systemd as its init +// system. This functions similarly to systemd's `sd_booted(3)`: internally, it +// checks whether /run/systemd/system/ exists and is a directory. +// http://www.freedesktop.org/software/systemd/man/sd_booted.html +func IsRunningSystemd() bool { + fi, err := os.Lstat("/run/systemd/system") + if err != nil { + return false + } + return fi.IsDir() +} + +// GetMachineID returns a host's 128-bit machine ID as a string. This functions +// similarly to systemd's `sd_id128_get_machine`: internally, it simply reads +// the contents of /etc/machine-id +// http://www.freedesktop.org/software/systemd/man/sd_id128_get_machine.html +func GetMachineID() (string, error) { + machineID, err := ioutil.ReadFile("/etc/machine-id") + if err != nil { + return "", fmt.Errorf("failed to read /etc/machine-id: %v", err) + } + return strings.TrimSpace(string(machineID)), nil +} diff --git a/vendor/github.com/coreos/go-systemd/util/util_cgo.go b/vendor/github.com/coreos/go-systemd/util/util_cgo.go new file mode 100644 index 000000000..22c0d6099 --- /dev/null +++ b/vendor/github.com/coreos/go-systemd/util/util_cgo.go @@ -0,0 +1,174 @@ +// Copyright 2016 CoreOS, Inc. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// +build cgo + +package util + +// #include +// #include +// #include +// +// int +// my_sd_pid_get_owner_uid(void *f, pid_t pid, uid_t *uid) +// { +// int (*sd_pid_get_owner_uid)(pid_t, uid_t *); +// +// sd_pid_get_owner_uid = (int (*)(pid_t, uid_t *))f; +// return sd_pid_get_owner_uid(pid, uid); +// } +// +// int +// my_sd_pid_get_unit(void *f, pid_t pid, char **unit) +// { +// int (*sd_pid_get_unit)(pid_t, char **); +// +// sd_pid_get_unit = (int (*)(pid_t, char **))f; +// return sd_pid_get_unit(pid, unit); +// } +// +// int +// my_sd_pid_get_slice(void *f, pid_t pid, char **slice) +// { +// int (*sd_pid_get_slice)(pid_t, char **); +// +// sd_pid_get_slice = (int (*)(pid_t, char **))f; +// return sd_pid_get_slice(pid, slice); +// } +// +// int +// am_session_leader() +// { +// return (getsid(0) == getpid()); +// } +import "C" +import ( + "fmt" + "syscall" + "unsafe" + + "github.com/coreos/pkg/dlopen" +) + +var libsystemdNames = []string{ + // systemd < 209 + "libsystemd-login.so.0", + "libsystemd-login.so", + + // systemd >= 209 merged libsystemd-login into libsystemd proper + "libsystemd.so.0", + "libsystemd.so", +} + +func getRunningSlice() (slice string, err error) { + var h *dlopen.LibHandle + h, err = dlopen.GetHandle(libsystemdNames) + if err != nil { + return + } + defer func() { + if err1 := h.Close(); err1 != nil { + err = err1 + } + }() + + sd_pid_get_slice, err := h.GetSymbolPointer("sd_pid_get_slice") + if err != nil { + return + } + + var s string + sl := C.CString(s) + defer C.free(unsafe.Pointer(sl)) + + ret := C.my_sd_pid_get_slice(sd_pid_get_slice, 0, &sl) + if ret < 0 { + err = fmt.Errorf("error calling sd_pid_get_slice: %v", syscall.Errno(-ret)) + return + } + + return C.GoString(sl), nil +} + +func runningFromSystemService() (ret bool, err error) { + var h *dlopen.LibHandle + h, err = dlopen.GetHandle(libsystemdNames) + if err != nil { + return + } + defer func() { + if err1 := h.Close(); err1 != nil { + err = err1 + } + }() + + sd_pid_get_owner_uid, err := h.GetSymbolPointer("sd_pid_get_owner_uid") + if err != nil { + return + } + + var uid C.uid_t + errno := C.my_sd_pid_get_owner_uid(sd_pid_get_owner_uid, 0, &uid) + serrno := syscall.Errno(-errno) + // when we're running from a unit file, sd_pid_get_owner_uid returns + // ENOENT (systemd <220) or ENXIO (systemd >=220) + switch { + case errno >= 0: + ret = false + case serrno == syscall.ENOENT, serrno == syscall.ENXIO: + // Since the implementation of sessions in systemd relies on + // the `pam_systemd` module, using the sd_pid_get_owner_uid + // heuristic alone can result in false positives if that module + // (or PAM itself) is not present or properly configured on the + // system. As such, we also check if we're the session leader, + // which should be the case if we're invoked from a unit file, + // but not if e.g. we're invoked from the command line from a + // user's login session + ret = C.am_session_leader() == 1 + default: + err = fmt.Errorf("error calling sd_pid_get_owner_uid: %v", syscall.Errno(-errno)) + } + return +} + +func currentUnitName() (unit string, err error) { + var h *dlopen.LibHandle + h, err = dlopen.GetHandle(libsystemdNames) + if err != nil { + return + } + defer func() { + if err1 := h.Close(); err1 != nil { + err = err1 + } + }() + + sd_pid_get_unit, err := h.GetSymbolPointer("sd_pid_get_unit") + if err != nil { + return + } + + var s string + u := C.CString(s) + defer C.free(unsafe.Pointer(u)) + + ret := C.my_sd_pid_get_unit(sd_pid_get_unit, 0, &u) + if ret < 0 { + err = fmt.Errorf("error calling sd_pid_get_unit: %v", syscall.Errno(-ret)) + return + } + + unit = C.GoString(u) + return +} diff --git a/vendor/github.com/coreos/go-systemd/util/util_stub.go b/vendor/github.com/coreos/go-systemd/util/util_stub.go new file mode 100644 index 000000000..477589e12 --- /dev/null +++ b/vendor/github.com/coreos/go-systemd/util/util_stub.go @@ -0,0 +1,23 @@ +// Copyright 2016 CoreOS, Inc. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// +build !cgo + +package util + +func getRunningSlice() (string, error) { return "", ErrNoCGO } + +func runningFromSystemService() (bool, error) { return false, ErrNoCGO } + +func currentUnitName() (string, error) { return "", ErrNoCGO } diff --git a/vendor/github.com/coreos/pkg/LICENSE b/vendor/github.com/coreos/pkg/LICENSE new file mode 100644 index 000000000..5c304d1a4 --- /dev/null +++ b/vendor/github.com/coreos/pkg/LICENSE @@ -0,0 +1,201 @@ +Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "{}" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright {yyyy} {name of copyright owner} + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/vendor/github.com/coreos/pkg/NOTICE b/vendor/github.com/coreos/pkg/NOTICE new file mode 100644 index 000000000..b39ddfa5c --- /dev/null +++ b/vendor/github.com/coreos/pkg/NOTICE @@ -0,0 +1,5 @@ +CoreOS Project +Copyright 2014 CoreOS, Inc + +This product includes software developed at CoreOS, Inc. +(http://www.coreos.com/). diff --git a/vendor/github.com/coreos/pkg/dlopen/dlopen.go b/vendor/github.com/coreos/pkg/dlopen/dlopen.go new file mode 100644 index 000000000..23774f612 --- /dev/null +++ b/vendor/github.com/coreos/pkg/dlopen/dlopen.go @@ -0,0 +1,82 @@ +// Copyright 2016 CoreOS, Inc. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// Package dlopen provides some convenience functions to dlopen a library and +// get its symbols. +package dlopen + +// #cgo LDFLAGS: -ldl +// #include +// #include +import "C" +import ( + "errors" + "fmt" + "unsafe" +) + +var ErrSoNotFound = errors.New("unable to open a handle to the library") + +// LibHandle represents an open handle to a library (.so) +type LibHandle struct { + Handle unsafe.Pointer + Libname string +} + +// GetHandle tries to get a handle to a library (.so), attempting to access it +// by the names specified in libs and returning the first that is successfully +// opened. Callers are responsible for closing the handler. If no library can +// be successfully opened, an error is returned. +func GetHandle(libs []string) (*LibHandle, error) { + for _, name := range libs { + libname := C.CString(name) + defer C.free(unsafe.Pointer(libname)) + handle := C.dlopen(libname, C.RTLD_LAZY) + if handle != nil { + h := &LibHandle{ + Handle: handle, + Libname: name, + } + return h, nil + } + } + return nil, ErrSoNotFound +} + +// GetSymbolPointer takes a symbol name and returns a pointer to the symbol. +func (l *LibHandle) GetSymbolPointer(symbol string) (unsafe.Pointer, error) { + sym := C.CString(symbol) + defer C.free(unsafe.Pointer(sym)) + + C.dlerror() + p := C.dlsym(l.Handle, sym) + e := C.dlerror() + if e != nil { + return nil, fmt.Errorf("error resolving symbol %q: %v", symbol, errors.New(C.GoString(e))) + } + + return p, nil +} + +// Close closes a LibHandle. +func (l *LibHandle) Close() error { + C.dlerror() + C.dlclose(l.Handle) + e := C.dlerror() + if e != nil { + return fmt.Errorf("error closing %v: %v", l.Libname, errors.New(C.GoString(e))) + } + + return nil +} diff --git a/vendor/github.com/coreos/pkg/dlopen/dlopen_example.go b/vendor/github.com/coreos/pkg/dlopen/dlopen_example.go new file mode 100644 index 000000000..48a660104 --- /dev/null +++ b/vendor/github.com/coreos/pkg/dlopen/dlopen_example.go @@ -0,0 +1,56 @@ +// Copyright 2015 CoreOS, Inc. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. +// +// +build linux + +package dlopen + +// #include +// #include +// +// int +// my_strlen(void *f, const char *s) +// { +// size_t (*strlen)(const char *); +// +// strlen = (size_t (*)(const char *))f; +// return strlen(s); +// } +import "C" + +import ( + "fmt" + "unsafe" +) + +func strlen(libs []string, s string) (int, error) { + h, err := GetHandle(libs) + if err != nil { + return -1, fmt.Errorf(`couldn't get a handle to the library: %v`, err) + } + defer h.Close() + + f := "strlen" + cs := C.CString(s) + defer C.free(unsafe.Pointer(cs)) + + strlen, err := h.GetSymbolPointer(f) + if err != nil { + return -1, fmt.Errorf(`couldn't get symbol %q: %v`, f, err) + } + + len := C.my_strlen(strlen, cs) + + return int(len), nil +} diff --git a/vendor/github.com/docker/docker/pkg/mflag/LICENSE b/vendor/github.com/cyphar/filepath-securejoin/LICENSE similarity index 92% rename from vendor/github.com/docker/docker/pkg/mflag/LICENSE rename to vendor/github.com/cyphar/filepath-securejoin/LICENSE index 9b4f4a294..bec842f29 100644 --- a/vendor/github.com/docker/docker/pkg/mflag/LICENSE +++ b/vendor/github.com/cyphar/filepath-securejoin/LICENSE @@ -1,4 +1,5 @@ -Copyright (c) 2014-2016 The Docker & Go Authors. All rights reserved. +Copyright (C) 2014-2015 Docker Inc & Go Authors. All rights reserved. +Copyright (C) 2017 SUSE LLC. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are diff --git a/vendor/github.com/cyphar/filepath-securejoin/README.md b/vendor/github.com/cyphar/filepath-securejoin/README.md new file mode 100644 index 000000000..49b2baa9f --- /dev/null +++ b/vendor/github.com/cyphar/filepath-securejoin/README.md @@ -0,0 +1,65 @@ +## `filepath-securejoin` ## + +[![Build Status](https://travis-ci.org/cyphar/filepath-securejoin.svg?branch=master)](https://travis-ci.org/cyphar/filepath-securejoin) + +An implementation of `SecureJoin`, a [candidate for inclusion in the Go +standard library][go#20126]. The purpose of this function is to be a "secure" +alternative to `filepath.Join`, and in particular it provides certain +guarantees that are not provided by `filepath.Join`. + +This is the function prototype: + +```go +func SecureJoin(root, unsafePath string) (string, error) +``` + +This library **guarantees** the following: + +* If no error is set, the resulting string **must** be a child path of + `SecureJoin` and will not contain any symlink path components (they will all + be expanded). + +* When expanding symlinks, all symlink path components **must** be resolved + relative to the provided root. In particular, this can be considered a + userspace implementation of how `chroot(2)` operates on file paths. Note that + these symlinks will **not** be expanded lexically (`filepath.Clean` is not + called on the input before processing). + +* Non-existant path components are unaffected by `SecureJoin` (similar to + `filepath.EvalSymlinks`'s semantics). + +* The returned path will always be `filepath.Clean`ed and thus not contain any + `..` components. + +A (trivial) implementation of this function on GNU/Linux systems could be done +with the following (note that this requires root privileges and is far more +opaque than the implementation in this library, and also requires that +`readlink` is inside the `root` path): + +```go +package securejoin + +import ( + "os/exec" + "path/filepath" +) + +func SecureJoin(root, unsafePath string) (string, error) { + unsafePath = string(filepath.Separator) + unsafePath + cmd := exec.Command("chroot", root, + "readlink", "--canonicalize-missing", "--no-newline", unsafePath) + output, err := cmd.CombinedOutput() + if err != nil { + return "", err + } + expanded := string(output) + return filepath.Join(root, expanded), nil +} +``` + +[go#20126]: https://github.com/golang/go/issues/20126 + +### License ### + +The license of this project is the same as Go, which is a BSD 3-clause license +available in the `LICENSE` file. diff --git a/vendor/github.com/cyphar/filepath-securejoin/join.go b/vendor/github.com/cyphar/filepath-securejoin/join.go new file mode 100644 index 000000000..f20985479 --- /dev/null +++ b/vendor/github.com/cyphar/filepath-securejoin/join.go @@ -0,0 +1,135 @@ +// Copyright (C) 2014-2015 Docker Inc & Go Authors. All rights reserved. +// Copyright (C) 2017 SUSE LLC. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package securejoin is an implementation of the hopefully-soon-to-be-included +// SecureJoin helper that is meant to be part of the "path/filepath" package. +// The purpose of this project is to provide a PoC implementation to make the +// SecureJoin proposal (https://github.com/golang/go/issues/20126) more +// tangible. +package securejoin + +import ( + "bytes" + "fmt" + "os" + "path/filepath" + "strings" + "syscall" + + "github.com/pkg/errors" +) + +// ErrSymlinkLoop is returned by SecureJoinVFS when too many symlinks have been +// evaluated in attempting to securely join the two given paths. +var ErrSymlinkLoop = fmt.Errorf("SecureJoin: too many links") + +// IsNotExist tells you if err is an error that implies that either the path +// accessed does not exist (or path components don't exist). This is +// effectively a more broad version of os.IsNotExist. +func IsNotExist(err error) bool { + // If it's a bone-fide ENOENT just bail. + if os.IsNotExist(errors.Cause(err)) { + return true + } + + // Check that it's not actually an ENOTDIR, which in some cases is a more + // convoluted case of ENOENT (usually involving weird paths). + var errno error + switch err := errors.Cause(err).(type) { + case *os.PathError: + errno = err.Err + case *os.LinkError: + errno = err.Err + case *os.SyscallError: + errno = err.Err + } + return errno == syscall.ENOTDIR || errno == syscall.ENOENT +} + +// SecureJoinVFS joins the two given path components (similar to Join) except +// that the returned path is guaranteed to be scoped inside the provided root +// path (when evaluated). Any symbolic links in the path are evaluated with the +// given root treated as the root of the filesystem, similar to a chroot. The +// filesystem state is evaluated through the given VFS interface (if nil, the +// standard os.* family of functions are used). +// +// Note that the guarantees provided by this function only apply if the path +// components in the returned string are not modified (in other words are not +// replaced with symlinks on the filesystem) after this function has returned. +// Such a symlink race is necessarily out-of-scope of SecureJoin. +func SecureJoinVFS(root, unsafePath string, vfs VFS) (string, error) { + // Use the os.* VFS implementation if none was specified. + if vfs == nil { + vfs = osVFS{} + } + + var path bytes.Buffer + n := 0 + for unsafePath != "" { + if n > 255 { + return "", ErrSymlinkLoop + } + + // Next path component, p. + i := strings.IndexRune(unsafePath, filepath.Separator) + var p string + if i == -1 { + p, unsafePath = unsafePath, "" + } else { + p, unsafePath = unsafePath[:i], unsafePath[i+1:] + } + + // Create a cleaned path, using the lexical semantics of /../a, to + // create a "scoped" path component which can safely be joined to fullP + // for evaluation. At this point, path.String() doesn't contain any + // symlink components. + cleanP := filepath.Clean(string(filepath.Separator) + path.String() + p) + if cleanP == string(filepath.Separator) { + path.Reset() + continue + } + fullP := filepath.Clean(root + cleanP) + + // Figure out whether the path is a symlink. + fi, err := vfs.Lstat(fullP) + if err != nil && !IsNotExist(err) { + return "", err + } + // Treat non-existent path components the same as non-symlinks (we + // can't do any better here). + if IsNotExist(err) || fi.Mode()&os.ModeSymlink == 0 { + path.WriteString(p) + path.WriteRune(filepath.Separator) + continue + } + + // Only increment when we actually dereference a link. + n++ + + // It's a symlink, expand it by prepending it to the yet-unparsed path. + dest, err := vfs.Readlink(fullP) + if err != nil { + return "", err + } + // Absolute symlinks reset any work we've already done. + if filepath.IsAbs(dest) { + path.Reset() + } + unsafePath = dest + string(filepath.Separator) + unsafePath + } + + // We have to clean path.String() here because it may contain '..' + // components that are entirely lexical, but would be misleading otherwise. + // And finally do a final clean to ensure that root is also lexically + // clean. + fullP := filepath.Clean(string(filepath.Separator) + path.String()) + return filepath.Clean(root + fullP), nil +} + +// SecureJoin is a wrapper around SecureJoinVFS that just uses the os.* library +// of functions as the VFS. If in doubt, use this function over SecureJoinVFS. +func SecureJoin(root, unsafePath string) (string, error) { + return SecureJoinVFS(root, unsafePath, nil) +} diff --git a/vendor/github.com/cyphar/filepath-securejoin/vendor.conf b/vendor/github.com/cyphar/filepath-securejoin/vendor.conf new file mode 100644 index 000000000..66bb574b9 --- /dev/null +++ b/vendor/github.com/cyphar/filepath-securejoin/vendor.conf @@ -0,0 +1 @@ +github.com/pkg/errors v0.8.0 diff --git a/vendor/github.com/cyphar/filepath-securejoin/vfs.go b/vendor/github.com/cyphar/filepath-securejoin/vfs.go new file mode 100644 index 000000000..a82a5eae1 --- /dev/null +++ b/vendor/github.com/cyphar/filepath-securejoin/vfs.go @@ -0,0 +1,41 @@ +// Copyright (C) 2017 SUSE LLC. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package securejoin + +import "os" + +// In future this should be moved into a separate package, because now there +// are several projects (umoci and go-mtree) that are using this sort of +// interface. + +// VFS is the minimal interface necessary to use SecureJoinVFS. A nil VFS is +// equivalent to using the standard os.* family of functions. This is mainly +// used for the purposes of mock testing, but also can be used to otherwise use +// SecureJoin with VFS-like system. +type VFS interface { + // Lstat returns a FileInfo describing the named file. If the file is a + // symbolic link, the returned FileInfo describes the symbolic link. Lstat + // makes no attempt to follow the link. These semantics are identical to + // os.Lstat. + Lstat(name string) (os.FileInfo, error) + + // Readlink returns the destination of the named symbolic link. These + // semantics are identical to os.Readlink. + Readlink(name string) (string, error) +} + +// osVFS is the "nil" VFS, in that it just passes everything through to the os +// module. +type osVFS struct{} + +// Lstat returns a FileInfo describing the named file. If the file is a +// symbolic link, the returned FileInfo describes the symbolic link. Lstat +// makes no attempt to follow the link. These semantics are identical to +// os.Lstat. +func (o osVFS) Lstat(name string) (os.FileInfo, error) { return os.Lstat(name) } + +// Readlink returns the destination of the named symbolic link. These +// semantics are identical to os.Readlink. +func (o osVFS) Readlink(name string) (string, error) { return os.Readlink(name) } diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/go-units/LICENSE b/vendor/github.com/docker/cli/LICENSE similarity index 99% rename from vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/go-units/LICENSE rename to vendor/github.com/docker/cli/LICENSE index b55b37bc3..9c8e20ab8 100644 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/go-units/LICENSE +++ b/vendor/github.com/docker/cli/LICENSE @@ -176,7 +176,7 @@ END OF TERMS AND CONDITIONS - Copyright 2015 Docker, Inc. + Copyright 2013-2017 Docker, Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/NOTICE b/vendor/github.com/docker/cli/NOTICE similarity index 94% rename from vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/NOTICE rename to vendor/github.com/docker/cli/NOTICE index 6e6f469ab..0c74e15b0 100644 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/NOTICE +++ b/vendor/github.com/docker/cli/NOTICE @@ -1,5 +1,5 @@ Docker -Copyright 2012-2015 Docker, Inc. +Copyright 2012-2017 Docker, Inc. This product includes software developed at Docker, Inc. (https://www.docker.com). diff --git a/vendor/github.com/docker/cli/cli/config/configfile/file.go b/vendor/github.com/docker/cli/cli/config/configfile/file.go new file mode 100644 index 000000000..37e1533f4 --- /dev/null +++ b/vendor/github.com/docker/cli/cli/config/configfile/file.go @@ -0,0 +1,335 @@ +package configfile + +import ( + "encoding/base64" + "encoding/json" + "fmt" + "io" + "io/ioutil" + "os" + "path/filepath" + "strings" + + "github.com/docker/cli/cli/config/credentials" + "github.com/docker/cli/opts" + "github.com/docker/docker/api/types" + "github.com/pkg/errors" +) + +const ( + // This constant is only used for really old config files when the + // URL wasn't saved as part of the config file and it was just + // assumed to be this value. + defaultIndexServer = "https://index.docker.io/v1/" +) + +// ConfigFile ~/.docker/config.json file info +type ConfigFile struct { + AuthConfigs map[string]types.AuthConfig `json:"auths"` + HTTPHeaders map[string]string `json:"HttpHeaders,omitempty"` + PsFormat string `json:"psFormat,omitempty"` + ImagesFormat string `json:"imagesFormat,omitempty"` + NetworksFormat string `json:"networksFormat,omitempty"` + PluginsFormat string `json:"pluginsFormat,omitempty"` + VolumesFormat string `json:"volumesFormat,omitempty"` + StatsFormat string `json:"statsFormat,omitempty"` + DetachKeys string `json:"detachKeys,omitempty"` + CredentialsStore string `json:"credsStore,omitempty"` + CredentialHelpers map[string]string `json:"credHelpers,omitempty"` + Filename string `json:"-"` // Note: for internal use only + ServiceInspectFormat string `json:"serviceInspectFormat,omitempty"` + ServicesFormat string `json:"servicesFormat,omitempty"` + TasksFormat string `json:"tasksFormat,omitempty"` + SecretFormat string `json:"secretFormat,omitempty"` + ConfigFormat string `json:"configFormat,omitempty"` + NodesFormat string `json:"nodesFormat,omitempty"` + PruneFilters []string `json:"pruneFilters,omitempty"` + Proxies map[string]ProxyConfig `json:"proxies,omitempty"` + Experimental string `json:"experimental,omitempty"` + StackOrchestrator string `json:"stackOrchestrator,omitempty"` + Kubernetes *KubernetesConfig `json:"kubernetes,omitempty"` +} + +// ProxyConfig contains proxy configuration settings +type ProxyConfig struct { + HTTPProxy string `json:"httpProxy,omitempty"` + HTTPSProxy string `json:"httpsProxy,omitempty"` + NoProxy string `json:"noProxy,omitempty"` + FTPProxy string `json:"ftpProxy,omitempty"` +} + +// KubernetesConfig contains Kubernetes orchestrator settings +type KubernetesConfig struct { + AllNamespaces string `json:"allNamespaces,omitempty"` +} + +// New initializes an empty configuration file for the given filename 'fn' +func New(fn string) *ConfigFile { + return &ConfigFile{ + AuthConfigs: make(map[string]types.AuthConfig), + HTTPHeaders: make(map[string]string), + Filename: fn, + } +} + +// LegacyLoadFromReader reads the non-nested configuration data given and sets up the +// auth config information with given directory and populates the receiver object +func (configFile *ConfigFile) LegacyLoadFromReader(configData io.Reader) error { + b, err := ioutil.ReadAll(configData) + if err != nil { + return err + } + + if err := json.Unmarshal(b, &configFile.AuthConfigs); err != nil { + arr := strings.Split(string(b), "\n") + if len(arr) < 2 { + return errors.Errorf("The Auth config file is empty") + } + authConfig := types.AuthConfig{} + origAuth := strings.Split(arr[0], " = ") + if len(origAuth) != 2 { + return errors.Errorf("Invalid Auth config file") + } + authConfig.Username, authConfig.Password, err = decodeAuth(origAuth[1]) + if err != nil { + return err + } + authConfig.ServerAddress = defaultIndexServer + configFile.AuthConfigs[defaultIndexServer] = authConfig + } else { + for k, authConfig := range configFile.AuthConfigs { + authConfig.Username, authConfig.Password, err = decodeAuth(authConfig.Auth) + if err != nil { + return err + } + authConfig.Auth = "" + authConfig.ServerAddress = k + configFile.AuthConfigs[k] = authConfig + } + } + return nil +} + +// LoadFromReader reads the configuration data given and sets up the auth config +// information with given directory and populates the receiver object +func (configFile *ConfigFile) LoadFromReader(configData io.Reader) error { + if err := json.NewDecoder(configData).Decode(&configFile); err != nil { + return err + } + var err error + for addr, ac := range configFile.AuthConfigs { + ac.Username, ac.Password, err = decodeAuth(ac.Auth) + if err != nil { + return err + } + ac.Auth = "" + ac.ServerAddress = addr + configFile.AuthConfigs[addr] = ac + } + return checkKubernetesConfiguration(configFile.Kubernetes) +} + +// ContainsAuth returns whether there is authentication configured +// in this file or not. +func (configFile *ConfigFile) ContainsAuth() bool { + return configFile.CredentialsStore != "" || + len(configFile.CredentialHelpers) > 0 || + len(configFile.AuthConfigs) > 0 +} + +// GetAuthConfigs returns the mapping of repo to auth configuration +func (configFile *ConfigFile) GetAuthConfigs() map[string]types.AuthConfig { + return configFile.AuthConfigs +} + +// SaveToWriter encodes and writes out all the authorization information to +// the given writer +func (configFile *ConfigFile) SaveToWriter(writer io.Writer) error { + // Encode sensitive data into a new/temp struct + tmpAuthConfigs := make(map[string]types.AuthConfig, len(configFile.AuthConfigs)) + for k, authConfig := range configFile.AuthConfigs { + authCopy := authConfig + // encode and save the authstring, while blanking out the original fields + authCopy.Auth = encodeAuth(&authCopy) + authCopy.Username = "" + authCopy.Password = "" + authCopy.ServerAddress = "" + tmpAuthConfigs[k] = authCopy + } + + saveAuthConfigs := configFile.AuthConfigs + configFile.AuthConfigs = tmpAuthConfigs + defer func() { configFile.AuthConfigs = saveAuthConfigs }() + + data, err := json.MarshalIndent(configFile, "", "\t") + if err != nil { + return err + } + _, err = writer.Write(data) + return err +} + +// Save encodes and writes out all the authorization information +func (configFile *ConfigFile) Save() error { + if configFile.Filename == "" { + return errors.Errorf("Can't save config with empty filename") + } + + if err := os.MkdirAll(filepath.Dir(configFile.Filename), 0700); err != nil { + return err + } + f, err := os.OpenFile(configFile.Filename, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0600) + if err != nil { + return err + } + defer f.Close() + return configFile.SaveToWriter(f) +} + +// ParseProxyConfig computes proxy configuration by retrieving the config for the provided host and +// then checking this against any environment variables provided to the container +func (configFile *ConfigFile) ParseProxyConfig(host string, runOpts []string) map[string]*string { + var cfgKey string + + if _, ok := configFile.Proxies[host]; !ok { + cfgKey = "default" + } else { + cfgKey = host + } + + config := configFile.Proxies[cfgKey] + permitted := map[string]*string{ + "HTTP_PROXY": &config.HTTPProxy, + "HTTPS_PROXY": &config.HTTPSProxy, + "NO_PROXY": &config.NoProxy, + "FTP_PROXY": &config.FTPProxy, + } + m := opts.ConvertKVStringsToMapWithNil(runOpts) + for k := range permitted { + if *permitted[k] == "" { + continue + } + if _, ok := m[k]; !ok { + m[k] = permitted[k] + } + if _, ok := m[strings.ToLower(k)]; !ok { + m[strings.ToLower(k)] = permitted[k] + } + } + return m +} + +// encodeAuth creates a base64 encoded string to containing authorization information +func encodeAuth(authConfig *types.AuthConfig) string { + if authConfig.Username == "" && authConfig.Password == "" { + return "" + } + + authStr := authConfig.Username + ":" + authConfig.Password + msg := []byte(authStr) + encoded := make([]byte, base64.StdEncoding.EncodedLen(len(msg))) + base64.StdEncoding.Encode(encoded, msg) + return string(encoded) +} + +// decodeAuth decodes a base64 encoded string and returns username and password +func decodeAuth(authStr string) (string, string, error) { + if authStr == "" { + return "", "", nil + } + + decLen := base64.StdEncoding.DecodedLen(len(authStr)) + decoded := make([]byte, decLen) + authByte := []byte(authStr) + n, err := base64.StdEncoding.Decode(decoded, authByte) + if err != nil { + return "", "", err + } + if n > decLen { + return "", "", errors.Errorf("Something went wrong decoding auth config") + } + arr := strings.SplitN(string(decoded), ":", 2) + if len(arr) != 2 { + return "", "", errors.Errorf("Invalid auth configuration file") + } + password := strings.Trim(arr[1], "\x00") + return arr[0], password, nil +} + +// GetCredentialsStore returns a new credentials store from the settings in the +// configuration file +func (configFile *ConfigFile) GetCredentialsStore(registryHostname string) credentials.Store { + if helper := getConfiguredCredentialStore(configFile, registryHostname); helper != "" { + return newNativeStore(configFile, helper) + } + return credentials.NewFileStore(configFile) +} + +// var for unit testing. +var newNativeStore = func(configFile *ConfigFile, helperSuffix string) credentials.Store { + return credentials.NewNativeStore(configFile, helperSuffix) +} + +// GetAuthConfig for a repository from the credential store +func (configFile *ConfigFile) GetAuthConfig(registryHostname string) (types.AuthConfig, error) { + return configFile.GetCredentialsStore(registryHostname).Get(registryHostname) +} + +// getConfiguredCredentialStore returns the credential helper configured for the +// given registry, the default credsStore, or the empty string if neither are +// configured. +func getConfiguredCredentialStore(c *ConfigFile, registryHostname string) string { + if c.CredentialHelpers != nil && registryHostname != "" { + if helper, exists := c.CredentialHelpers[registryHostname]; exists { + return helper + } + } + return c.CredentialsStore +} + +// GetAllCredentials returns all of the credentials stored in all of the +// configured credential stores. +func (configFile *ConfigFile) GetAllCredentials() (map[string]types.AuthConfig, error) { + auths := make(map[string]types.AuthConfig) + addAll := func(from map[string]types.AuthConfig) { + for reg, ac := range from { + auths[reg] = ac + } + } + + defaultStore := configFile.GetCredentialsStore("") + newAuths, err := defaultStore.GetAll() + if err != nil { + return nil, err + } + addAll(newAuths) + + // Auth configs from a registry-specific helper should override those from the default store. + for registryHostname := range configFile.CredentialHelpers { + newAuth, err := configFile.GetAuthConfig(registryHostname) + if err != nil { + return nil, err + } + auths[registryHostname] = newAuth + } + return auths, nil +} + +// GetFilename returns the file name that this config file is based on. +func (configFile *ConfigFile) GetFilename() string { + return configFile.Filename +} + +func checkKubernetesConfiguration(kubeConfig *KubernetesConfig) error { + if kubeConfig == nil { + return nil + } + switch kubeConfig.AllNamespaces { + case "": + case "enabled": + case "disabled": + default: + return fmt.Errorf("invalid 'kubernetes.allNamespaces' value, should be 'enabled' or 'disabled': %s", kubeConfig.AllNamespaces) + } + return nil +} diff --git a/vendor/github.com/docker/cli/cli/config/credentials/credentials.go b/vendor/github.com/docker/cli/cli/config/credentials/credentials.go new file mode 100644 index 000000000..ca874cac5 --- /dev/null +++ b/vendor/github.com/docker/cli/cli/config/credentials/credentials.go @@ -0,0 +1,17 @@ +package credentials + +import ( + "github.com/docker/docker/api/types" +) + +// Store is the interface that any credentials store must implement. +type Store interface { + // Erase removes credentials from the store for a given server. + Erase(serverAddress string) error + // Get retrieves credentials from the store for a given server. + Get(serverAddress string) (types.AuthConfig, error) + // GetAll retrieves all the credentials from the store. + GetAll() (map[string]types.AuthConfig, error) + // Store saves credentials in the store. + Store(authConfig types.AuthConfig) error +} diff --git a/vendor/github.com/docker/cli/cli/config/credentials/default_store.go b/vendor/github.com/docker/cli/cli/config/credentials/default_store.go new file mode 100644 index 000000000..7a760f1a9 --- /dev/null +++ b/vendor/github.com/docker/cli/cli/config/credentials/default_store.go @@ -0,0 +1,21 @@ +package credentials + +import ( + "os/exec" +) + +// DetectDefaultStore return the default credentials store for the platform if +// the store executable is available. +func DetectDefaultStore(store string) string { + platformDefault := defaultCredentialsStore() + + // user defined or no default for platform + if store != "" || platformDefault == "" { + return store + } + + if _, err := exec.LookPath(remoteCredentialsPrefix + platformDefault); err == nil { + return platformDefault + } + return "" +} diff --git a/vendor/github.com/docker/cli/cli/config/credentials/default_store_darwin.go b/vendor/github.com/docker/cli/cli/config/credentials/default_store_darwin.go new file mode 100644 index 000000000..5d42dec62 --- /dev/null +++ b/vendor/github.com/docker/cli/cli/config/credentials/default_store_darwin.go @@ -0,0 +1,5 @@ +package credentials + +func defaultCredentialsStore() string { + return "osxkeychain" +} diff --git a/vendor/github.com/docker/cli/cli/config/credentials/default_store_linux.go b/vendor/github.com/docker/cli/cli/config/credentials/default_store_linux.go new file mode 100644 index 000000000..a9012c6d4 --- /dev/null +++ b/vendor/github.com/docker/cli/cli/config/credentials/default_store_linux.go @@ -0,0 +1,13 @@ +package credentials + +import ( + "os/exec" +) + +func defaultCredentialsStore() string { + if _, err := exec.LookPath("pass"); err == nil { + return "pass" + } + + return "secretservice" +} diff --git a/vendor/github.com/docker/cli/cli/config/credentials/default_store_unsupported.go b/vendor/github.com/docker/cli/cli/config/credentials/default_store_unsupported.go new file mode 100644 index 000000000..3028168ac --- /dev/null +++ b/vendor/github.com/docker/cli/cli/config/credentials/default_store_unsupported.go @@ -0,0 +1,7 @@ +// +build !windows,!darwin,!linux + +package credentials + +func defaultCredentialsStore() string { + return "" +} diff --git a/vendor/github.com/docker/cli/cli/config/credentials/default_store_windows.go b/vendor/github.com/docker/cli/cli/config/credentials/default_store_windows.go new file mode 100644 index 000000000..bb799ca61 --- /dev/null +++ b/vendor/github.com/docker/cli/cli/config/credentials/default_store_windows.go @@ -0,0 +1,5 @@ +package credentials + +func defaultCredentialsStore() string { + return "wincred" +} diff --git a/vendor/github.com/docker/cli/cli/config/credentials/file_store.go b/vendor/github.com/docker/cli/cli/config/credentials/file_store.go new file mode 100644 index 000000000..6ae681754 --- /dev/null +++ b/vendor/github.com/docker/cli/cli/config/credentials/file_store.go @@ -0,0 +1,64 @@ +package credentials + +import ( + "github.com/docker/docker/api/types" + "github.com/docker/docker/registry" +) + +type store interface { + Save() error + GetAuthConfigs() map[string]types.AuthConfig + GetFilename() string +} + +// fileStore implements a credentials store using +// the docker configuration file to keep the credentials in plain text. +type fileStore struct { + file store +} + +// NewFileStore creates a new file credentials store. +func NewFileStore(file store) Store { + return &fileStore{file: file} +} + +// Erase removes the given credentials from the file store. +func (c *fileStore) Erase(serverAddress string) error { + delete(c.file.GetAuthConfigs(), serverAddress) + return c.file.Save() +} + +// Get retrieves credentials for a specific server from the file store. +func (c *fileStore) Get(serverAddress string) (types.AuthConfig, error) { + authConfig, ok := c.file.GetAuthConfigs()[serverAddress] + if !ok { + // Maybe they have a legacy config file, we will iterate the keys converting + // them to the new format and testing + for r, ac := range c.file.GetAuthConfigs() { + if serverAddress == registry.ConvertToHostname(r) { + return ac, nil + } + } + + authConfig = types.AuthConfig{} + } + return authConfig, nil +} + +func (c *fileStore) GetAll() (map[string]types.AuthConfig, error) { + return c.file.GetAuthConfigs(), nil +} + +// Store saves the given credentials in the file store. +func (c *fileStore) Store(authConfig types.AuthConfig) error { + c.file.GetAuthConfigs()[authConfig.ServerAddress] = authConfig + return c.file.Save() +} + +func (c *fileStore) GetFilename() string { + return c.file.GetFilename() +} + +func (c *fileStore) IsFileStore() bool { + return true +} diff --git a/vendor/github.com/docker/cli/cli/config/credentials/native_store.go b/vendor/github.com/docker/cli/cli/config/credentials/native_store.go new file mode 100644 index 000000000..ef3aab4ad --- /dev/null +++ b/vendor/github.com/docker/cli/cli/config/credentials/native_store.go @@ -0,0 +1,143 @@ +package credentials + +import ( + "github.com/docker/docker-credential-helpers/client" + "github.com/docker/docker-credential-helpers/credentials" + "github.com/docker/docker/api/types" +) + +const ( + remoteCredentialsPrefix = "docker-credential-" + tokenUsername = "" +) + +// nativeStore implements a credentials store +// using native keychain to keep credentials secure. +// It piggybacks into a file store to keep users' emails. +type nativeStore struct { + programFunc client.ProgramFunc + fileStore Store +} + +// NewNativeStore creates a new native store that +// uses a remote helper program to manage credentials. +func NewNativeStore(file store, helperSuffix string) Store { + name := remoteCredentialsPrefix + helperSuffix + return &nativeStore{ + programFunc: client.NewShellProgramFunc(name), + fileStore: NewFileStore(file), + } +} + +// Erase removes the given credentials from the native store. +func (c *nativeStore) Erase(serverAddress string) error { + if err := client.Erase(c.programFunc, serverAddress); err != nil { + return err + } + + // Fallback to plain text store to remove email + return c.fileStore.Erase(serverAddress) +} + +// Get retrieves credentials for a specific server from the native store. +func (c *nativeStore) Get(serverAddress string) (types.AuthConfig, error) { + // load user email if it exist or an empty auth config. + auth, _ := c.fileStore.Get(serverAddress) + + creds, err := c.getCredentialsFromStore(serverAddress) + if err != nil { + return auth, err + } + auth.Username = creds.Username + auth.IdentityToken = creds.IdentityToken + auth.Password = creds.Password + + return auth, nil +} + +// GetAll retrieves all the credentials from the native store. +func (c *nativeStore) GetAll() (map[string]types.AuthConfig, error) { + auths, err := c.listCredentialsInStore() + if err != nil { + return nil, err + } + + // Emails are only stored in the file store. + // This call can be safely eliminated when emails are removed. + fileConfigs, _ := c.fileStore.GetAll() + + authConfigs := make(map[string]types.AuthConfig) + for registry := range auths { + creds, err := c.getCredentialsFromStore(registry) + if err != nil { + return nil, err + } + ac := fileConfigs[registry] // might contain Email + ac.Username = creds.Username + ac.Password = creds.Password + ac.IdentityToken = creds.IdentityToken + authConfigs[registry] = ac + } + + return authConfigs, nil +} + +// Store saves the given credentials in the file store. +func (c *nativeStore) Store(authConfig types.AuthConfig) error { + if err := c.storeCredentialsInStore(authConfig); err != nil { + return err + } + authConfig.Username = "" + authConfig.Password = "" + authConfig.IdentityToken = "" + + // Fallback to old credential in plain text to save only the email + return c.fileStore.Store(authConfig) +} + +// storeCredentialsInStore executes the command to store the credentials in the native store. +func (c *nativeStore) storeCredentialsInStore(config types.AuthConfig) error { + creds := &credentials.Credentials{ + ServerURL: config.ServerAddress, + Username: config.Username, + Secret: config.Password, + } + + if config.IdentityToken != "" { + creds.Username = tokenUsername + creds.Secret = config.IdentityToken + } + + return client.Store(c.programFunc, creds) +} + +// getCredentialsFromStore executes the command to get the credentials from the native store. +func (c *nativeStore) getCredentialsFromStore(serverAddress string) (types.AuthConfig, error) { + var ret types.AuthConfig + + creds, err := client.Get(c.programFunc, serverAddress) + if err != nil { + if credentials.IsErrCredentialsNotFound(err) { + // do not return an error if the credentials are not + // in the keychain. Let docker ask for new credentials. + return ret, nil + } + return ret, err + } + + if creds.Username == tokenUsername { + ret.IdentityToken = creds.Secret + } else { + ret.Password = creds.Secret + ret.Username = creds.Username + } + + ret.ServerAddress = serverAddress + return ret, nil +} + +// listCredentialsInStore returns a listing of stored credentials as a map of +// URL -> username. +func (c *nativeStore) listCredentialsInStore() (map[string]string, error) { + return client.List(c.programFunc) +} diff --git a/vendor/github.com/docker/cli/opts/config.go b/vendor/github.com/docker/cli/opts/config.go new file mode 100644 index 000000000..82fd2bce4 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/config.go @@ -0,0 +1,98 @@ +package opts + +import ( + "encoding/csv" + "fmt" + "os" + "strconv" + "strings" + + swarmtypes "github.com/docker/docker/api/types/swarm" +) + +// ConfigOpt is a Value type for parsing configs +type ConfigOpt struct { + values []*swarmtypes.ConfigReference +} + +// Set a new config value +func (o *ConfigOpt) Set(value string) error { + csvReader := csv.NewReader(strings.NewReader(value)) + fields, err := csvReader.Read() + if err != nil { + return err + } + + options := &swarmtypes.ConfigReference{ + File: &swarmtypes.ConfigReferenceFileTarget{ + UID: "0", + GID: "0", + Mode: 0444, + }, + } + + // support a simple syntax of --config foo + if len(fields) == 1 { + options.File.Name = fields[0] + options.ConfigName = fields[0] + o.values = append(o.values, options) + return nil + } + + for _, field := range fields { + parts := strings.SplitN(field, "=", 2) + key := strings.ToLower(parts[0]) + + if len(parts) != 2 { + return fmt.Errorf("invalid field '%s' must be a key=value pair", field) + } + + value := parts[1] + switch key { + case "source", "src": + options.ConfigName = value + case "target": + options.File.Name = value + case "uid": + options.File.UID = value + case "gid": + options.File.GID = value + case "mode": + m, err := strconv.ParseUint(value, 0, 32) + if err != nil { + return fmt.Errorf("invalid mode specified: %v", err) + } + + options.File.Mode = os.FileMode(m) + default: + return fmt.Errorf("invalid field in config request: %s", key) + } + } + + if options.ConfigName == "" { + return fmt.Errorf("source is required") + } + + o.values = append(o.values, options) + return nil +} + +// Type returns the type of this option +func (o *ConfigOpt) Type() string { + return "config" +} + +// String returns a string repr of this option +func (o *ConfigOpt) String() string { + configs := []string{} + for _, config := range o.values { + repr := fmt.Sprintf("%s -> %s", config.ConfigName, config.File.Name) + configs = append(configs, repr) + } + return strings.Join(configs, ", ") +} + +// Value returns the config requests +func (o *ConfigOpt) Value() []*swarmtypes.ConfigReference { + return o.values +} diff --git a/vendor/github.com/docker/cli/opts/duration.go b/vendor/github.com/docker/cli/opts/duration.go new file mode 100644 index 000000000..5dc6eeaa7 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/duration.go @@ -0,0 +1,64 @@ +package opts + +import ( + "time" + + "github.com/pkg/errors" +) + +// PositiveDurationOpt is an option type for time.Duration that uses a pointer. +// It behave similarly to DurationOpt but only allows positive duration values. +type PositiveDurationOpt struct { + DurationOpt +} + +// Set a new value on the option. Setting a negative duration value will cause +// an error to be returned. +func (d *PositiveDurationOpt) Set(s string) error { + err := d.DurationOpt.Set(s) + if err != nil { + return err + } + if *d.DurationOpt.value < 0 { + return errors.Errorf("duration cannot be negative") + } + return nil +} + +// DurationOpt is an option type for time.Duration that uses a pointer. This +// allows us to get nil values outside, instead of defaulting to 0 +type DurationOpt struct { + value *time.Duration +} + +// NewDurationOpt creates a DurationOpt with the specified duration +func NewDurationOpt(value *time.Duration) *DurationOpt { + return &DurationOpt{ + value: value, + } +} + +// Set a new value on the option +func (d *DurationOpt) Set(s string) error { + v, err := time.ParseDuration(s) + d.value = &v + return err +} + +// Type returns the type of this option, which will be displayed in `--help` output +func (d *DurationOpt) Type() string { + return "duration" +} + +// String returns a string repr of this option +func (d *DurationOpt) String() string { + if d.value != nil { + return d.value.String() + } + return "" +} + +// Value returns the time.Duration +func (d *DurationOpt) Value() *time.Duration { + return d.value +} diff --git a/vendor/github.com/docker/cli/opts/env.go b/vendor/github.com/docker/cli/opts/env.go new file mode 100644 index 000000000..e6ddd7330 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/env.go @@ -0,0 +1,46 @@ +package opts + +import ( + "fmt" + "os" + "runtime" + "strings" +) + +// ValidateEnv validates an environment variable and returns it. +// If no value is specified, it returns the current value using os.Getenv. +// +// As on ParseEnvFile and related to #16585, environment variable names +// are not validate what so ever, it's up to application inside docker +// to validate them or not. +// +// The only validation here is to check if name is empty, per #25099 +func ValidateEnv(val string) (string, error) { + arr := strings.Split(val, "=") + if arr[0] == "" { + return "", fmt.Errorf("invalid environment variable: %s", val) + } + if len(arr) > 1 { + return val, nil + } + if !doesEnvExist(val) { + return val, nil + } + return fmt.Sprintf("%s=%s", val, os.Getenv(val)), nil +} + +func doesEnvExist(name string) bool { + for _, entry := range os.Environ() { + parts := strings.SplitN(entry, "=", 2) + if runtime.GOOS == "windows" { + // Environment variable are case-insensitive on Windows. PaTh, path and PATH are equivalent. + if strings.EqualFold(parts[0], name) { + return true + } + } + if parts[0] == name { + return true + } + } + return false +} diff --git a/vendor/github.com/docker/cli/opts/envfile.go b/vendor/github.com/docker/cli/opts/envfile.go new file mode 100644 index 000000000..10054c896 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/envfile.go @@ -0,0 +1,22 @@ +package opts + +import ( + "os" +) + +// ParseEnvFile reads a file with environment variables enumerated by lines +// +// ``Environment variable names used by the utilities in the Shell and +// Utilities volume of IEEE Std 1003.1-2001 consist solely of uppercase +// letters, digits, and the '_' (underscore) from the characters defined in +// Portable Character Set and do not begin with a digit. *But*, other +// characters may be permitted by an implementation; applications shall +// tolerate the presence of such names.'' +// -- http://pubs.opengroup.org/onlinepubs/009695399/basedefs/xbd_chap08.html +// +// As of #16585, it's up to application inside docker to validate or not +// environment variables, that's why we just strip leading whitespace and +// nothing more. +func ParseEnvFile(filename string) ([]string, error) { + return parseKeyValueFile(filename, os.Getenv) +} diff --git a/vendor/github.com/docker/cli/opts/file.go b/vendor/github.com/docker/cli/opts/file.go new file mode 100644 index 000000000..281905949 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/file.go @@ -0,0 +1,71 @@ +package opts + +import ( + "bufio" + "bytes" + "fmt" + "os" + "strings" + "unicode" + "unicode/utf8" +) + +var whiteSpaces = " \t" + +// ErrBadKey typed error for bad environment variable +type ErrBadKey struct { + msg string +} + +func (e ErrBadKey) Error() string { + return fmt.Sprintf("poorly formatted environment: %s", e.msg) +} + +func parseKeyValueFile(filename string, emptyFn func(string) string) ([]string, error) { + fh, err := os.Open(filename) + if err != nil { + return []string{}, err + } + defer fh.Close() + + lines := []string{} + scanner := bufio.NewScanner(fh) + currentLine := 0 + utf8bom := []byte{0xEF, 0xBB, 0xBF} + for scanner.Scan() { + scannedBytes := scanner.Bytes() + if !utf8.Valid(scannedBytes) { + return []string{}, fmt.Errorf("env file %s contains invalid utf8 bytes at line %d: %v", filename, currentLine+1, scannedBytes) + } + // We trim UTF8 BOM + if currentLine == 0 { + scannedBytes = bytes.TrimPrefix(scannedBytes, utf8bom) + } + // trim the line from all leading whitespace first + line := strings.TrimLeftFunc(string(scannedBytes), unicode.IsSpace) + currentLine++ + // line is not empty, and not starting with '#' + if len(line) > 0 && !strings.HasPrefix(line, "#") { + data := strings.SplitN(line, "=", 2) + + // trim the front of a variable, but nothing else + variable := strings.TrimLeft(data[0], whiteSpaces) + if strings.ContainsAny(variable, whiteSpaces) { + return []string{}, ErrBadKey{fmt.Sprintf("variable '%s' has white spaces", variable)} + } + + if len(data) > 1 { + // pass the value through, no trimming + lines = append(lines, fmt.Sprintf("%s=%s", variable, data[1])) + } else { + var value string + if emptyFn != nil { + value = emptyFn(line) + } + // if only a pass-through variable is given, clean it up. + lines = append(lines, fmt.Sprintf("%s=%s", strings.TrimSpace(line), value)) + } + } + } + return lines, scanner.Err() +} diff --git a/vendor/github.com/docker/cli/opts/hosts.go b/vendor/github.com/docker/cli/opts/hosts.go new file mode 100644 index 000000000..594cccf2f --- /dev/null +++ b/vendor/github.com/docker/cli/opts/hosts.go @@ -0,0 +1,165 @@ +package opts + +import ( + "fmt" + "net" + "net/url" + "strconv" + "strings" +) + +var ( + // DefaultHTTPPort Default HTTP Port used if only the protocol is provided to -H flag e.g. dockerd -H tcp:// + // These are the IANA registered port numbers for use with Docker + // see http://www.iana.org/assignments/service-names-port-numbers/service-names-port-numbers.xhtml?search=docker + DefaultHTTPPort = 2375 // Default HTTP Port + // DefaultTLSHTTPPort Default HTTP Port used when TLS enabled + DefaultTLSHTTPPort = 2376 // Default TLS encrypted HTTP Port + // DefaultUnixSocket Path for the unix socket. + // Docker daemon by default always listens on the default unix socket + DefaultUnixSocket = "/var/run/docker.sock" + // DefaultTCPHost constant defines the default host string used by docker on Windows + DefaultTCPHost = fmt.Sprintf("tcp://%s:%d", DefaultHTTPHost, DefaultHTTPPort) + // DefaultTLSHost constant defines the default host string used by docker for TLS sockets + DefaultTLSHost = fmt.Sprintf("tcp://%s:%d", DefaultHTTPHost, DefaultTLSHTTPPort) + // DefaultNamedPipe defines the default named pipe used by docker on Windows + DefaultNamedPipe = `//./pipe/docker_engine` +) + +// ValidateHost validates that the specified string is a valid host and returns it. +func ValidateHost(val string) (string, error) { + host := strings.TrimSpace(val) + // The empty string means default and is not handled by parseDockerDaemonHost + if host != "" { + _, err := parseDockerDaemonHost(host) + if err != nil { + return val, err + } + } + // Note: unlike most flag validators, we don't return the mutated value here + // we need to know what the user entered later (using ParseHost) to adjust for TLS + return val, nil +} + +// ParseHost and set defaults for a Daemon host string +func ParseHost(defaultToTLS bool, val string) (string, error) { + host := strings.TrimSpace(val) + if host == "" { + if defaultToTLS { + host = DefaultTLSHost + } else { + host = DefaultHost + } + } else { + var err error + host, err = parseDockerDaemonHost(host) + if err != nil { + return val, err + } + } + return host, nil +} + +// parseDockerDaemonHost parses the specified address and returns an address that will be used as the host. +// Depending of the address specified, this may return one of the global Default* strings defined in hosts.go. +func parseDockerDaemonHost(addr string) (string, error) { + addrParts := strings.SplitN(addr, "://", 2) + if len(addrParts) == 1 && addrParts[0] != "" { + addrParts = []string{"tcp", addrParts[0]} + } + + switch addrParts[0] { + case "tcp": + return ParseTCPAddr(addrParts[1], DefaultTCPHost) + case "unix": + return parseSimpleProtoAddr("unix", addrParts[1], DefaultUnixSocket) + case "npipe": + return parseSimpleProtoAddr("npipe", addrParts[1], DefaultNamedPipe) + case "fd": + return addr, nil + default: + return "", fmt.Errorf("Invalid bind address format: %s", addr) + } +} + +// parseSimpleProtoAddr parses and validates that the specified address is a valid +// socket address for simple protocols like unix and npipe. It returns a formatted +// socket address, either using the address parsed from addr, or the contents of +// defaultAddr if addr is a blank string. +func parseSimpleProtoAddr(proto, addr, defaultAddr string) (string, error) { + addr = strings.TrimPrefix(addr, proto+"://") + if strings.Contains(addr, "://") { + return "", fmt.Errorf("Invalid proto, expected %s: %s", proto, addr) + } + if addr == "" { + addr = defaultAddr + } + return fmt.Sprintf("%s://%s", proto, addr), nil +} + +// ParseTCPAddr parses and validates that the specified address is a valid TCP +// address. It returns a formatted TCP address, either using the address parsed +// from tryAddr, or the contents of defaultAddr if tryAddr is a blank string. +// tryAddr is expected to have already been Trim()'d +// defaultAddr must be in the full `tcp://host:port` form +func ParseTCPAddr(tryAddr string, defaultAddr string) (string, error) { + if tryAddr == "" || tryAddr == "tcp://" { + return defaultAddr, nil + } + addr := strings.TrimPrefix(tryAddr, "tcp://") + if strings.Contains(addr, "://") || addr == "" { + return "", fmt.Errorf("Invalid proto, expected tcp: %s", tryAddr) + } + + defaultAddr = strings.TrimPrefix(defaultAddr, "tcp://") + defaultHost, defaultPort, err := net.SplitHostPort(defaultAddr) + if err != nil { + return "", err + } + // url.Parse fails for trailing colon on IPv6 brackets on Go 1.5, but + // not 1.4. See https://github.com/golang/go/issues/12200 and + // https://github.com/golang/go/issues/6530. + if strings.HasSuffix(addr, "]:") { + addr += defaultPort + } + + u, err := url.Parse("tcp://" + addr) + if err != nil { + return "", err + } + host, port, err := net.SplitHostPort(u.Host) + if err != nil { + // try port addition once + host, port, err = net.SplitHostPort(net.JoinHostPort(u.Host, defaultPort)) + } + if err != nil { + return "", fmt.Errorf("Invalid bind address format: %s", tryAddr) + } + + if host == "" { + host = defaultHost + } + if port == "" { + port = defaultPort + } + p, err := strconv.Atoi(port) + if err != nil && p == 0 { + return "", fmt.Errorf("Invalid bind address format: %s", tryAddr) + } + + return fmt.Sprintf("tcp://%s%s", net.JoinHostPort(host, port), u.Path), nil +} + +// ValidateExtraHost validates that the specified string is a valid extrahost and returns it. +// ExtraHost is in the form of name:ip where the ip has to be a valid ip (IPv4 or IPv6). +func ValidateExtraHost(val string) (string, error) { + // allow for IPv6 addresses in extra hosts by only splitting on first ":" + arr := strings.SplitN(val, ":", 2) + if len(arr) != 2 || len(arr[0]) == 0 { + return "", fmt.Errorf("bad format for add-host: %q", val) + } + if _, err := ValidateIPAddress(arr[1]); err != nil { + return "", fmt.Errorf("invalid IP address in add-host: %q", arr[1]) + } + return val, nil +} diff --git a/vendor/github.com/docker/cli/opts/hosts_unix.go b/vendor/github.com/docker/cli/opts/hosts_unix.go new file mode 100644 index 000000000..611407a9d --- /dev/null +++ b/vendor/github.com/docker/cli/opts/hosts_unix.go @@ -0,0 +1,8 @@ +// +build !windows + +package opts + +import "fmt" + +// DefaultHost constant defines the default host string used by docker on other hosts than Windows +var DefaultHost = fmt.Sprintf("unix://%s", DefaultUnixSocket) diff --git a/vendor/github.com/docker/cli/opts/hosts_windows.go b/vendor/github.com/docker/cli/opts/hosts_windows.go new file mode 100644 index 000000000..7c239e00f --- /dev/null +++ b/vendor/github.com/docker/cli/opts/hosts_windows.go @@ -0,0 +1,6 @@ +// +build windows + +package opts + +// DefaultHost constant defines the default host string used by docker on Windows +var DefaultHost = "npipe://" + DefaultNamedPipe diff --git a/vendor/github.com/docker/cli/opts/ip.go b/vendor/github.com/docker/cli/opts/ip.go new file mode 100644 index 000000000..fb03b5011 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/ip.go @@ -0,0 +1,47 @@ +package opts + +import ( + "fmt" + "net" +) + +// IPOpt holds an IP. It is used to store values from CLI flags. +type IPOpt struct { + *net.IP +} + +// NewIPOpt creates a new IPOpt from a reference net.IP and a +// string representation of an IP. If the string is not a valid +// IP it will fallback to the specified reference. +func NewIPOpt(ref *net.IP, defaultVal string) *IPOpt { + o := &IPOpt{ + IP: ref, + } + o.Set(defaultVal) + return o +} + +// Set sets an IPv4 or IPv6 address from a given string. If the given +// string is not parseable as an IP address it returns an error. +func (o *IPOpt) Set(val string) error { + ip := net.ParseIP(val) + if ip == nil { + return fmt.Errorf("%s is not an ip address", val) + } + *o.IP = ip + return nil +} + +// String returns the IP address stored in the IPOpt. If stored IP is a +// nil pointer, it returns an empty string. +func (o *IPOpt) String() string { + if *o.IP == nil { + return "" + } + return o.IP.String() +} + +// Type returns the type of the option +func (o *IPOpt) Type() string { + return "ip" +} diff --git a/vendor/github.com/docker/cli/opts/mount.go b/vendor/github.com/docker/cli/opts/mount.go new file mode 100644 index 000000000..3aa984942 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/mount.go @@ -0,0 +1,174 @@ +package opts + +import ( + "encoding/csv" + "fmt" + "os" + "strconv" + "strings" + + mounttypes "github.com/docker/docker/api/types/mount" + "github.com/docker/go-units" +) + +// MountOpt is a Value type for parsing mounts +type MountOpt struct { + values []mounttypes.Mount +} + +// Set a new mount value +// nolint: gocyclo +func (m *MountOpt) Set(value string) error { + csvReader := csv.NewReader(strings.NewReader(value)) + fields, err := csvReader.Read() + if err != nil { + return err + } + + mount := mounttypes.Mount{} + + volumeOptions := func() *mounttypes.VolumeOptions { + if mount.VolumeOptions == nil { + mount.VolumeOptions = &mounttypes.VolumeOptions{ + Labels: make(map[string]string), + } + } + if mount.VolumeOptions.DriverConfig == nil { + mount.VolumeOptions.DriverConfig = &mounttypes.Driver{} + } + return mount.VolumeOptions + } + + bindOptions := func() *mounttypes.BindOptions { + if mount.BindOptions == nil { + mount.BindOptions = new(mounttypes.BindOptions) + } + return mount.BindOptions + } + + tmpfsOptions := func() *mounttypes.TmpfsOptions { + if mount.TmpfsOptions == nil { + mount.TmpfsOptions = new(mounttypes.TmpfsOptions) + } + return mount.TmpfsOptions + } + + setValueOnMap := func(target map[string]string, value string) { + parts := strings.SplitN(value, "=", 2) + if len(parts) == 1 { + target[value] = "" + } else { + target[parts[0]] = parts[1] + } + } + + mount.Type = mounttypes.TypeVolume // default to volume mounts + // Set writable as the default + for _, field := range fields { + parts := strings.SplitN(field, "=", 2) + key := strings.ToLower(parts[0]) + + if len(parts) == 1 { + switch key { + case "readonly", "ro": + mount.ReadOnly = true + continue + case "volume-nocopy": + volumeOptions().NoCopy = true + continue + } + } + + if len(parts) != 2 { + return fmt.Errorf("invalid field '%s' must be a key=value pair", field) + } + + value := parts[1] + switch key { + case "type": + mount.Type = mounttypes.Type(strings.ToLower(value)) + case "source", "src": + mount.Source = value + case "target", "dst", "destination": + mount.Target = value + case "readonly", "ro": + mount.ReadOnly, err = strconv.ParseBool(value) + if err != nil { + return fmt.Errorf("invalid value for %s: %s", key, value) + } + case "consistency": + mount.Consistency = mounttypes.Consistency(strings.ToLower(value)) + case "bind-propagation": + bindOptions().Propagation = mounttypes.Propagation(strings.ToLower(value)) + case "volume-nocopy": + volumeOptions().NoCopy, err = strconv.ParseBool(value) + if err != nil { + return fmt.Errorf("invalid value for volume-nocopy: %s", value) + } + case "volume-label": + setValueOnMap(volumeOptions().Labels, value) + case "volume-driver": + volumeOptions().DriverConfig.Name = value + case "volume-opt": + if volumeOptions().DriverConfig.Options == nil { + volumeOptions().DriverConfig.Options = make(map[string]string) + } + setValueOnMap(volumeOptions().DriverConfig.Options, value) + case "tmpfs-size": + sizeBytes, err := units.RAMInBytes(value) + if err != nil { + return fmt.Errorf("invalid value for %s: %s", key, value) + } + tmpfsOptions().SizeBytes = sizeBytes + case "tmpfs-mode": + ui64, err := strconv.ParseUint(value, 8, 32) + if err != nil { + return fmt.Errorf("invalid value for %s: %s", key, value) + } + tmpfsOptions().Mode = os.FileMode(ui64) + default: + return fmt.Errorf("unexpected key '%s' in '%s'", key, field) + } + } + + if mount.Type == "" { + return fmt.Errorf("type is required") + } + + if mount.Target == "" { + return fmt.Errorf("target is required") + } + + if mount.VolumeOptions != nil && mount.Type != mounttypes.TypeVolume { + return fmt.Errorf("cannot mix 'volume-*' options with mount type '%s'", mount.Type) + } + if mount.BindOptions != nil && mount.Type != mounttypes.TypeBind { + return fmt.Errorf("cannot mix 'bind-*' options with mount type '%s'", mount.Type) + } + if mount.TmpfsOptions != nil && mount.Type != mounttypes.TypeTmpfs { + return fmt.Errorf("cannot mix 'tmpfs-*' options with mount type '%s'", mount.Type) + } + + m.values = append(m.values, mount) + return nil +} + +// Type returns the type of this option +func (m *MountOpt) Type() string { + return "mount" +} + +// String returns a string repr of this option +func (m *MountOpt) String() string { + mounts := []string{} + for _, mount := range m.values { + repr := fmt.Sprintf("%s %s %s", mount.Type, mount.Source, mount.Target) + mounts = append(mounts, repr) + } + return strings.Join(mounts, ", ") +} + +// Value returns the mounts +func (m *MountOpt) Value() []mounttypes.Mount { + return m.values +} diff --git a/vendor/github.com/docker/cli/opts/network.go b/vendor/github.com/docker/cli/opts/network.go new file mode 100644 index 000000000..ec4967ff3 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/network.go @@ -0,0 +1,106 @@ +package opts + +import ( + "encoding/csv" + "fmt" + "regexp" + "strings" +) + +const ( + networkOptName = "name" + networkOptAlias = "alias" + driverOpt = "driver-opt" +) + +// NetworkAttachmentOpts represents the network options for endpoint creation +type NetworkAttachmentOpts struct { + Target string + Aliases []string + DriverOpts map[string]string +} + +// NetworkOpt represents a network config in swarm mode. +type NetworkOpt struct { + options []NetworkAttachmentOpts +} + +// Set networkopts value +func (n *NetworkOpt) Set(value string) error { + longSyntax, err := regexp.MatchString(`\w+=\w+(,\w+=\w+)*`, value) + if err != nil { + return err + } + + var netOpt NetworkAttachmentOpts + if longSyntax { + csvReader := csv.NewReader(strings.NewReader(value)) + fields, err := csvReader.Read() + if err != nil { + return err + } + + netOpt.Aliases = []string{} + for _, field := range fields { + parts := strings.SplitN(field, "=", 2) + + if len(parts) < 2 { + return fmt.Errorf("invalid field %s", field) + } + + key := strings.TrimSpace(strings.ToLower(parts[0])) + value := strings.TrimSpace(strings.ToLower(parts[1])) + + switch key { + case networkOptName: + netOpt.Target = value + case networkOptAlias: + netOpt.Aliases = append(netOpt.Aliases, value) + case driverOpt: + key, value, err = parseDriverOpt(value) + if err == nil { + if netOpt.DriverOpts == nil { + netOpt.DriverOpts = make(map[string]string) + } + netOpt.DriverOpts[key] = value + } else { + return err + } + default: + return fmt.Errorf("invalid field key %s", key) + } + } + if len(netOpt.Target) == 0 { + return fmt.Errorf("network name/id is not specified") + } + } else { + netOpt.Target = value + } + n.options = append(n.options, netOpt) + return nil +} + +// Type returns the type of this option +func (n *NetworkOpt) Type() string { + return "network" +} + +// Value returns the networkopts +func (n *NetworkOpt) Value() []NetworkAttachmentOpts { + return n.options +} + +// String returns the network opts as a string +func (n *NetworkOpt) String() string { + return "" +} + +func parseDriverOpt(driverOpt string) (string, string, error) { + parts := strings.SplitN(driverOpt, "=", 2) + if len(parts) != 2 { + return "", "", fmt.Errorf("invalid key value pair format in driver options") + } + key := strings.TrimSpace(strings.ToLower(parts[0])) + value := strings.TrimSpace(strings.ToLower(parts[1])) + return key, value, nil +} diff --git a/vendor/github.com/docker/cli/opts/opts.go b/vendor/github.com/docker/cli/opts/opts.go new file mode 100644 index 000000000..51519e03b --- /dev/null +++ b/vendor/github.com/docker/cli/opts/opts.go @@ -0,0 +1,524 @@ +package opts + +import ( + "fmt" + "math/big" + "net" + "path" + "regexp" + "strconv" + "strings" + + "github.com/docker/docker/api/types/filters" + units "github.com/docker/go-units" +) + +var ( + alphaRegexp = regexp.MustCompile(`[a-zA-Z]`) + domainRegexp = regexp.MustCompile(`^(:?(:?[a-zA-Z0-9]|(:?[a-zA-Z0-9][a-zA-Z0-9\-]*[a-zA-Z0-9]))(:?\.(:?[a-zA-Z0-9]|(:?[a-zA-Z0-9][a-zA-Z0-9\-]*[a-zA-Z0-9])))*)\.?\s*$`) +) + +// ListOpts holds a list of values and a validation function. +type ListOpts struct { + values *[]string + validator ValidatorFctType +} + +// NewListOpts creates a new ListOpts with the specified validator. +func NewListOpts(validator ValidatorFctType) ListOpts { + var values []string + return *NewListOptsRef(&values, validator) +} + +// NewListOptsRef creates a new ListOpts with the specified values and validator. +func NewListOptsRef(values *[]string, validator ValidatorFctType) *ListOpts { + return &ListOpts{ + values: values, + validator: validator, + } +} + +func (opts *ListOpts) String() string { + if len(*opts.values) == 0 { + return "" + } + return fmt.Sprintf("%v", *opts.values) +} + +// Set validates if needed the input value and adds it to the +// internal slice. +func (opts *ListOpts) Set(value string) error { + if opts.validator != nil { + v, err := opts.validator(value) + if err != nil { + return err + } + value = v + } + (*opts.values) = append((*opts.values), value) + return nil +} + +// Delete removes the specified element from the slice. +func (opts *ListOpts) Delete(key string) { + for i, k := range *opts.values { + if k == key { + (*opts.values) = append((*opts.values)[:i], (*opts.values)[i+1:]...) + return + } + } +} + +// GetMap returns the content of values in a map in order to avoid +// duplicates. +func (opts *ListOpts) GetMap() map[string]struct{} { + ret := make(map[string]struct{}) + for _, k := range *opts.values { + ret[k] = struct{}{} + } + return ret +} + +// GetAll returns the values of slice. +func (opts *ListOpts) GetAll() []string { + return (*opts.values) +} + +// GetAllOrEmpty returns the values of the slice +// or an empty slice when there are no values. +func (opts *ListOpts) GetAllOrEmpty() []string { + v := *opts.values + if v == nil { + return make([]string, 0) + } + return v +} + +// Get checks the existence of the specified key. +func (opts *ListOpts) Get(key string) bool { + for _, k := range *opts.values { + if k == key { + return true + } + } + return false +} + +// Len returns the amount of element in the slice. +func (opts *ListOpts) Len() int { + return len((*opts.values)) +} + +// Type returns a string name for this Option type +func (opts *ListOpts) Type() string { + return "list" +} + +// WithValidator returns the ListOpts with validator set. +func (opts *ListOpts) WithValidator(validator ValidatorFctType) *ListOpts { + opts.validator = validator + return opts +} + +// NamedOption is an interface that list and map options +// with names implement. +type NamedOption interface { + Name() string +} + +// NamedListOpts is a ListOpts with a configuration name. +// This struct is useful to keep reference to the assigned +// field name in the internal configuration struct. +type NamedListOpts struct { + name string + ListOpts +} + +var _ NamedOption = &NamedListOpts{} + +// NewNamedListOptsRef creates a reference to a new NamedListOpts struct. +func NewNamedListOptsRef(name string, values *[]string, validator ValidatorFctType) *NamedListOpts { + return &NamedListOpts{ + name: name, + ListOpts: *NewListOptsRef(values, validator), + } +} + +// Name returns the name of the NamedListOpts in the configuration. +func (o *NamedListOpts) Name() string { + return o.name +} + +// MapOpts holds a map of values and a validation function. +type MapOpts struct { + values map[string]string + validator ValidatorFctType +} + +// Set validates if needed the input value and add it to the +// internal map, by splitting on '='. +func (opts *MapOpts) Set(value string) error { + if opts.validator != nil { + v, err := opts.validator(value) + if err != nil { + return err + } + value = v + } + vals := strings.SplitN(value, "=", 2) + if len(vals) == 1 { + (opts.values)[vals[0]] = "" + } else { + (opts.values)[vals[0]] = vals[1] + } + return nil +} + +// GetAll returns the values of MapOpts as a map. +func (opts *MapOpts) GetAll() map[string]string { + return opts.values +} + +func (opts *MapOpts) String() string { + return fmt.Sprintf("%v", opts.values) +} + +// Type returns a string name for this Option type +func (opts *MapOpts) Type() string { + return "map" +} + +// NewMapOpts creates a new MapOpts with the specified map of values and a validator. +func NewMapOpts(values map[string]string, validator ValidatorFctType) *MapOpts { + if values == nil { + values = make(map[string]string) + } + return &MapOpts{ + values: values, + validator: validator, + } +} + +// NamedMapOpts is a MapOpts struct with a configuration name. +// This struct is useful to keep reference to the assigned +// field name in the internal configuration struct. +type NamedMapOpts struct { + name string + MapOpts +} + +var _ NamedOption = &NamedMapOpts{} + +// NewNamedMapOpts creates a reference to a new NamedMapOpts struct. +func NewNamedMapOpts(name string, values map[string]string, validator ValidatorFctType) *NamedMapOpts { + return &NamedMapOpts{ + name: name, + MapOpts: *NewMapOpts(values, validator), + } +} + +// Name returns the name of the NamedMapOpts in the configuration. +func (o *NamedMapOpts) Name() string { + return o.name +} + +// ValidatorFctType defines a validator function that returns a validated string and/or an error. +type ValidatorFctType func(val string) (string, error) + +// ValidatorFctListType defines a validator function that returns a validated list of string and/or an error +type ValidatorFctListType func(val string) ([]string, error) + +// ValidateIPAddress validates an Ip address. +func ValidateIPAddress(val string) (string, error) { + var ip = net.ParseIP(strings.TrimSpace(val)) + if ip != nil { + return ip.String(), nil + } + return "", fmt.Errorf("%s is not an ip address", val) +} + +// ValidateMACAddress validates a MAC address. +func ValidateMACAddress(val string) (string, error) { + _, err := net.ParseMAC(strings.TrimSpace(val)) + if err != nil { + return "", err + } + return val, nil +} + +// ValidateDNSSearch validates domain for resolvconf search configuration. +// A zero length domain is represented by a dot (.). +func ValidateDNSSearch(val string) (string, error) { + if val = strings.Trim(val, " "); val == "." { + return val, nil + } + return validateDomain(val) +} + +func validateDomain(val string) (string, error) { + if alphaRegexp.FindString(val) == "" { + return "", fmt.Errorf("%s is not a valid domain", val) + } + ns := domainRegexp.FindSubmatch([]byte(val)) + if len(ns) > 0 && len(ns[1]) < 255 { + return string(ns[1]), nil + } + return "", fmt.Errorf("%s is not a valid domain", val) +} + +// ValidateLabel validates that the specified string is a valid label, and returns it. +// Labels are in the form on key=value. +func ValidateLabel(val string) (string, error) { + if strings.Count(val, "=") < 1 { + return "", fmt.Errorf("bad attribute format: %s", val) + } + return val, nil +} + +// ValidateSysctl validates a sysctl and returns it. +func ValidateSysctl(val string) (string, error) { + validSysctlMap := map[string]bool{ + "kernel.msgmax": true, + "kernel.msgmnb": true, + "kernel.msgmni": true, + "kernel.sem": true, + "kernel.shmall": true, + "kernel.shmmax": true, + "kernel.shmmni": true, + "kernel.shm_rmid_forced": true, + } + validSysctlPrefixes := []string{ + "net.", + "fs.mqueue.", + } + arr := strings.Split(val, "=") + if len(arr) < 2 { + return "", fmt.Errorf("sysctl '%s' is not whitelisted", val) + } + if validSysctlMap[arr[0]] { + return val, nil + } + + for _, vp := range validSysctlPrefixes { + if strings.HasPrefix(arr[0], vp) { + return val, nil + } + } + return "", fmt.Errorf("sysctl '%s' is not whitelisted", val) +} + +// FilterOpt is a flag type for validating filters +type FilterOpt struct { + filter filters.Args +} + +// NewFilterOpt returns a new FilterOpt +func NewFilterOpt() FilterOpt { + return FilterOpt{filter: filters.NewArgs()} +} + +func (o *FilterOpt) String() string { + repr, err := filters.ToParam(o.filter) + if err != nil { + return "invalid filters" + } + return repr +} + +// Set sets the value of the opt by parsing the command line value +func (o *FilterOpt) Set(value string) error { + var err error + o.filter, err = filters.ParseFlag(value, o.filter) + return err +} + +// Type returns the option type +func (o *FilterOpt) Type() string { + return "filter" +} + +// Value returns the value of this option +func (o *FilterOpt) Value() filters.Args { + return o.filter +} + +// NanoCPUs is a type for fixed point fractional number. +type NanoCPUs int64 + +// String returns the string format of the number +func (c *NanoCPUs) String() string { + if *c == 0 { + return "" + } + return big.NewRat(c.Value(), 1e9).FloatString(3) +} + +// Set sets the value of the NanoCPU by passing a string +func (c *NanoCPUs) Set(value string) error { + cpus, err := ParseCPUs(value) + *c = NanoCPUs(cpus) + return err +} + +// Type returns the type +func (c *NanoCPUs) Type() string { + return "decimal" +} + +// Value returns the value in int64 +func (c *NanoCPUs) Value() int64 { + return int64(*c) +} + +// ParseCPUs takes a string ratio and returns an integer value of nano cpus +func ParseCPUs(value string) (int64, error) { + cpu, ok := new(big.Rat).SetString(value) + if !ok { + return 0, fmt.Errorf("failed to parse %v as a rational number", value) + } + nano := cpu.Mul(cpu, big.NewRat(1e9, 1)) + if !nano.IsInt() { + return 0, fmt.Errorf("value is too precise") + } + return nano.Num().Int64(), nil +} + +// ParseLink parses and validates the specified string as a link format (name:alias) +func ParseLink(val string) (string, string, error) { + if val == "" { + return "", "", fmt.Errorf("empty string specified for links") + } + arr := strings.Split(val, ":") + if len(arr) > 2 { + return "", "", fmt.Errorf("bad format for links: %s", val) + } + if len(arr) == 1 { + return val, val, nil + } + // This is kept because we can actually get a HostConfig with links + // from an already created container and the format is not `foo:bar` + // but `/foo:/c1/bar` + if strings.HasPrefix(arr[0], "/") { + _, alias := path.Split(arr[1]) + return arr[0][1:], alias, nil + } + return arr[0], arr[1], nil +} + +// ValidateLink validates that the specified string has a valid link format (containerName:alias). +func ValidateLink(val string) (string, error) { + _, _, err := ParseLink(val) + return val, err +} + +// MemBytes is a type for human readable memory bytes (like 128M, 2g, etc) +type MemBytes int64 + +// String returns the string format of the human readable memory bytes +func (m *MemBytes) String() string { + // NOTE: In spf13/pflag/flag.go, "0" is considered as "zero value" while "0 B" is not. + // We return "0" in case value is 0 here so that the default value is hidden. + // (Sometimes "default 0 B" is actually misleading) + if m.Value() != 0 { + return units.BytesSize(float64(m.Value())) + } + return "0" +} + +// Set sets the value of the MemBytes by passing a string +func (m *MemBytes) Set(value string) error { + val, err := units.RAMInBytes(value) + *m = MemBytes(val) + return err +} + +// Type returns the type +func (m *MemBytes) Type() string { + return "bytes" +} + +// Value returns the value in int64 +func (m *MemBytes) Value() int64 { + return int64(*m) +} + +// UnmarshalJSON is the customized unmarshaler for MemBytes +func (m *MemBytes) UnmarshalJSON(s []byte) error { + if len(s) <= 2 || s[0] != '"' || s[len(s)-1] != '"' { + return fmt.Errorf("invalid size: %q", s) + } + val, err := units.RAMInBytes(string(s[1 : len(s)-1])) + *m = MemBytes(val) + return err +} + +// MemSwapBytes is a type for human readable memory bytes (like 128M, 2g, etc). +// It differs from MemBytes in that -1 is valid and the default. +type MemSwapBytes int64 + +// Set sets the value of the MemSwapBytes by passing a string +func (m *MemSwapBytes) Set(value string) error { + if value == "-1" { + *m = MemSwapBytes(-1) + return nil + } + val, err := units.RAMInBytes(value) + *m = MemSwapBytes(val) + return err +} + +// Type returns the type +func (m *MemSwapBytes) Type() string { + return "bytes" +} + +// Value returns the value in int64 +func (m *MemSwapBytes) Value() int64 { + return int64(*m) +} + +func (m *MemSwapBytes) String() string { + b := MemBytes(*m) + return b.String() +} + +// UnmarshalJSON is the customized unmarshaler for MemSwapBytes +func (m *MemSwapBytes) UnmarshalJSON(s []byte) error { + b := MemBytes(*m) + return b.UnmarshalJSON(s) +} + +// NullableBool is a type for tri-state boolean options +type NullableBool struct { + b *bool +} + +// Type returns the type +func (n *NullableBool) Type() string { + return "" +} + +// Value returns the value in *bool +func (n *NullableBool) Value() *bool { + return n.b +} + +// Set sets the value. If value is empty string or "auto", nil is set. +// Otherwise true or false are set based on flag.Bool behavior. +func (n *NullableBool) Set(value string) error { + if value != "auto" && value != "" { + b, err := strconv.ParseBool(value) + if err != nil { + return err + } + n.b = &b + } + return nil +} + +func (n *NullableBool) String() string { + if n.b == nil { + return "auto" + } + return strconv.FormatBool(*n.b) +} diff --git a/vendor/github.com/docker/cli/opts/opts_unix.go b/vendor/github.com/docker/cli/opts/opts_unix.go new file mode 100644 index 000000000..2766a43a0 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/opts_unix.go @@ -0,0 +1,6 @@ +// +build !windows + +package opts + +// DefaultHTTPHost Default HTTP Host used if only port is provided to -H flag e.g. dockerd -H tcp://:8080 +const DefaultHTTPHost = "localhost" diff --git a/vendor/github.com/docker/cli/opts/opts_windows.go b/vendor/github.com/docker/cli/opts/opts_windows.go new file mode 100644 index 000000000..98b7251a9 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/opts_windows.go @@ -0,0 +1,56 @@ +package opts + +// TODO Windows. Identify bug in GOLang 1.5.1+ and/or Windows Server 2016 TP5. +// @jhowardmsft, @swernli. +// +// On Windows, this mitigates a problem with the default options of running +// a docker client against a local docker daemon on TP5. +// +// What was found that if the default host is "localhost", even if the client +// (and daemon as this is local) is not physically on a network, and the DNS +// cache is flushed (ipconfig /flushdns), then the client will pause for +// exactly one second when connecting to the daemon for calls. For example +// using docker run windowsservercore cmd, the CLI will send a create followed +// by an attach. You see the delay between the attach finishing and the attach +// being seen by the daemon. +// +// Here's some daemon debug logs with additional debug spew put in. The +// AfterWriteJSON log is the very last thing the daemon does as part of the +// create call. The POST /attach is the second CLI call. Notice the second +// time gap. +// +// time="2015-11-06T13:38:37.259627400-08:00" level=debug msg="After createRootfs" +// time="2015-11-06T13:38:37.263626300-08:00" level=debug msg="After setHostConfig" +// time="2015-11-06T13:38:37.267631200-08:00" level=debug msg="before createContainerPl...." +// time="2015-11-06T13:38:37.271629500-08:00" level=debug msg=ToDiskLocking.... +// time="2015-11-06T13:38:37.275643200-08:00" level=debug msg="loggin event...." +// time="2015-11-06T13:38:37.277627600-08:00" level=debug msg="logged event...." +// time="2015-11-06T13:38:37.279631800-08:00" level=debug msg="In defer func" +// time="2015-11-06T13:38:37.282628100-08:00" level=debug msg="After daemon.create" +// time="2015-11-06T13:38:37.286651700-08:00" level=debug msg="return 2" +// time="2015-11-06T13:38:37.289629500-08:00" level=debug msg="Returned from daemon.ContainerCreate" +// time="2015-11-06T13:38:37.311629100-08:00" level=debug msg="After WriteJSON" +// ... 1 second gap here.... +// time="2015-11-06T13:38:38.317866200-08:00" level=debug msg="Calling POST /v1.22/containers/984758282b842f779e805664b2c95d563adc9a979c8a3973e68c807843ee4757/attach" +// time="2015-11-06T13:38:38.326882500-08:00" level=info msg="POST /v1.22/containers/984758282b842f779e805664b2c95d563adc9a979c8a3973e68c807843ee4757/attach?stderr=1&stdin=1&stdout=1&stream=1" +// +// We suspect this is either a bug introduced in GOLang 1.5.1, or that a change +// in GOLang 1.5.1 (from 1.4.3) is exposing a bug in Windows. In theory, +// the Windows networking stack is supposed to resolve "localhost" internally, +// without hitting DNS, or even reading the hosts file (which is why localhost +// is commented out in the hosts file on Windows). +// +// We have validated that working around this using the actual IPv4 localhost +// address does not cause the delay. +// +// This does not occur with the docker client built with 1.4.3 on the same +// Windows build, regardless of whether the daemon is built using 1.5.1 +// or 1.4.3. It does not occur on Linux. We also verified we see the same thing +// on a cross-compiled Windows binary (from Linux). +// +// Final note: This is a mitigation, not a 'real' fix. It is still susceptible +// to the delay if a user were to do 'docker run -H=tcp://localhost:2375...' +// explicitly. + +// DefaultHTTPHost Default HTTP Host used if only port is provided to -H flag e.g. dockerd -H tcp://:8080 +const DefaultHTTPHost = "127.0.0.1" diff --git a/vendor/github.com/docker/cli/opts/parse.go b/vendor/github.com/docker/cli/opts/parse.go new file mode 100644 index 000000000..679759ded --- /dev/null +++ b/vendor/github.com/docker/cli/opts/parse.go @@ -0,0 +1,99 @@ +package opts + +import ( + "fmt" + "os" + "strconv" + "strings" + + "github.com/docker/docker/api/types/container" +) + +// ReadKVStrings reads a file of line terminated key=value pairs, and overrides any keys +// present in the file with additional pairs specified in the override parameter +func ReadKVStrings(files []string, override []string) ([]string, error) { + return readKVStrings(files, override, nil) +} + +// ReadKVEnvStrings reads a file of line terminated key=value pairs, and overrides any keys +// present in the file with additional pairs specified in the override parameter. +// If a key has no value, it will get the value from the environment. +func ReadKVEnvStrings(files []string, override []string) ([]string, error) { + return readKVStrings(files, override, os.Getenv) +} + +func readKVStrings(files []string, override []string, emptyFn func(string) string) ([]string, error) { + variables := []string{} + for _, ef := range files { + parsedVars, err := parseKeyValueFile(ef, emptyFn) + if err != nil { + return nil, err + } + variables = append(variables, parsedVars...) + } + // parse the '-e' and '--env' after, to allow override + variables = append(variables, override...) + + return variables, nil +} + +// ConvertKVStringsToMap converts ["key=value"] to {"key":"value"} +func ConvertKVStringsToMap(values []string) map[string]string { + result := make(map[string]string, len(values)) + for _, value := range values { + kv := strings.SplitN(value, "=", 2) + if len(kv) == 1 { + result[kv[0]] = "" + } else { + result[kv[0]] = kv[1] + } + } + + return result +} + +// ConvertKVStringsToMapWithNil converts ["key=value"] to {"key":"value"} +// but set unset keys to nil - meaning the ones with no "=" in them. +// We use this in cases where we need to distinguish between +// FOO= and FOO +// where the latter case just means FOO was mentioned but not given a value +func ConvertKVStringsToMapWithNil(values []string) map[string]*string { + result := make(map[string]*string, len(values)) + for _, value := range values { + kv := strings.SplitN(value, "=", 2) + if len(kv) == 1 { + result[kv[0]] = nil + } else { + result[kv[0]] = &kv[1] + } + } + + return result +} + +// ParseRestartPolicy returns the parsed policy or an error indicating what is incorrect +func ParseRestartPolicy(policy string) (container.RestartPolicy, error) { + p := container.RestartPolicy{} + + if policy == "" { + return p, nil + } + + parts := strings.Split(policy, ":") + + if len(parts) > 2 { + return p, fmt.Errorf("invalid restart policy format") + } + if len(parts) == 2 { + count, err := strconv.Atoi(parts[1]) + if err != nil { + return p, fmt.Errorf("maximum retry count must be an integer") + } + + p.MaximumRetryCount = count + } + + p.Name = parts[0] + + return p, nil +} diff --git a/vendor/github.com/docker/cli/opts/port.go b/vendor/github.com/docker/cli/opts/port.go new file mode 100644 index 000000000..201aefafc --- /dev/null +++ b/vendor/github.com/docker/cli/opts/port.go @@ -0,0 +1,167 @@ +package opts + +import ( + "encoding/csv" + "fmt" + "regexp" + "strconv" + "strings" + + "github.com/docker/docker/api/types/swarm" + "github.com/docker/go-connections/nat" + "github.com/sirupsen/logrus" +) + +const ( + portOptTargetPort = "target" + portOptPublishedPort = "published" + portOptProtocol = "protocol" + portOptMode = "mode" +) + +// PortOpt represents a port config in swarm mode. +type PortOpt struct { + ports []swarm.PortConfig +} + +// Set a new port value +// nolint: gocyclo +func (p *PortOpt) Set(value string) error { + longSyntax, err := regexp.MatchString(`\w+=\w+(,\w+=\w+)*`, value) + if err != nil { + return err + } + if longSyntax { + csvReader := csv.NewReader(strings.NewReader(value)) + fields, err := csvReader.Read() + if err != nil { + return err + } + + pConfig := swarm.PortConfig{} + for _, field := range fields { + parts := strings.SplitN(field, "=", 2) + if len(parts) != 2 { + return fmt.Errorf("invalid field %s", field) + } + + key := strings.ToLower(parts[0]) + value := strings.ToLower(parts[1]) + + switch key { + case portOptProtocol: + if value != string(swarm.PortConfigProtocolTCP) && value != string(swarm.PortConfigProtocolUDP) && value != string(swarm.PortConfigProtocolSCTP) { + return fmt.Errorf("invalid protocol value %s", value) + } + + pConfig.Protocol = swarm.PortConfigProtocol(value) + case portOptMode: + if value != string(swarm.PortConfigPublishModeIngress) && value != string(swarm.PortConfigPublishModeHost) { + return fmt.Errorf("invalid publish mode value %s", value) + } + + pConfig.PublishMode = swarm.PortConfigPublishMode(value) + case portOptTargetPort: + tPort, err := strconv.ParseUint(value, 10, 16) + if err != nil { + return err + } + + pConfig.TargetPort = uint32(tPort) + case portOptPublishedPort: + pPort, err := strconv.ParseUint(value, 10, 16) + if err != nil { + return err + } + + pConfig.PublishedPort = uint32(pPort) + default: + return fmt.Errorf("invalid field key %s", key) + } + } + + if pConfig.TargetPort == 0 { + return fmt.Errorf("missing mandatory field %q", portOptTargetPort) + } + + if pConfig.PublishMode == "" { + pConfig.PublishMode = swarm.PortConfigPublishModeIngress + } + + if pConfig.Protocol == "" { + pConfig.Protocol = swarm.PortConfigProtocolTCP + } + + p.ports = append(p.ports, pConfig) + } else { + // short syntax + portConfigs := []swarm.PortConfig{} + ports, portBindingMap, err := nat.ParsePortSpecs([]string{value}) + if err != nil { + return err + } + for _, portBindings := range portBindingMap { + for _, portBinding := range portBindings { + if portBinding.HostIP != "" { + return fmt.Errorf("hostip is not supported") + } + } + } + + for port := range ports { + portConfig, err := ConvertPortToPortConfig(port, portBindingMap) + if err != nil { + return err + } + portConfigs = append(portConfigs, portConfig...) + } + p.ports = append(p.ports, portConfigs...) + } + return nil +} + +// Type returns the type of this option +func (p *PortOpt) Type() string { + return "port" +} + +// String returns a string repr of this option +func (p *PortOpt) String() string { + ports := []string{} + for _, port := range p.ports { + repr := fmt.Sprintf("%v:%v/%s/%s", port.PublishedPort, port.TargetPort, port.Protocol, port.PublishMode) + ports = append(ports, repr) + } + return strings.Join(ports, ", ") +} + +// Value returns the ports +func (p *PortOpt) Value() []swarm.PortConfig { + return p.ports +} + +// ConvertPortToPortConfig converts ports to the swarm type +func ConvertPortToPortConfig( + port nat.Port, + portBindings map[nat.Port][]nat.PortBinding, +) ([]swarm.PortConfig, error) { + ports := []swarm.PortConfig{} + + for _, binding := range portBindings[port] { + if binding.HostIP != "" && binding.HostIP != "0.0.0.0" { + logrus.Warnf("ignoring IP-address (%s:%s:%s) service will listen on '0.0.0.0'", binding.HostIP, binding.HostPort, port) + } + hostPort, err := strconv.ParseUint(binding.HostPort, 10, 16) + if err != nil && binding.HostPort != "" { + return nil, fmt.Errorf("invalid hostport binding (%s) for port (%s)", binding.HostPort, port.Port()) + } + ports = append(ports, swarm.PortConfig{ + //TODO Name: ? + Protocol: swarm.PortConfigProtocol(strings.ToLower(port.Proto())), + TargetPort: uint32(port.Int()), + PublishedPort: uint32(hostPort), + PublishMode: swarm.PortConfigPublishModeIngress, + }) + } + return ports, nil +} diff --git a/vendor/github.com/docker/cli/opts/quotedstring.go b/vendor/github.com/docker/cli/opts/quotedstring.go new file mode 100644 index 000000000..09c68a526 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/quotedstring.go @@ -0,0 +1,37 @@ +package opts + +// QuotedString is a string that may have extra quotes around the value. The +// quotes are stripped from the value. +type QuotedString struct { + value *string +} + +// Set sets a new value +func (s *QuotedString) Set(val string) error { + *s.value = trimQuotes(val) + return nil +} + +// Type returns the type of the value +func (s *QuotedString) Type() string { + return "string" +} + +func (s *QuotedString) String() string { + return *s.value +} + +func trimQuotes(value string) string { + lastIndex := len(value) - 1 + for _, char := range []byte{'\'', '"'} { + if value[0] == char && value[lastIndex] == char { + return value[1:lastIndex] + } + } + return value +} + +// NewQuotedString returns a new quoted string option +func NewQuotedString(value *string) *QuotedString { + return &QuotedString{value: value} +} diff --git a/vendor/github.com/docker/cli/opts/runtime.go b/vendor/github.com/docker/cli/opts/runtime.go new file mode 100644 index 000000000..4361b3ce0 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/runtime.go @@ -0,0 +1,79 @@ +package opts + +import ( + "fmt" + "strings" + + "github.com/docker/docker/api/types" +) + +// RuntimeOpt defines a map of Runtimes +type RuntimeOpt struct { + name string + stockRuntimeName string + values *map[string]types.Runtime +} + +// NewNamedRuntimeOpt creates a new RuntimeOpt +func NewNamedRuntimeOpt(name string, ref *map[string]types.Runtime, stockRuntime string) *RuntimeOpt { + if ref == nil { + ref = &map[string]types.Runtime{} + } + return &RuntimeOpt{name: name, values: ref, stockRuntimeName: stockRuntime} +} + +// Name returns the name of the NamedListOpts in the configuration. +func (o *RuntimeOpt) Name() string { + return o.name +} + +// Set validates and updates the list of Runtimes +func (o *RuntimeOpt) Set(val string) error { + parts := strings.SplitN(val, "=", 2) + if len(parts) != 2 { + return fmt.Errorf("invalid runtime argument: %s", val) + } + + parts[0] = strings.TrimSpace(parts[0]) + parts[1] = strings.TrimSpace(parts[1]) + if parts[0] == "" || parts[1] == "" { + return fmt.Errorf("invalid runtime argument: %s", val) + } + + parts[0] = strings.ToLower(parts[0]) + if parts[0] == o.stockRuntimeName { + return fmt.Errorf("runtime name '%s' is reserved", o.stockRuntimeName) + } + + if _, ok := (*o.values)[parts[0]]; ok { + return fmt.Errorf("runtime '%s' was already defined", parts[0]) + } + + (*o.values)[parts[0]] = types.Runtime{Path: parts[1]} + + return nil +} + +// String returns Runtime values as a string. +func (o *RuntimeOpt) String() string { + var out []string + for k := range *o.values { + out = append(out, k) + } + + return fmt.Sprintf("%v", out) +} + +// GetMap returns a map of Runtimes (name: path) +func (o *RuntimeOpt) GetMap() map[string]types.Runtime { + if o.values != nil { + return *o.values + } + + return map[string]types.Runtime{} +} + +// Type returns the type of the option +func (o *RuntimeOpt) Type() string { + return "runtime" +} diff --git a/vendor/github.com/docker/cli/opts/secret.go b/vendor/github.com/docker/cli/opts/secret.go new file mode 100644 index 000000000..a1fde54d9 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/secret.go @@ -0,0 +1,98 @@ +package opts + +import ( + "encoding/csv" + "fmt" + "os" + "strconv" + "strings" + + swarmtypes "github.com/docker/docker/api/types/swarm" +) + +// SecretOpt is a Value type for parsing secrets +type SecretOpt struct { + values []*swarmtypes.SecretReference +} + +// Set a new secret value +func (o *SecretOpt) Set(value string) error { + csvReader := csv.NewReader(strings.NewReader(value)) + fields, err := csvReader.Read() + if err != nil { + return err + } + + options := &swarmtypes.SecretReference{ + File: &swarmtypes.SecretReferenceFileTarget{ + UID: "0", + GID: "0", + Mode: 0444, + }, + } + + // support a simple syntax of --secret foo + if len(fields) == 1 { + options.File.Name = fields[0] + options.SecretName = fields[0] + o.values = append(o.values, options) + return nil + } + + for _, field := range fields { + parts := strings.SplitN(field, "=", 2) + key := strings.ToLower(parts[0]) + + if len(parts) != 2 { + return fmt.Errorf("invalid field '%s' must be a key=value pair", field) + } + + value := parts[1] + switch key { + case "source", "src": + options.SecretName = value + case "target": + options.File.Name = value + case "uid": + options.File.UID = value + case "gid": + options.File.GID = value + case "mode": + m, err := strconv.ParseUint(value, 0, 32) + if err != nil { + return fmt.Errorf("invalid mode specified: %v", err) + } + + options.File.Mode = os.FileMode(m) + default: + return fmt.Errorf("invalid field in secret request: %s", key) + } + } + + if options.SecretName == "" { + return fmt.Errorf("source is required") + } + + o.values = append(o.values, options) + return nil +} + +// Type returns the type of this option +func (o *SecretOpt) Type() string { + return "secret" +} + +// String returns a string repr of this option +func (o *SecretOpt) String() string { + secrets := []string{} + for _, secret := range o.values { + repr := fmt.Sprintf("%s -> %s", secret.SecretName, secret.File.Name) + secrets = append(secrets, repr) + } + return strings.Join(secrets, ", ") +} + +// Value returns the secret requests +func (o *SecretOpt) Value() []*swarmtypes.SecretReference { + return o.values +} diff --git a/vendor/github.com/docker/cli/opts/throttledevice.go b/vendor/github.com/docker/cli/opts/throttledevice.go new file mode 100644 index 000000000..0959efae3 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/throttledevice.go @@ -0,0 +1,108 @@ +package opts + +import ( + "fmt" + "strconv" + "strings" + + "github.com/docker/docker/api/types/blkiodev" + "github.com/docker/go-units" +) + +// ValidatorThrottleFctType defines a validator function that returns a validated struct and/or an error. +type ValidatorThrottleFctType func(val string) (*blkiodev.ThrottleDevice, error) + +// ValidateThrottleBpsDevice validates that the specified string has a valid device-rate format. +func ValidateThrottleBpsDevice(val string) (*blkiodev.ThrottleDevice, error) { + split := strings.SplitN(val, ":", 2) + if len(split) != 2 { + return nil, fmt.Errorf("bad format: %s", val) + } + if !strings.HasPrefix(split[0], "/dev/") { + return nil, fmt.Errorf("bad format for device path: %s", val) + } + rate, err := units.RAMInBytes(split[1]) + if err != nil { + return nil, fmt.Errorf("invalid rate for device: %s. The correct format is :[]. Number must be a positive integer. Unit is optional and can be kb, mb, or gb", val) + } + if rate < 0 { + return nil, fmt.Errorf("invalid rate for device: %s. The correct format is :[]. Number must be a positive integer. Unit is optional and can be kb, mb, or gb", val) + } + + return &blkiodev.ThrottleDevice{ + Path: split[0], + Rate: uint64(rate), + }, nil +} + +// ValidateThrottleIOpsDevice validates that the specified string has a valid device-rate format. +func ValidateThrottleIOpsDevice(val string) (*blkiodev.ThrottleDevice, error) { + split := strings.SplitN(val, ":", 2) + if len(split) != 2 { + return nil, fmt.Errorf("bad format: %s", val) + } + if !strings.HasPrefix(split[0], "/dev/") { + return nil, fmt.Errorf("bad format for device path: %s", val) + } + rate, err := strconv.ParseUint(split[1], 10, 64) + if err != nil { + return nil, fmt.Errorf("invalid rate for device: %s. The correct format is :. Number must be a positive integer", val) + } + if rate < 0 { + return nil, fmt.Errorf("invalid rate for device: %s. The correct format is :. Number must be a positive integer", val) + } + + return &blkiodev.ThrottleDevice{Path: split[0], Rate: rate}, nil +} + +// ThrottledeviceOpt defines a map of ThrottleDevices +type ThrottledeviceOpt struct { + values []*blkiodev.ThrottleDevice + validator ValidatorThrottleFctType +} + +// NewThrottledeviceOpt creates a new ThrottledeviceOpt +func NewThrottledeviceOpt(validator ValidatorThrottleFctType) ThrottledeviceOpt { + values := []*blkiodev.ThrottleDevice{} + return ThrottledeviceOpt{ + values: values, + validator: validator, + } +} + +// Set validates a ThrottleDevice and sets its name as a key in ThrottledeviceOpt +func (opt *ThrottledeviceOpt) Set(val string) error { + var value *blkiodev.ThrottleDevice + if opt.validator != nil { + v, err := opt.validator(val) + if err != nil { + return err + } + value = v + } + (opt.values) = append((opt.values), value) + return nil +} + +// String returns ThrottledeviceOpt values as a string. +func (opt *ThrottledeviceOpt) String() string { + var out []string + for _, v := range opt.values { + out = append(out, v.String()) + } + + return fmt.Sprintf("%v", out) +} + +// GetList returns a slice of pointers to ThrottleDevices. +func (opt *ThrottledeviceOpt) GetList() []*blkiodev.ThrottleDevice { + var throttledevice []*blkiodev.ThrottleDevice + throttledevice = append(throttledevice, opt.values...) + + return throttledevice +} + +// Type returns the option type +func (opt *ThrottledeviceOpt) Type() string { + return "list" +} diff --git a/vendor/github.com/docker/cli/opts/ulimit.go b/vendor/github.com/docker/cli/opts/ulimit.go new file mode 100644 index 000000000..5adfe3085 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/ulimit.go @@ -0,0 +1,57 @@ +package opts + +import ( + "fmt" + + "github.com/docker/go-units" +) + +// UlimitOpt defines a map of Ulimits +type UlimitOpt struct { + values *map[string]*units.Ulimit +} + +// NewUlimitOpt creates a new UlimitOpt +func NewUlimitOpt(ref *map[string]*units.Ulimit) *UlimitOpt { + if ref == nil { + ref = &map[string]*units.Ulimit{} + } + return &UlimitOpt{ref} +} + +// Set validates a Ulimit and sets its name as a key in UlimitOpt +func (o *UlimitOpt) Set(val string) error { + l, err := units.ParseUlimit(val) + if err != nil { + return err + } + + (*o.values)[l.Name] = l + + return nil +} + +// String returns Ulimit values as a string. +func (o *UlimitOpt) String() string { + var out []string + for _, v := range *o.values { + out = append(out, v.String()) + } + + return fmt.Sprintf("%v", out) +} + +// GetList returns a slice of pointers to Ulimits. +func (o *UlimitOpt) GetList() []*units.Ulimit { + var ulimits []*units.Ulimit + for _, v := range *o.values { + ulimits = append(ulimits, v) + } + + return ulimits +} + +// Type returns the option type +func (o *UlimitOpt) Type() string { + return "ulimit" +} diff --git a/vendor/github.com/docker/cli/opts/weightdevice.go b/vendor/github.com/docker/cli/opts/weightdevice.go new file mode 100644 index 000000000..46ce9b656 --- /dev/null +++ b/vendor/github.com/docker/cli/opts/weightdevice.go @@ -0,0 +1,84 @@ +package opts + +import ( + "fmt" + "strconv" + "strings" + + "github.com/docker/docker/api/types/blkiodev" +) + +// ValidatorWeightFctType defines a validator function that returns a validated struct and/or an error. +type ValidatorWeightFctType func(val string) (*blkiodev.WeightDevice, error) + +// ValidateWeightDevice validates that the specified string has a valid device-weight format. +func ValidateWeightDevice(val string) (*blkiodev.WeightDevice, error) { + split := strings.SplitN(val, ":", 2) + if len(split) != 2 { + return nil, fmt.Errorf("bad format: %s", val) + } + if !strings.HasPrefix(split[0], "/dev/") { + return nil, fmt.Errorf("bad format for device path: %s", val) + } + weight, err := strconv.ParseUint(split[1], 10, 0) + if err != nil { + return nil, fmt.Errorf("invalid weight for device: %s", val) + } + if weight > 0 && (weight < 10 || weight > 1000) { + return nil, fmt.Errorf("invalid weight for device: %s", val) + } + + return &blkiodev.WeightDevice{ + Path: split[0], + Weight: uint16(weight), + }, nil +} + +// WeightdeviceOpt defines a map of WeightDevices +type WeightdeviceOpt struct { + values []*blkiodev.WeightDevice + validator ValidatorWeightFctType +} + +// NewWeightdeviceOpt creates a new WeightdeviceOpt +func NewWeightdeviceOpt(validator ValidatorWeightFctType) WeightdeviceOpt { + values := []*blkiodev.WeightDevice{} + return WeightdeviceOpt{ + values: values, + validator: validator, + } +} + +// Set validates a WeightDevice and sets its name as a key in WeightdeviceOpt +func (opt *WeightdeviceOpt) Set(val string) error { + var value *blkiodev.WeightDevice + if opt.validator != nil { + v, err := opt.validator(val) + if err != nil { + return err + } + value = v + } + (opt.values) = append((opt.values), value) + return nil +} + +// String returns WeightdeviceOpt values as a string. +func (opt *WeightdeviceOpt) String() string { + var out []string + for _, v := range opt.values { + out = append(out, v.String()) + } + + return fmt.Sprintf("%v", out) +} + +// GetList returns a slice of pointers to WeightDevices. +func (opt *WeightdeviceOpt) GetList() []*blkiodev.WeightDevice { + return opt.values +} + +// Type returns the option type +func (opt *WeightdeviceOpt) Type() string { + return "list" +} diff --git a/vendor/github.com/docker/distribution/AUTHORS b/vendor/github.com/docker/distribution/AUTHORS new file mode 100644 index 000000000..252ff8aa2 --- /dev/null +++ b/vendor/github.com/docker/distribution/AUTHORS @@ -0,0 +1,182 @@ +a-palchikov +Aaron Lehmann +Aaron Schlesinger +Aaron Vinson +Adam Duke +Adam Enger +Adrian Mouat +Ahmet Alp Balkan +Alex Chan +Alex Elman +Alexey Gladkov +allencloud +amitshukla +Amy Lindburg +Andrew Hsu +Andrew Meredith +Andrew T Nguyen +Andrey Kostov +Andy Goldstein +Anis Elleuch +Anton Tiurin +Antonio Mercado +Antonio Murdaca +Anusha Ragunathan +Arien Holthuizen +Arnaud Porterie +Arthur Baars +Asuka Suzuki +Avi Miller +Ayose Cazorla +BadZen +Ben Bodenmiller +Ben Firshman +bin liu +Brian Bland +burnettk +Carson A +Cezar Sa Espinola +Charles Smith +Chris Dillon +cuiwei13 +cyli +Daisuke Fujita +Daniel Huhn +Darren Shepherd +Dave Trombley +Dave Tucker +David Lawrence +David Verhasselt +David Xia +davidli +Dejan Golja +Derek McGowan +Diogo Mónica +DJ Enriquez +Donald Huang +Doug Davis +Edgar Lee +Eric Yang +Fabio Berchtold +Fabio Huser +farmerworking +Felix Yan +Florentin Raud +Frank Chen +Frederick F. Kautz IV +gabriell nascimento +Gleb Schukin +harche +Henri Gomez +Hu Keping +Hua Wang +HuKeping +Ian Babrou +igayoso +Jack Griffin +James Findley +Jason Freidman +Jason Heiss +Jeff Nickoloff +Jess Frazelle +Jessie Frazelle +jhaohai +Jianqing Wang +Jihoon Chung +Joao Fernandes +John Mulhausen +John Starks +Jon Johnson +Jon Poler +Jonathan Boulle +Jordan Liggitt +Josh Chorlton +Josh Hawn +Julien Fernandez +Ke Xu +Keerthan Mala +Kelsey Hightower +Kenneth Lim +Kenny Leung +Li Yi +Liu Hua +liuchang0812 +Lloyd Ramey +Louis Kottmann +Luke Carpenter +Marcus Martins +Mary Anthony +Matt Bentley +Matt Duch +Matt Moore +Matt Robenolt +Matthew Green +Michael Prokop +Michal Minar +Michal Minář +Mike Brown +Miquel Sabaté +Misty Stanley-Jones +Misty Stanley-Jones +Morgan Bauer +moxiegirl +Nathan Sullivan +nevermosby +Nghia Tran +Nikita Tarasov +Noah Treuhaft +Nuutti Kotivuori +Oilbeater +Olivier Gambier +Olivier Jacques +Omer Cohen +Patrick Devine +Phil Estes +Philip Misiowiec +Pierre-Yves Ritschard +Qiao Anran +Randy Barlow +Richard Scothern +Rodolfo Carvalho +Rusty Conover +Sean Boran +Sebastiaan van Stijn +Sebastien Coavoux +Serge Dubrouski +Sharif Nassar +Shawn Falkner-Horine +Shreyas Karnik +Simon Thulbourn +spacexnice +Spencer Rinehart +Stan Hu +Stefan Majewsky +Stefan Weil +Stephen J Day +Sungho Moon +Sven Dowideit +Sylvain Baubeau +Ted Reed +tgic +Thomas Sjögren +Tianon Gravi +Tibor Vass +Tonis Tiigi +Tony Holdstock-Brown +Trevor Pounds +Troels Thomsen +Victor Vieux +Victoria Bialas +Vincent Batts +Vincent Demeester +Vincent Giersch +W. Trevor King +weiyuan.yl +xg.song +xiekeyang +Yann ROBERT +yaoyao.xyy +yuexiao-wang +yuzou +zhouhaibing089 +姜继忠 diff --git a/vendor/github.com/docker/distribution/BUILDING.md b/vendor/github.com/docker/distribution/BUILDING.md new file mode 100644 index 000000000..2981d016b --- /dev/null +++ b/vendor/github.com/docker/distribution/BUILDING.md @@ -0,0 +1,117 @@ + +# Building the registry source + +## Use-case + +This is useful if you intend to actively work on the registry. + +### Alternatives + +Most people should use the [official Registry docker image](https://hub.docker.com/r/library/registry/). + +People looking for advanced operational use cases might consider rolling their own image with a custom Dockerfile inheriting `FROM registry:2`. + +OS X users who want to run natively can do so following [the instructions here](https://github.com/docker/docker.github.io/blob/master/registry/recipes/osx-setup-guide.md). + +### Gotchas + +You are expected to know your way around with go & git. + +If you are a casual user with no development experience, and no preliminary knowledge of go, building from source is probably not a good solution for you. + +## Build the development environment + +The first prerequisite of properly building distribution targets is to have a Go +development environment setup. Please follow [How to Write Go Code](https://golang.org/doc/code.html) +for proper setup. If done correctly, you should have a GOROOT and GOPATH set in the +environment. + +If a Go development environment is setup, one can use `go get` to install the +`registry` command from the current latest: + + go get github.com/docker/distribution/cmd/registry + +The above will install the source repository into the `GOPATH`. + +Now create the directory for the registry data (this might require you to set permissions properly) + + mkdir -p /var/lib/registry + +... or alternatively `export REGISTRY_STORAGE_FILESYSTEM_ROOTDIRECTORY=/somewhere` if you want to store data into another location. + +The `registry` +binary can then be run with the following: + + $ $GOPATH/bin/registry --version + $GOPATH/bin/registry github.com/docker/distribution v2.0.0-alpha.1+unknown + +> __NOTE:__ While you do not need to use `go get` to checkout the distribution +> project, for these build instructions to work, the project must be checked +> out in the correct location in the `GOPATH`. This should almost always be +> `$GOPATH/src/github.com/docker/distribution`. + +The registry can be run with the default config using the following +incantation: + + $ $GOPATH/bin/registry serve $GOPATH/src/github.com/docker/distribution/cmd/registry/config-example.yml + INFO[0000] endpoint local-5003 disabled, skipping app.id=34bbec38-a91a-494a-9a3f-b72f9010081f version=v2.0.0-alpha.1+unknown + INFO[0000] endpoint local-8083 disabled, skipping app.id=34bbec38-a91a-494a-9a3f-b72f9010081f version=v2.0.0-alpha.1+unknown + INFO[0000] listening on :5000 app.id=34bbec38-a91a-494a-9a3f-b72f9010081f version=v2.0.0-alpha.1+unknown + INFO[0000] debug server listening localhost:5001 + +If it is working, one should see the above log messages. + +### Repeatable Builds + +For the full development experience, one should `cd` into +`$GOPATH/src/github.com/docker/distribution`. From there, the regular `go` +commands, such as `go test`, should work per package (please see +[Developing](#developing) if they don't work). + +A `Makefile` has been provided as a convenience to support repeatable builds. +Please install the following into `GOPATH` for it to work: + + go get github.com/golang/lint/golint + +Once these commands are available in the `GOPATH`, run `make` to get a full +build: + + $ make + + clean + + fmt + + vet + + lint + + build + github.com/docker/docker/vendor/src/code.google.com/p/go/src/pkg/archive/tar + github.com/sirupsen/logrus + github.com/docker/libtrust + ... + github.com/yvasiyarov/gorelic + github.com/docker/distribution/registry/handlers + github.com/docker/distribution/cmd/registry + + test + ... + ok github.com/docker/distribution/digest 7.875s + ok github.com/docker/distribution/manifest 0.028s + ok github.com/docker/distribution/notifications 17.322s + ? github.com/docker/distribution/registry [no test files] + ok github.com/docker/distribution/registry/api/v2 0.101s + ? github.com/docker/distribution/registry/auth [no test files] + ok github.com/docker/distribution/registry/auth/silly 0.011s + ... + + /Users/sday/go/src/github.com/docker/distribution/bin/registry + + /Users/sday/go/src/github.com/docker/distribution/bin/registry-api-descriptor-template + + binaries + +The above provides a repeatable build using the contents of the vendor +directory. This includes formatting, vetting, linting, building, +testing and generating tagged binaries. We can verify this worked by running +the registry binary generated in the "./bin" directory: + + $ ./bin/registry --version + ./bin/registry github.com/docker/distribution v2.0.0-alpha.2-80-g16d8b2c.m + +### Optional build tags + +Optional [build tags](http://golang.org/pkg/go/build/) can be provided using +the environment variable `DOCKER_BUILDTAGS`. diff --git a/vendor/github.com/docker/distribution/CHANGELOG.md b/vendor/github.com/docker/distribution/CHANGELOG.md new file mode 100644 index 000000000..9e46af37f --- /dev/null +++ b/vendor/github.com/docker/distribution/CHANGELOG.md @@ -0,0 +1,106 @@ +# Changelog + +## 2.6.0 (2017-01-18) + +#### Storage +- S3: fixed bug in delete due to read-after-write inconsistency +- S3: allow EC2 IAM roles to be used when authorizing region endpoints +- S3: add Object ACL Support +- S3: fix delete method's notion of subpaths +- S3: use multipart upload API in `Move` method for performance +- S3: add v2 signature signing for legacy S3 clones +- Swift: add simple heuristic to detect incomplete DLOs during read ops +- Swift: support different user and tenant domains +- Swift: bulk deletes in chunks +- Aliyun OSS: fix delete method's notion of subpaths +- Aliyun OSS: optimize data copy after upload finishes +- Azure: close leaking response body +- Fix storage drivers dropping non-EOF errors when listing repositories +- Compare path properly when listing repositories in catalog +- Add a foreign layer URL host whitelist +- Improve catalog enumerate runtime + +#### Registry +- Export `storage.CreateOptions` in top-level package +- Enable notifications to endpoints that use self-signed certificates +- Properly validate multi-URL foreign layers +- Add control over validation of URLs in pushed manifests +- Proxy mode: fix socket leak when pull is cancelled +- Tag service: properly handle error responses on HEAD request +- Support for custom authentication URL in proxying registry +- Add configuration option to disable access logging +- Add notification filtering by target media type +- Manifest: `References()` returns all children +- Honor `X-Forwarded-Port` and Forwarded headers +- Reference: Preserve tag and digest in With* functions +- Add policy configuration for enforcing repository classes + +#### Client +- Changes the client Tags `All()` method to follow links +- Allow registry clients to connect via HTTP2 +- Better handling of OAuth errors in client + +#### Spec +- Manifest: clarify relationship between urls and foreign layers +- Authorization: add support for repository classes + +#### Manifest +- Override media type returned from `Stat()` for existing manifests +- Add plugin mediatype to distribution manifest + +#### Docs +- Document `TOOMANYREQUESTS` error code +- Document required Let's Encrypt port +- Improve documentation around implementation of OAuth2 +- Improve documentation for configuration + +#### Auth +- Add support for registry type in scope +- Add support for using v2 ping challenges for v1 +- Add leeway to JWT `nbf` and `exp` checking +- htpasswd: dynamically parse htpasswd file +- Fix missing auth headers with PATCH HTTP request when pushing to default port + +#### Dockerfile +- Update to go1.7 +- Reorder Dockerfile steps for better layer caching + +#### Notes + +Documentation has moved to the documentation repository at +`github.com/docker/docker.github.io/tree/master/registry` + +The registry is go 1.7 compliant, and passes newer, more restrictive `lint` and `vet` ing. + + +## 2.5.0 (2016-06-14) + +#### Storage +- Ensure uploads directory is cleaned after upload is committed +- Add ability to cap concurrent operations in filesystem driver +- S3: Add 'us-gov-west-1' to the valid region list +- Swift: Handle ceph not returning Last-Modified header for HEAD requests +- Add redirect middleware + +#### Registry +- Add support for blobAccessController middleware +- Add support for layers from foreign sources +- Remove signature store +- Add support for Let's Encrypt +- Correct yaml key names in configuration + +#### Client +- Add option to get content digest from manifest get + +#### Spec +- Update the auth spec scope grammar to reflect the fact that hostnames are optionally supported +- Clarify API documentation around catalog fetch behavior + +#### API +- Support returning HTTP 429 (Too Many Requests) + +#### Documentation +- Update auth documentation examples to show "expires in" as int + +#### Docker Image +- Use Alpine Linux as base image diff --git a/vendor/github.com/docker/distribution/CONTRIBUTING.md b/vendor/github.com/docker/distribution/CONTRIBUTING.md new file mode 100644 index 000000000..4c067d9e7 --- /dev/null +++ b/vendor/github.com/docker/distribution/CONTRIBUTING.md @@ -0,0 +1,148 @@ +# Contributing to the registry + +## Before reporting an issue... + +### If your problem is with... + + - automated builds + - your account on the [Docker Hub](https://hub.docker.com/) + - any other [Docker Hub](https://hub.docker.com/) issue + +Then please do not report your issue here - you should instead report it to [https://support.docker.com](https://support.docker.com) + +### If you... + + - need help setting up your registry + - can't figure out something + - are not sure what's going on or what your problem is + +Then please do not open an issue here yet - you should first try one of the following support forums: + + - irc: #docker-distribution on freenode + - mailing-list: or https://groups.google.com/a/dockerproject.org/forum/#!forum/distribution + +### Reporting security issues + +The Docker maintainers take security seriously. If you discover a security +issue, please bring it to their attention right away! + +Please **DO NOT** file a public issue, instead send your report privately to +[security@docker.com](mailto:security@docker.com). + +## Reporting an issue properly + +By following these simple rules you will get better and faster feedback on your issue. + + - search the bugtracker for an already reported issue + +### If you found an issue that describes your problem: + + - please read other user comments first, and confirm this is the same issue: a given error condition might be indicative of different problems - you may also find a workaround in the comments + - please refrain from adding "same thing here" or "+1" comments + - you don't need to comment on an issue to get notified of updates: just hit the "subscribe" button + - comment if you have some new, technical and relevant information to add to the case + - __DO NOT__ comment on closed issues or merged PRs. If you think you have a related problem, open up a new issue and reference the PR or issue. + +### If you have not found an existing issue that describes your problem: + + 1. create a new issue, with a succinct title that describes your issue: + - bad title: "It doesn't work with my docker" + - good title: "Private registry push fail: 400 error with E_INVALID_DIGEST" + 2. copy the output of: + - `docker version` + - `docker info` + - `docker exec registry --version` + 3. copy the command line you used to launch your Registry + 4. restart your docker daemon in debug mode (add `-D` to the daemon launch arguments) + 5. reproduce your problem and get your docker daemon logs showing the error + 6. if relevant, copy your registry logs that show the error + 7. provide any relevant detail about your specific Registry configuration (e.g., storage backend used) + 8. indicate if you are using an enterprise proxy, Nginx, or anything else between you and your Registry + +## Contributing a patch for a known bug, or a small correction + +You should follow the basic GitHub workflow: + + 1. fork + 2. commit a change + 3. make sure the tests pass + 4. PR + +Additionally, you must [sign your commits](https://github.com/docker/docker/blob/master/CONTRIBUTING.md#sign-your-work). It's very simple: + + - configure your name with git: `git config user.name "Real Name" && git config user.email mail@example.com` + - sign your commits using `-s`: `git commit -s -m "My commit"` + +Some simple rules to ensure quick merge: + + - clearly point to the issue(s) you want to fix in your PR comment (e.g., `closes #12345`) + - prefer multiple (smaller) PRs addressing individual issues over a big one trying to address multiple issues at once + - if you need to amend your PR following comments, please squash instead of adding more commits + +## Contributing new features + +You are heavily encouraged to first discuss what you want to do. You can do so on the irc channel, or by opening an issue that clearly describes the use case you want to fulfill, or the problem you are trying to solve. + +If this is a major new feature, you should then submit a proposal that describes your technical solution and reasoning. +If you did discuss it first, this will likely be greenlighted very fast. It's advisable to address all feedback on this proposal before starting actual work. + +Then you should submit your implementation, clearly linking to the issue (and possible proposal). + +Your PR will be reviewed by the community, then ultimately by the project maintainers, before being merged. + +It's mandatory to: + + - interact respectfully with other community members and maintainers - more generally, you are expected to abide by the [Docker community rules](https://github.com/docker/docker/blob/master/CONTRIBUTING.md#docker-community-guidelines) + - address maintainers' comments and modify your submission accordingly + - write tests for any new code + +Complying to these simple rules will greatly accelerate the review process, and will ensure you have a pleasant experience in contributing code to the Registry. + +Have a look at a great, successful contribution: the [Swift driver PR](https://github.com/docker/distribution/pull/493) + +## Coding Style + +Unless explicitly stated, we follow all coding guidelines from the Go +community. While some of these standards may seem arbitrary, they somehow seem +to result in a solid, consistent codebase. + +It is possible that the code base does not currently comply with these +guidelines. We are not looking for a massive PR that fixes this, since that +goes against the spirit of the guidelines. All new contributions should make a +best effort to clean up and make the code base better than they left it. +Obviously, apply your best judgement. Remember, the goal here is to make the +code base easier for humans to navigate and understand. Always keep that in +mind when nudging others to comply. + +The rules: + +1. All code should be formatted with `gofmt -s`. +2. All code should pass the default levels of + [`golint`](https://github.com/golang/lint). +3. All code should follow the guidelines covered in [Effective + Go](http://golang.org/doc/effective_go.html) and [Go Code Review + Comments](https://github.com/golang/go/wiki/CodeReviewComments). +4. Comment the code. Tell us the why, the history and the context. +5. Document _all_ declarations and methods, even private ones. Declare + expectations, caveats and anything else that may be important. If a type + gets exported, having the comments already there will ensure it's ready. +6. Variable name length should be proportional to its context and no longer. + `noCommaALongVariableNameLikeThisIsNotMoreClearWhenASimpleCommentWouldDo`. + In practice, short methods will have short variable names and globals will + have longer names. +7. No underscores in package names. If you need a compound name, step back, + and re-examine why you need a compound name. If you still think you need a + compound name, lose the underscore. +8. No utils or helpers packages. If a function is not general enough to + warrant its own package, it has not been written generally enough to be a + part of a util package. Just leave it unexported and well-documented. +9. All tests should run with `go test` and outside tooling should not be + required. No, we don't need another unit testing framework. Assertion + packages are acceptable if they provide _real_ incremental value. +10. Even though we call these "rules" above, they are actually just + guidelines. Since you've read all the rules, you now know that. + +If you are having trouble getting into the mood of idiomatic Go, we recommend +reading through [Effective Go](http://golang.org/doc/effective_go.html). The +[Go Blog](http://blog.golang.org/) is also a great resource. Drinking the +kool-aid is a lot easier than going thirsty. diff --git a/vendor/github.com/docker/distribution/Dockerfile b/vendor/github.com/docker/distribution/Dockerfile new file mode 100644 index 000000000..e0d74684a --- /dev/null +++ b/vendor/github.com/docker/distribution/Dockerfile @@ -0,0 +1,21 @@ +FROM golang:1.10-alpine + +ENV DISTRIBUTION_DIR /go/src/github.com/docker/distribution +ENV DOCKER_BUILDTAGS include_oss include_gcs + +ARG GOOS=linux +ARG GOARCH=amd64 + +RUN set -ex \ + && apk add --no-cache make git + +WORKDIR $DISTRIBUTION_DIR +COPY . $DISTRIBUTION_DIR +COPY cmd/registry/config-dev.yml /etc/docker/registry/config.yml + +RUN make PREFIX=/go clean binaries + +VOLUME ["/var/lib/registry"] +EXPOSE 5000 +ENTRYPOINT ["registry"] +CMD ["serve", "/etc/docker/registry/config.yml"] diff --git a/vendor/github.com/docker/distribution/MAINTAINERS b/vendor/github.com/docker/distribution/MAINTAINERS new file mode 100644 index 000000000..3183620c5 --- /dev/null +++ b/vendor/github.com/docker/distribution/MAINTAINERS @@ -0,0 +1,243 @@ +# Distribution maintainers file +# +# This file describes who runs the docker/distribution project and how. +# This is a living document - if you see something out of date or missing, speak up! +# +# It is structured to be consumable by both humans and programs. +# To extract its contents programmatically, use any TOML-compliant parser. +# + +[Rules] + + [Rules.maintainers] + + title = "What is a maintainer?" + + text = """ +There are different types of maintainers, with different responsibilities, but +all maintainers have 3 things in common: + +1) They share responsibility in the project's success. +2) They have made a long-term, recurring time investment to improve the project. +3) They spend that time doing whatever needs to be done, not necessarily what +is the most interesting or fun. + +Maintainers are often under-appreciated, because their work is harder to appreciate. +It's easy to appreciate a really cool and technically advanced feature. It's harder +to appreciate the absence of bugs, the slow but steady improvement in stability, +or the reliability of a release process. But those things distinguish a good +project from a great one. +""" + + [Rules.reviewer] + + title = "What is a reviewer?" + + text = """ +A reviewer is a core role within the project. +They share in reviewing issues and pull requests and their LGTM count towards the +required LGTM count to merge a code change into the project. + +Reviewers are part of the organization but do not have write access. +Becoming a reviewer is a core aspect in the journey to becoming a maintainer. +""" + + [Rules.adding-maintainers] + + title = "How are maintainers added?" + + text = """ +Maintainers are first and foremost contributors that have shown they are +committed to the long term success of a project. Contributors wanting to become +maintainers are expected to be deeply involved in contributing code, pull +request review, and triage of issues in the project for more than three months. + +Just contributing does not make you a maintainer, it is about building trust +with the current maintainers of the project and being a person that they can +depend on and trust to make decisions in the best interest of the project. + +Periodically, the existing maintainers curate a list of contributors that have +shown regular activity on the project over the prior months. From this list, +maintainer candidates are selected and proposed on the maintainers mailing list. + +After a candidate has been announced on the maintainers mailing list, the +existing maintainers are given five business days to discuss the candidate, +raise objections and cast their vote. Candidates must be approved by at least 66% of the current maintainers by adding their vote on the mailing +list. Only maintainers of the repository that the candidate is proposed for are +allowed to vote. + +If a candidate is approved, a maintainer will contact the candidate to invite +the candidate to open a pull request that adds the contributor to the +MAINTAINERS file. The candidate becomes a maintainer once the pull request is +merged. +""" + + [Rules.stepping-down-policy] + + title = "Stepping down policy" + + text = """ +Life priorities, interests, and passions can change. If you're a maintainer but +feel you must remove yourself from the list, inform other maintainers that you +intend to step down, and if possible, help find someone to pick up your work. +At the very least, ensure your work can be continued where you left off. + +After you've informed other maintainers, create a pull request to remove +yourself from the MAINTAINERS file. +""" + + [Rules.inactive-maintainers] + + title = "Removal of inactive maintainers" + + text = """ +Similar to the procedure for adding new maintainers, existing maintainers can +be removed from the list if they do not show significant activity on the +project. Periodically, the maintainers review the list of maintainers and their +activity over the last three months. + +If a maintainer has shown insufficient activity over this period, a neutral +person will contact the maintainer to ask if they want to continue being +a maintainer. If the maintainer decides to step down as a maintainer, they +open a pull request to be removed from the MAINTAINERS file. + +If the maintainer wants to remain a maintainer, but is unable to perform the +required duties they can be removed with a vote of at least 66% of +the current maintainers. An e-mail is sent to the +mailing list, inviting maintainers of the project to vote. The voting period is +five business days. Issues related to a maintainer's performance should be +discussed with them among the other maintainers so that they are not surprised +by a pull request removing them. +""" + + [Rules.decisions] + + title = "How are decisions made?" + + text = """ +Short answer: EVERYTHING IS A PULL REQUEST. + +distribution is an open-source project with an open design philosophy. This means +that the repository is the source of truth for EVERY aspect of the project, +including its philosophy, design, road map, and APIs. *If it's part of the +project, it's in the repo. If it's in the repo, it's part of the project.* + +As a result, all decisions can be expressed as changes to the repository. An +implementation change is a change to the source code. An API change is a change +to the API specification. A philosophy change is a change to the philosophy +manifesto, and so on. + +All decisions affecting distribution, big and small, follow the same 3 steps: + +* Step 1: Open a pull request. Anyone can do this. + +* Step 2: Discuss the pull request. Anyone can do this. + +* Step 3: Merge or refuse the pull request. Who does this depends on the nature +of the pull request and which areas of the project it affects. +""" + + [Rules.DCO] + + title = "Helping contributors with the DCO" + + text = """ +The [DCO or `Sign your work`]( +https://github.com/moby/moby/blob/master/CONTRIBUTING.md#sign-your-work) +requirement is not intended as a roadblock or speed bump. + +Some distribution contributors are not as familiar with `git`, or have used a web +based editor, and thus asking them to `git commit --amend -s` is not the best +way forward. + +In this case, maintainers can update the commits based on clause (c) of the DCO. +The most trivial way for a contributor to allow the maintainer to do this, is to +add a DCO signature in a pull requests's comment, or a maintainer can simply +note that the change is sufficiently trivial that it does not substantially +change the existing contribution - i.e., a spelling change. + +When you add someone's DCO, please also add your own to keep a log. +""" + + [Rules."no direct push"] + + title = "I'm a maintainer. Should I make pull requests too?" + + text = """ +Yes. Nobody should ever push to master directly. All changes should be +made through a pull request. +""" + + [Rules.tsc] + + title = "Conflict Resolution and technical disputes" + + text = """ +distribution defers to the [Technical Steering Committee](https://github.com/moby/tsc) for escalations and resolution on disputes for technical matters." + """ + + [Rules.meta] + + title = "How is this process changed?" + + text = "Just like everything else: by making a pull request :)" + +# Current project organization +[Org] + + [Org.Maintainers] + people = [ + "dmcgowan", + "dmp42", + "stevvooe", + ] + [Org.Reviewers] + people = [ + "manishtomar", + "caervs", + "davidswu", + "RobbKistler" + ] + +[people] + +# A reference list of all people associated with the project. +# All other sections should refer to people by their canonical key +# in the people section. + + # ADD YOURSELF HERE IN ALPHABETICAL ORDER + + [people.caervs] + Name = "Ryan Abrams" + Email = "rdabrams@gmail.com" + GitHub = "caervs" + + [people.davidswu] + Name = "David Wu" + Email = "dwu7401@gmail.com" + GitHub = "davidswu" + + [people.dmcgowan] + Name = "Derek McGowan" + Email = "derek@mcgstyle.net" + GitHub = "dmcgowan" + + [people.dmp42] + Name = "Olivier Gambier" + Email = "olivier@docker.com" + GitHub = "dmp42" + + [people.manishtomar] + Name = "Manish Tomar" + Email = "manish.tomar@docker.com" + GitHub = "manishtomar" + + [people.RobbKistler] + Name = "Robb Kistler" + Email = "robb.kistler@docker.com" + GitHub = "RobbKistler" + + [people.stevvooe] + Name = "Stephen Day" + Email = "stephen.day@docker.com" + GitHub = "stevvooe" diff --git a/vendor/github.com/docker/distribution/Makefile b/vendor/github.com/docker/distribution/Makefile new file mode 100644 index 000000000..df89bb0c1 --- /dev/null +++ b/vendor/github.com/docker/distribution/Makefile @@ -0,0 +1,105 @@ +# Root directory of the project (absolute path). +ROOTDIR=$(dir $(abspath $(lastword $(MAKEFILE_LIST)))) + +# Used to populate version variable in main package. +VERSION=$(shell git describe --match 'v[0-9]*' --dirty='.m' --always) +REVISION=$(shell git rev-parse HEAD)$(shell if ! git diff --no-ext-diff --quiet --exit-code; then echo .m; fi) + + +PKG=github.com/docker/distribution + +# Project packages. +PACKAGES=$(shell go list -tags "${BUILDTAGS}" ./... | grep -v /vendor/) +INTEGRATION_PACKAGE=${PKG} +COVERAGE_PACKAGES=$(filter-out ${PKG}/registry/storage/driver/%,${PACKAGES}) + + +# Project binaries. +COMMANDS=registry digest registry-api-descriptor-template + +# Allow turning off function inlining and variable registerization +ifeq (${DISABLE_OPTIMIZATION},true) + GO_GCFLAGS=-gcflags "-N -l" + VERSION:="$(VERSION)-noopt" +endif + +WHALE = "+" + +# Go files +# +TESTFLAGS_RACE= +GOFILES=$(shell find . -type f -name '*.go') +GO_TAGS=$(if $(BUILDTAGS),-tags "$(BUILDTAGS)",) +GO_LDFLAGS=-ldflags '-s -w -X $(PKG)/version.Version=$(VERSION) -X $(PKG)/version.Revision=$(REVISION) -X $(PKG)/version.Package=$(PKG) $(EXTRA_LDFLAGS)' + +BINARIES=$(addprefix bin/,$(COMMANDS)) + +# Flags passed to `go test` +TESTFLAGS ?= -v $(TESTFLAGS_RACE) +TESTFLAGS_PARALLEL ?= 8 + +.PHONY: all build binaries check clean test test-race test-full integration coverage +.DEFAULT: all + +all: binaries + +AUTHORS: .mailmap .git/HEAD + git log --format='%aN <%aE>' | sort -fu > $@ + +# This only needs to be generated by hand when cutting full releases. +version/version.go: + @echo "$(WHALE) $@" + ./version/version.sh > $@ + +check: ## run all linters (TODO: enable "unused", "varcheck", "ineffassign", "unconvert", "staticheck", "goimports", "structcheck") + @echo "$(WHALE) $@" + gometalinter --config .gometalinter.json ./... + +test: ## run tests, except integration test with test.short + @echo "$(WHALE) $@" + @go test ${GO_TAGS} -test.short ${TESTFLAGS} $(filter-out ${INTEGRATION_PACKAGE},${PACKAGES}) + +test-race: ## run tests, except integration test with test.short and race + @echo "$(WHALE) $@" + @go test ${GO_TAGS} -race -test.short ${TESTFLAGS} $(filter-out ${INTEGRATION_PACKAGE},${PACKAGES}) + +test-full: ## run tests, except integration tests + @echo "$(WHALE) $@" + @go test ${GO_TAGS} ${TESTFLAGS} $(filter-out ${INTEGRATION_PACKAGE},${PACKAGES}) + +integration: ## run integration tests + @echo "$(WHALE) $@" + @go test ${TESTFLAGS} -parallel ${TESTFLAGS_PARALLEL} ${INTEGRATION_PACKAGE} + +coverage: ## generate coverprofiles from the unit tests + @echo "$(WHALE) $@" + @rm -f coverage.txt + @go test ${GO_TAGS} -i ${TESTFLAGS} $(filter-out ${INTEGRATION_PACKAGE},${COVERAGE_PACKAGES}) 2> /dev/null + @( for pkg in $(filter-out ${INTEGRATION_PACKAGE},${COVERAGE_PACKAGES}); do \ + go test ${GO_TAGS} ${TESTFLAGS} \ + -cover \ + -coverprofile=profile.out \ + -covermode=atomic $$pkg || exit; \ + if [ -f profile.out ]; then \ + cat profile.out >> coverage.txt; \ + rm profile.out; \ + fi; \ + done ) + +FORCE: + +# Build a binary from a cmd. +bin/%: cmd/% FORCE + @echo "$(WHALE) $@${BINARY_SUFFIX}" + @go build ${GO_GCFLAGS} ${GO_BUILD_FLAGS} -o $@${BINARY_SUFFIX} ${GO_LDFLAGS} ${GO_TAGS} ./$< + +binaries: $(BINARIES) ## build binaries + @echo "$(WHALE) $@" + +build: + @echo "$(WHALE) $@" + @go build ${GO_GCFLAGS} ${GO_BUILD_FLAGS} ${GO_LDFLAGS} ${GO_TAGS} $(PACKAGES) + +clean: ## clean up binaries + @echo "$(WHALE) $@" + @rm -f $(BINARIES) diff --git a/vendor/github.com/docker/distribution/README.md b/vendor/github.com/docker/distribution/README.md new file mode 100644 index 000000000..998878850 --- /dev/null +++ b/vendor/github.com/docker/distribution/README.md @@ -0,0 +1,130 @@ +# Distribution + +The Docker toolset to pack, ship, store, and deliver content. + +This repository's main product is the Docker Registry 2.0 implementation +for storing and distributing Docker images. It supersedes the +[docker/docker-registry](https://github.com/docker/docker-registry) +project with a new API design, focused around security and performance. + + + +[![Circle CI](https://circleci.com/gh/docker/distribution/tree/master.svg?style=svg)](https://circleci.com/gh/docker/distribution/tree/master) +[![GoDoc](https://godoc.org/github.com/docker/distribution?status.svg)](https://godoc.org/github.com/docker/distribution) + +This repository contains the following components: + +|**Component** |Description | +|--------------------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| +| **registry** | An implementation of the [Docker Registry HTTP API V2](docs/spec/api.md) for use with docker 1.6+. | +| **libraries** | A rich set of libraries for interacting with distribution components. Please see [godoc](https://godoc.org/github.com/docker/distribution) for details. **Note**: These libraries are **unstable**. | +| **specifications** | _Distribution_ related specifications are available in [docs/spec](docs/spec) | +| **documentation** | Docker's full documentation set is available at [docs.docker.com](https://docs.docker.com). This repository [contains the subset](docs/) related just to the registry. | + +### How does this integrate with Docker engine? + +This project should provide an implementation to a V2 API for use in the [Docker +core project](https://github.com/docker/docker). The API should be embeddable +and simplify the process of securely pulling and pushing content from `docker` +daemons. + +### What are the long term goals of the Distribution project? + +The _Distribution_ project has the further long term goal of providing a +secure tool chain for distributing content. The specifications, APIs and tools +should be as useful with Docker as they are without. + +Our goal is to design a professional grade and extensible content distribution +system that allow users to: + +* Enjoy an efficient, secured and reliable way to store, manage, package and + exchange content +* Hack/roll their own on top of healthy open-source components +* Implement their own home made solution through good specs, and solid + extensions mechanism. + +## More about Registry 2.0 + +The new registry implementation provides the following benefits: + +- faster push and pull +- new, more efficient implementation +- simplified deployment +- pluggable storage backend +- webhook notifications + +For information on upcoming functionality, please see [ROADMAP.md](ROADMAP.md). + +### Who needs to deploy a registry? + +By default, Docker users pull images from Docker's public registry instance. +[Installing Docker](https://docs.docker.com/engine/installation/) gives users this +ability. Users can also push images to a repository on Docker's public registry, +if they have a [Docker Hub](https://hub.docker.com/) account. + +For some users and even companies, this default behavior is sufficient. For +others, it is not. + +For example, users with their own software products may want to maintain a +registry for private, company images. Also, you may wish to deploy your own +image repository for images used to test or in continuous integration. For these +use cases and others, [deploying your own registry instance](https://github.com/docker/docker.github.io/blob/master/registry/deploying.md) +may be the better choice. + +### Migration to Registry 2.0 + +For those who have previously deployed their own registry based on the Registry +1.0 implementation and wish to deploy a Registry 2.0 while retaining images, +data migration is required. A tool to assist with migration efforts has been +created. For more information see [docker/migrator](https://github.com/docker/migrator). + +## Contribute + +Please see [CONTRIBUTING.md](CONTRIBUTING.md) for details on how to contribute +issues, fixes, and patches to this project. If you are contributing code, see +the instructions for [building a development environment](BUILDING.md). + +## Support + +If any issues are encountered while using the _Distribution_ project, several +avenues are available for support: + + + + + + + + + + + + + + + + + + +
+ IRC + + #docker-distribution on FreeNode +
+ Issue Tracker + + github.com/docker/distribution/issues +
+ Google Groups + + https://groups.google.com/a/dockerproject.org/forum/#!forum/distribution +
+ Mailing List + + docker@dockerproject.org +
+ + +## License + +This project is distributed under [Apache License, Version 2.0](LICENSE). diff --git a/vendor/github.com/docker/distribution/RELEASE-CHECKLIST.md b/vendor/github.com/docker/distribution/RELEASE-CHECKLIST.md new file mode 100644 index 000000000..e959c2a41 --- /dev/null +++ b/vendor/github.com/docker/distribution/RELEASE-CHECKLIST.md @@ -0,0 +1,43 @@ +## Registry Release Checklist + +10. Compile release notes detailing features and since the last release. + + Update the `CHANGELOG.md` file and create a PR to master with the updates. +Once that PR has been approved by maintainers the change may be cherry-picked +to the release branch (new release branches may be forked from this commit). + +20. Update the version file: `https://github.com/docker/distribution/blob/master/version/version.go` + +30. Update the `MAINTAINERS` (if necessary), `AUTHORS` and `.mailmap` files. + +``` +make AUTHORS +``` + +40. Create a signed tag. + + Distribution uses semantic versioning. Tags are of the format +`vx.y.z[-rcn]`. You will need PGP installed and a PGP key which has been added +to your Github account. The comment for the tag should include the release +notes, use previous tags as a guide for formatting consistently. Run +`git tag -s vx.y.z[-rcn]` to create tag and `git -v vx.y.z[-rcn]` to verify tag, +check comment and correct commit hash. + +50. Push the signed tag + +60. Create a new [release](https://github.com/docker/distribution/releases). In the case of a release candidate, tick the `pre-release` checkbox. + +70. Update the registry binary in [distribution library image repo](https://github.com/docker/distribution-library-image) by running the update script and opening a pull request. + +80. Update the official image. Add the new version in the [official images repo](https://github.com/docker-library/official-images) by appending a new version to the `registry/registry` file with the git hash pointed to by the signed tag. Update the major version to point to the latest version and the minor version to point to new patch release if necessary. +e.g. to release `2.3.1` + + `2.3.1 (new)` + + `2.3.0 -> 2.3.0` can be removed + + `2 -> 2.3.1` + + `2.3 -> 2.3.1` + +90. Build a new distribution/registry image on [Docker hub](https://hub.docker.com/u/distribution/dashboard) by adding a new automated build with the new tag and re-building the images. diff --git a/vendor/github.com/docker/distribution/ROADMAP.md b/vendor/github.com/docker/distribution/ROADMAP.md new file mode 100644 index 000000000..aa3bd3bff --- /dev/null +++ b/vendor/github.com/docker/distribution/ROADMAP.md @@ -0,0 +1,266 @@ +# Roadmap + +The Distribution Project consists of several components, some of which are +still being defined. This document defines the high-level goals of the +project, identifies the current components, and defines the release- +relationship to the Docker Platform. + +* [Distribution Goals](#distribution-goals) +* [Distribution Components](#distribution-components) +* [Project Planning](#project-planning): release-relationship to the Docker Platform. + +This road map is a living document, providing an overview of the goals and +considerations made in respect of the future of the project. + +## Distribution Goals + +- Replace the existing [docker registry](github.com/docker/docker-registry) + implementation as the primary implementation. +- Replace the existing push and pull code in the docker engine with the + distribution package. +- Define a strong data model for distributing docker images +- Provide a flexible distribution tool kit for use in the docker platform +- Unlock new distribution models + +## Distribution Components + +Components of the Distribution Project are managed via github [milestones](https://github.com/docker/distribution/milestones). Upcoming +features and bugfixes for a component will be added to the relevant milestone. If a feature or +bugfix is not part of a milestone, it is currently unscheduled for +implementation. + +* [Registry](#registry) +* [Distribution Package](#distribution-package) + +*** + +### Registry + +The new Docker registry is the main portion of the distribution repository. +Registry 2.0 is the first release of the next-generation registry. This was +primarily focused on implementing the [new registry +API](https://github.com/docker/distribution/blob/master/docs/spec/api.md), +with a focus on security and performance. + +Following from the Distribution project goals above, we have a set of goals +for registry v2 that we would like to follow in the design. New features +should be compared against these goals. + +#### Data Storage and Distribution First + +The registry's first goal is to provide a reliable, consistent storage +location for Docker images. The registry should only provide the minimal +amount of indexing required to fetch image data and no more. + +This means we should be selective in new features and API additions, including +those that may require expensive, ever growing indexes. Requests should be +servable in "constant time". + +#### Content Addressability + +All data objects used in the registry API should be content addressable. +Content identifiers should be secure and verifiable. This provides a secure, +reliable base from which to build more advanced content distribution systems. + +#### Content Agnostic + +In the past, changes to the image format would require large changes in Docker +and the Registry. By decoupling the distribution and image format, we can +allow the formats to progress without having to coordinate between the two. +This means that we should be focused on decoupling Docker from the registry +just as much as decoupling the registry from Docker. Such an approach will +allow us to unlock new distribution models that haven't been possible before. + +We can take this further by saying that the new registry should be content +agnostic. The registry provides a model of names, tags, manifests and content +addresses and that model can be used to work with content. + +#### Simplicity + +The new registry should be closer to a microservice component than its +predecessor. This means it should have a narrower API and a low number of +service dependencies. It should be easy to deploy. + +This means that other solutions should be explored before changing the API or +adding extra dependencies. If functionality is required, can it be added as an +extension or companion service. + +#### Extensibility + +The registry should provide extension points to add functionality. By keeping +the scope narrow, but providing the ability to add functionality. + +Features like search, indexing, synchronization and registry explorers fall +into this category. No such feature should be added unless we've found it +impossible to do through an extension. + +#### Active Feature Discussions + +The following are feature discussions that are currently active. + +If you don't see your favorite, unimplemented feature, feel free to contact us +via IRC or the mailing list and we can talk about adding it. The goal here is +to make sure that new features go through a rigid design process before +landing in the registry. + +##### Proxying to other Registries + +A _pull-through caching_ mode exists for the registry, but is restricted from +within the docker client to only mirror the official Docker Hub. This functionality +can be expanded when image provenance has been specified and implemented in the +distribution project. + +##### Metadata storage + +Metadata for the registry is currently stored with the manifest and layer data on +the storage backend. While this is a big win for simplicity and reliably maintaining +state, it comes with the cost of consistency and high latency. The mutable registry +metadata operations should be abstracted behind an API which will allow ACID compliant +storage systems to handle metadata. + +##### Peer to Peer transfer + +Discussion has started here: https://docs.google.com/document/d/1rYDpSpJiQWmCQy8Cuiaa3NH-Co33oK_SC9HeXYo87QA/edit + +##### Indexing, Search and Discovery + +The original registry provided some implementation of search for use with +private registries. Support has been elided from V2 since we'd like to both +decouple search functionality from the registry. The makes the registry +simpler to deploy, especially in use cases where search is not needed, and +let's us decouple the image format from the registry. + +There are explorations into using the catalog API and notification system to +build external indexes. The current line of thought is that we will define a +common search API to index and query docker images. Such a system could be run +as a companion to a registry or set of registries to power discovery. + +The main issue with search and discovery is that there are so many ways to +accomplish it. There are two aspects to this project. The first is deciding on +how it will be done, including an API definition that can work with changing +data formats. The second is the process of integrating with `docker search`. +We expect that someone attempts to address the problem with the existing tools +and propose it as a standard search API or uses it to inform a standardization +process. Once this has been explored, we integrate with the docker client. + +Please see the following for more detail: + +- https://github.com/docker/distribution/issues/206 + +##### Deletes + +> __NOTE:__ Deletes are a much asked for feature. Before requesting this +feature or participating in discussion, we ask that you read this section in +full and understand the problems behind deletes. + +While, at first glance, implementing deleting seems simple, there are a number +mitigating factors that make many solutions not ideal or even pathological in +the context of a registry. The following paragraph discuss the background and +approaches that could be applied to arrive at a solution. + +The goal of deletes in any system is to remove unused or unneeded data. Only +data requested for deletion should be removed and no other data. Removing +unintended data is worse than _not_ removing data that was requested for +removal but ideally, both are supported. Generally, according to this rule, we +err on holding data longer than needed, ensuring that it is only removed when +we can be certain that it can be removed. With the current behavior, we opt to +hold onto the data forever, ensuring that data cannot be incorrectly removed. + +To understand the problems with implementing deletes, one must understand the +data model. All registry data is stored in a filesystem layout, implemented on +a "storage driver", effectively a _virtual file system_ (VFS). The storage +system must assume that this VFS layer will be eventually consistent and has +poor read- after-write consistency, since this is the lower common denominator +among the storage drivers. This is mitigated by writing values in reverse- +dependent order, but makes wider transactional operations unsafe. + +Layered on the VFS model is a content-addressable _directed, acyclic graph_ +(DAG) made up of blobs. Manifests reference layers. Tags reference manifests. +Since the same data can be referenced by multiple manifests, we only store +data once, even if it is in different repositories. Thus, we have a set of +blobs, referenced by tags and manifests. If we want to delete a blob we need +to be certain that it is no longer referenced by another manifest or tag. When +we delete a manifest, we also can try to delete the referenced blobs. Deciding +whether or not a blob has an active reference is the crux of the problem. + +Conceptually, deleting a manifest and its resources is quite simple. Just find +all the manifests, enumerate the referenced blobs and delete the blobs not in +that set. An astute observer will recognize this as a garbage collection +problem. As with garbage collection in programming languages, this is very +simple when one always has a consistent view. When one adds parallelism and an +inconsistent view of data, it becomes very challenging. + +A simple example can demonstrate this. Let's say we are deleting a manifest +_A_ in one process. We scan the manifest and decide that all the blobs are +ready for deletion. Concurrently, we have another process accepting a new +manifest _B_ referencing one or more blobs from the manifest _A_. Manifest _B_ +is accepted and all the blobs are considered present, so the operation +proceeds. The original process then deletes the referenced blobs, assuming +they were unreferenced. The manifest _B_, which we thought had all of its data +present, can no longer be served by the registry, since the dependent data has +been deleted. + +Deleting data from the registry safely requires some way to coordinate this +operation. The following approaches are being considered: + +- _Reference Counting_ - Maintain a count of references to each blob. This is + challenging for a number of reasons: 1. maintaining a consistent consensus + of reference counts across a set of Registries and 2. Building the initial + list of reference counts for an existing registry. These challenges can be + met with a consensus protocol like Paxos or Raft in the first case and a + necessary but simple scan in the second.. +- _Lock the World GC_ - Halt all writes to the data store. Walk the data store + and find all blob references. Delete all unreferenced blobs. This approach + is very simple but requires disabling writes for a period of time while the + service reads all data. This is slow and expensive but very accurate and + effective. +- _Generational GC_ - Do something similar to above but instead of blocking + writes, writes are sent to another storage backend while reads are broadcast + to the new and old backends. GC is then performed on the read-only portion. + Because writes land in the new backend, the data in the read-only section + can be safely deleted. The main drawbacks of this approach are complexity + and coordination. +- _Centralized Oracle_ - Using a centralized, transactional database, we can + know exactly which data is referenced at any given time. This avoids + coordination problem by managing this data in a single location. We trade + off metadata scalability for simplicity and performance. This is a very good + option for most registry deployments. This would create a bottleneck for + registry metadata. However, metadata is generally not the main bottleneck + when serving images. + +Please let us know if other solutions exist that we have yet to enumerate. +Note that for any approach, implementation is a massive consideration. For +example, a mark-sweep based solution may seem simple but the amount of work in +coordination offset the extra work it might take to build a _Centralized +Oracle_. We'll accept proposals for any solution but please coordinate with us +before dropping code. + +At this time, we have traded off simplicity and ease of deployment for disk +space. Simplicity and ease of deployment tend to reduce developer involvement, +which is currently the most expensive resource in software engineering. Taking +on any solution for deletes will greatly effect these factors, trading off +very cheap disk space for a complex deployment and operational story. + +Please see the following issues for more detail: + +- https://github.com/docker/distribution/issues/422 +- https://github.com/docker/distribution/issues/461 +- https://github.com/docker/distribution/issues/462 + +### Distribution Package + +At its core, the Distribution Project is a set of Go packages that make up +Distribution Components. At this time, most of these packages make up the +Registry implementation. + +The package itself is considered unstable. If you're using it, please take care to vendor the dependent version. + +For feature additions, please see the Registry section. In the future, we may break out a +separate Roadmap for distribution-specific features that apply to more than +just the registry. + +*** + +### Project Planning + +An [Open-Source Planning Process](https://github.com/docker/distribution/wiki/Open-Source-Planning-Process) is used to define the Roadmap. [Project Pages](https://github.com/docker/distribution/wiki) define the goals for each Milestone and identify current progress. diff --git a/vendor/github.com/docker/distribution/blobs.go b/vendor/github.com/docker/distribution/blobs.go new file mode 100644 index 000000000..c0e9261be --- /dev/null +++ b/vendor/github.com/docker/distribution/blobs.go @@ -0,0 +1,265 @@ +package distribution + +import ( + "context" + "errors" + "fmt" + "io" + "net/http" + "time" + + "github.com/docker/distribution/reference" + "github.com/opencontainers/go-digest" + "github.com/opencontainers/image-spec/specs-go/v1" +) + +var ( + // ErrBlobExists returned when blob already exists + ErrBlobExists = errors.New("blob exists") + + // ErrBlobDigestUnsupported when blob digest is an unsupported version. + ErrBlobDigestUnsupported = errors.New("unsupported blob digest") + + // ErrBlobUnknown when blob is not found. + ErrBlobUnknown = errors.New("unknown blob") + + // ErrBlobUploadUnknown returned when upload is not found. + ErrBlobUploadUnknown = errors.New("blob upload unknown") + + // ErrBlobInvalidLength returned when the blob has an expected length on + // commit, meaning mismatched with the descriptor or an invalid value. + ErrBlobInvalidLength = errors.New("blob invalid length") +) + +// ErrBlobInvalidDigest returned when digest check fails. +type ErrBlobInvalidDigest struct { + Digest digest.Digest + Reason error +} + +func (err ErrBlobInvalidDigest) Error() string { + return fmt.Sprintf("invalid digest for referenced layer: %v, %v", + err.Digest, err.Reason) +} + +// ErrBlobMounted returned when a blob is mounted from another repository +// instead of initiating an upload session. +type ErrBlobMounted struct { + From reference.Canonical + Descriptor Descriptor +} + +func (err ErrBlobMounted) Error() string { + return fmt.Sprintf("blob mounted from: %v to: %v", + err.From, err.Descriptor) +} + +// Descriptor describes targeted content. Used in conjunction with a blob +// store, a descriptor can be used to fetch, store and target any kind of +// blob. The struct also describes the wire protocol format. Fields should +// only be added but never changed. +type Descriptor struct { + // MediaType describe the type of the content. All text based formats are + // encoded as utf-8. + MediaType string `json:"mediaType,omitempty"` + + // Size in bytes of content. + Size int64 `json:"size,omitempty"` + + // Digest uniquely identifies the content. A byte stream can be verified + // against this digest. + Digest digest.Digest `json:"digest,omitempty"` + + // URLs contains the source URLs of this content. + URLs []string `json:"urls,omitempty"` + + // Annotations contains arbitrary metadata relating to the targeted content. + Annotations map[string]string `json:"annotations,omitempty"` + + // Platform describes the platform which the image in the manifest runs on. + // This should only be used when referring to a manifest. + Platform *v1.Platform `json:"platform,omitempty"` + + // NOTE: Before adding a field here, please ensure that all + // other options have been exhausted. Much of the type relationships + // depend on the simplicity of this type. +} + +// Descriptor returns the descriptor, to make it satisfy the Describable +// interface. Note that implementations of Describable are generally objects +// which can be described, not simply descriptors; this exception is in place +// to make it more convenient to pass actual descriptors to functions that +// expect Describable objects. +func (d Descriptor) Descriptor() Descriptor { + return d +} + +// BlobStatter makes blob descriptors available by digest. The service may +// provide a descriptor of a different digest if the provided digest is not +// canonical. +type BlobStatter interface { + // Stat provides metadata about a blob identified by the digest. If the + // blob is unknown to the describer, ErrBlobUnknown will be returned. + Stat(ctx context.Context, dgst digest.Digest) (Descriptor, error) +} + +// BlobDeleter enables deleting blobs from storage. +type BlobDeleter interface { + Delete(ctx context.Context, dgst digest.Digest) error +} + +// BlobEnumerator enables iterating over blobs from storage +type BlobEnumerator interface { + Enumerate(ctx context.Context, ingester func(dgst digest.Digest) error) error +} + +// BlobDescriptorService manages metadata about a blob by digest. Most +// implementations will not expose such an interface explicitly. Such mappings +// should be maintained by interacting with the BlobIngester. Hence, this is +// left off of BlobService and BlobStore. +type BlobDescriptorService interface { + BlobStatter + + // SetDescriptor assigns the descriptor to the digest. The provided digest and + // the digest in the descriptor must map to identical content but they may + // differ on their algorithm. The descriptor must have the canonical + // digest of the content and the digest algorithm must match the + // annotators canonical algorithm. + // + // Such a facility can be used to map blobs between digest domains, with + // the restriction that the algorithm of the descriptor must match the + // canonical algorithm (ie sha256) of the annotator. + SetDescriptor(ctx context.Context, dgst digest.Digest, desc Descriptor) error + + // Clear enables descriptors to be unlinked + Clear(ctx context.Context, dgst digest.Digest) error +} + +// BlobDescriptorServiceFactory creates middleware for BlobDescriptorService. +type BlobDescriptorServiceFactory interface { + BlobAccessController(svc BlobDescriptorService) BlobDescriptorService +} + +// ReadSeekCloser is the primary reader type for blob data, combining +// io.ReadSeeker with io.Closer. +type ReadSeekCloser interface { + io.ReadSeeker + io.Closer +} + +// BlobProvider describes operations for getting blob data. +type BlobProvider interface { + // Get returns the entire blob identified by digest along with the descriptor. + Get(ctx context.Context, dgst digest.Digest) ([]byte, error) + + // Open provides a ReadSeekCloser to the blob identified by the provided + // descriptor. If the blob is not known to the service, an error will be + // returned. + Open(ctx context.Context, dgst digest.Digest) (ReadSeekCloser, error) +} + +// BlobServer can serve blobs via http. +type BlobServer interface { + // ServeBlob attempts to serve the blob, identified by dgst, via http. The + // service may decide to redirect the client elsewhere or serve the data + // directly. + // + // This handler only issues successful responses, such as 2xx or 3xx, + // meaning it serves data or issues a redirect. If the blob is not + // available, an error will be returned and the caller may still issue a + // response. + // + // The implementation may serve the same blob from a different digest + // domain. The appropriate headers will be set for the blob, unless they + // have already been set by the caller. + ServeBlob(ctx context.Context, w http.ResponseWriter, r *http.Request, dgst digest.Digest) error +} + +// BlobIngester ingests blob data. +type BlobIngester interface { + // Put inserts the content p into the blob service, returning a descriptor + // or an error. + Put(ctx context.Context, mediaType string, p []byte) (Descriptor, error) + + // Create allocates a new blob writer to add a blob to this service. The + // returned handle can be written to and later resumed using an opaque + // identifier. With this approach, one can Close and Resume a BlobWriter + // multiple times until the BlobWriter is committed or cancelled. + Create(ctx context.Context, options ...BlobCreateOption) (BlobWriter, error) + + // Resume attempts to resume a write to a blob, identified by an id. + Resume(ctx context.Context, id string) (BlobWriter, error) +} + +// BlobCreateOption is a general extensible function argument for blob creation +// methods. A BlobIngester may choose to honor any or none of the given +// BlobCreateOptions, which can be specific to the implementation of the +// BlobIngester receiving them. +// TODO (brianbland): unify this with ManifestServiceOption in the future +type BlobCreateOption interface { + Apply(interface{}) error +} + +// CreateOptions is a collection of blob creation modifiers relevant to general +// blob storage intended to be configured by the BlobCreateOption.Apply method. +type CreateOptions struct { + Mount struct { + ShouldMount bool + From reference.Canonical + // Stat allows to pass precalculated descriptor to link and return. + // Blob access check will be skipped if set. + Stat *Descriptor + } +} + +// BlobWriter provides a handle for inserting data into a blob store. +// Instances should be obtained from BlobWriteService.Writer and +// BlobWriteService.Resume. If supported by the store, a writer can be +// recovered with the id. +type BlobWriter interface { + io.WriteCloser + io.ReaderFrom + + // Size returns the number of bytes written to this blob. + Size() int64 + + // ID returns the identifier for this writer. The ID can be used with the + // Blob service to later resume the write. + ID() string + + // StartedAt returns the time this blob write was started. + StartedAt() time.Time + + // Commit completes the blob writer process. The content is verified + // against the provided provisional descriptor, which may result in an + // error. Depending on the implementation, written data may be validated + // against the provisional descriptor fields. If MediaType is not present, + // the implementation may reject the commit or assign "application/octet- + // stream" to the blob. The returned descriptor may have a different + // digest depending on the blob store, referred to as the canonical + // descriptor. + Commit(ctx context.Context, provisional Descriptor) (canonical Descriptor, err error) + + // Cancel ends the blob write without storing any data and frees any + // associated resources. Any data written thus far will be lost. Cancel + // implementations should allow multiple calls even after a commit that + // result in a no-op. This allows use of Cancel in a defer statement, + // increasing the assurance that it is correctly called. + Cancel(ctx context.Context) error +} + +// BlobService combines the operations to access, read and write blobs. This +// can be used to describe remote blob services. +type BlobService interface { + BlobStatter + BlobProvider + BlobIngester +} + +// BlobStore represent the entire suite of blob related operations. Such an +// implementation can access, read, write, delete and serve blobs. +type BlobStore interface { + BlobService + BlobServer + BlobDeleter +} diff --git a/vendor/github.com/docker/distribution/doc.go b/vendor/github.com/docker/distribution/doc.go new file mode 100644 index 000000000..bdd8cb708 --- /dev/null +++ b/vendor/github.com/docker/distribution/doc.go @@ -0,0 +1,7 @@ +// Package distribution will define the interfaces for the components of +// docker distribution. The goal is to allow users to reliably package, ship +// and store content related to docker images. +// +// This is currently a work in progress. More details are available in the +// README.md. +package distribution diff --git a/vendor/github.com/docker/distribution/errors.go b/vendor/github.com/docker/distribution/errors.go new file mode 100644 index 000000000..8e0b788d6 --- /dev/null +++ b/vendor/github.com/docker/distribution/errors.go @@ -0,0 +1,119 @@ +package distribution + +import ( + "errors" + "fmt" + "strings" + + "github.com/opencontainers/go-digest" +) + +// ErrAccessDenied is returned when an access to a requested resource is +// denied. +var ErrAccessDenied = errors.New("access denied") + +// ErrManifestNotModified is returned when a conditional manifest GetByTag +// returns nil due to the client indicating it has the latest version +var ErrManifestNotModified = errors.New("manifest not modified") + +// ErrUnsupported is returned when an unimplemented or unsupported action is +// performed +var ErrUnsupported = errors.New("operation unsupported") + +// ErrSchemaV1Unsupported is returned when a client tries to upload a schema v1 +// manifest but the registry is configured to reject it +var ErrSchemaV1Unsupported = errors.New("manifest schema v1 unsupported") + +// ErrTagUnknown is returned if the given tag is not known by the tag service +type ErrTagUnknown struct { + Tag string +} + +func (err ErrTagUnknown) Error() string { + return fmt.Sprintf("unknown tag=%s", err.Tag) +} + +// ErrRepositoryUnknown is returned if the named repository is not known by +// the registry. +type ErrRepositoryUnknown struct { + Name string +} + +func (err ErrRepositoryUnknown) Error() string { + return fmt.Sprintf("unknown repository name=%s", err.Name) +} + +// ErrRepositoryNameInvalid should be used to denote an invalid repository +// name. Reason may set, indicating the cause of invalidity. +type ErrRepositoryNameInvalid struct { + Name string + Reason error +} + +func (err ErrRepositoryNameInvalid) Error() string { + return fmt.Sprintf("repository name %q invalid: %v", err.Name, err.Reason) +} + +// ErrManifestUnknown is returned if the manifest is not known by the +// registry. +type ErrManifestUnknown struct { + Name string + Tag string +} + +func (err ErrManifestUnknown) Error() string { + return fmt.Sprintf("unknown manifest name=%s tag=%s", err.Name, err.Tag) +} + +// ErrManifestUnknownRevision is returned when a manifest cannot be found by +// revision within a repository. +type ErrManifestUnknownRevision struct { + Name string + Revision digest.Digest +} + +func (err ErrManifestUnknownRevision) Error() string { + return fmt.Sprintf("unknown manifest name=%s revision=%s", err.Name, err.Revision) +} + +// ErrManifestUnverified is returned when the registry is unable to verify +// the manifest. +type ErrManifestUnverified struct{} + +func (ErrManifestUnverified) Error() string { + return "unverified manifest" +} + +// ErrManifestVerification provides a type to collect errors encountered +// during manifest verification. Currently, it accepts errors of all types, +// but it may be narrowed to those involving manifest verification. +type ErrManifestVerification []error + +func (errs ErrManifestVerification) Error() string { + var parts []string + for _, err := range errs { + parts = append(parts, err.Error()) + } + + return fmt.Sprintf("errors verifying manifest: %v", strings.Join(parts, ",")) +} + +// ErrManifestBlobUnknown returned when a referenced blob cannot be found. +type ErrManifestBlobUnknown struct { + Digest digest.Digest +} + +func (err ErrManifestBlobUnknown) Error() string { + return fmt.Sprintf("unknown blob %v on manifest", err.Digest) +} + +// ErrManifestNameInvalid should be used to denote an invalid manifest +// name. Reason may set, indicating the cause of invalidity. +type ErrManifestNameInvalid struct { + Name string + Reason error +} + +func (err ErrManifestNameInvalid) Error() string { + return fmt.Sprintf("manifest name %q invalid: %v", err.Name, err.Reason) +} diff --git a/vendor/github.com/docker/distribution/manifests.go b/vendor/github.com/docker/distribution/manifests.go new file mode 100644 index 000000000..1816baea1 --- /dev/null +++ b/vendor/github.com/docker/distribution/manifests.go @@ -0,0 +1,125 @@ +package distribution + +import ( + "context" + "fmt" + "mime" + + "github.com/opencontainers/go-digest" +) + +// Manifest represents a registry object specifying a set of +// references and an optional target +type Manifest interface { + // References returns a list of objects which make up this manifest. + // A reference is anything which can be represented by a + // distribution.Descriptor. These can consist of layers, resources or other + // manifests. + // + // While no particular order is required, implementations should return + // them from highest to lowest priority. For example, one might want to + // return the base layer before the top layer. + References() []Descriptor + + // Payload provides the serialized format of the manifest, in addition to + // the media type. + Payload() (mediaType string, payload []byte, err error) +} + +// ManifestBuilder creates a manifest allowing one to include dependencies. +// Instances can be obtained from a version-specific manifest package. Manifest +// specific data is passed into the function which creates the builder. +type ManifestBuilder interface { + // Build creates the manifest from his builder. + Build(ctx context.Context) (Manifest, error) + + // References returns a list of objects which have been added to this + // builder. The dependencies are returned in the order they were added, + // which should be from base to head. + References() []Descriptor + + // AppendReference includes the given object in the manifest after any + // existing dependencies. If the add fails, such as when adding an + // unsupported dependency, an error may be returned. + // + // The destination of the reference is dependent on the manifest type and + // the dependency type. + AppendReference(dependency Describable) error +} + +// ManifestService describes operations on image manifests. +type ManifestService interface { + // Exists returns true if the manifest exists. + Exists(ctx context.Context, dgst digest.Digest) (bool, error) + + // Get retrieves the manifest specified by the given digest + Get(ctx context.Context, dgst digest.Digest, options ...ManifestServiceOption) (Manifest, error) + + // Put creates or updates the given manifest returning the manifest digest + Put(ctx context.Context, manifest Manifest, options ...ManifestServiceOption) (digest.Digest, error) + + // Delete removes the manifest specified by the given digest. Deleting + // a manifest that doesn't exist will return ErrManifestNotFound + Delete(ctx context.Context, dgst digest.Digest) error +} + +// ManifestEnumerator enables iterating over manifests +type ManifestEnumerator interface { + // Enumerate calls ingester for each manifest. + Enumerate(ctx context.Context, ingester func(digest.Digest) error) error +} + +// Describable is an interface for descriptors +type Describable interface { + Descriptor() Descriptor +} + +// ManifestMediaTypes returns the supported media types for manifests. +func ManifestMediaTypes() (mediaTypes []string) { + for t := range mappings { + if t != "" { + mediaTypes = append(mediaTypes, t) + } + } + return +} + +// UnmarshalFunc implements manifest unmarshalling a given MediaType +type UnmarshalFunc func([]byte) (Manifest, Descriptor, error) + +var mappings = make(map[string]UnmarshalFunc, 0) + +// UnmarshalManifest looks up manifest unmarshal functions based on +// MediaType +func UnmarshalManifest(ctHeader string, p []byte) (Manifest, Descriptor, error) { + // Need to look up by the actual media type, not the raw contents of + // the header. Strip semicolons and anything following them. + var mediaType string + if ctHeader != "" { + var err error + mediaType, _, err = mime.ParseMediaType(ctHeader) + if err != nil { + return nil, Descriptor{}, err + } + } + + unmarshalFunc, ok := mappings[mediaType] + if !ok { + unmarshalFunc, ok = mappings[""] + if !ok { + return nil, Descriptor{}, fmt.Errorf("unsupported manifest media type and no default available: %s", mediaType) + } + } + + return unmarshalFunc(p) +} + +// RegisterManifestSchema registers an UnmarshalFunc for a given schema type. This +// should be called from specific +func RegisterManifestSchema(mediaType string, u UnmarshalFunc) error { + if _, ok := mappings[mediaType]; ok { + return fmt.Errorf("manifest media type registration would overwrite existing: %s", mediaType) + } + mappings[mediaType] = u + return nil +} diff --git a/vendor/github.com/docker/distribution/metrics/prometheus.go b/vendor/github.com/docker/distribution/metrics/prometheus.go new file mode 100644 index 000000000..b5a532144 --- /dev/null +++ b/vendor/github.com/docker/distribution/metrics/prometheus.go @@ -0,0 +1,13 @@ +package metrics + +import "github.com/docker/go-metrics" + +const ( + // NamespacePrefix is the namespace of prometheus metrics + NamespacePrefix = "registry" +) + +var ( + // StorageNamespace is the prometheus namespace of blob/cache related operations + StorageNamespace = metrics.NewNamespace(NamespacePrefix, "storage", nil) +) diff --git a/vendor/github.com/docker/distribution/reference/helpers.go b/vendor/github.com/docker/distribution/reference/helpers.go index a8f46cedd..978df7eab 100644 --- a/vendor/github.com/docker/distribution/reference/helpers.go +++ b/vendor/github.com/docker/distribution/reference/helpers.go @@ -1,5 +1,7 @@ package reference +import "path" + // IsNameOnly returns true if reference only contains a repo name. func IsNameOnly(ref Named) bool { if _, ok := ref.(NamedTagged); ok { @@ -14,7 +16,7 @@ func IsNameOnly(ref Named) bool { // FamiliarName returns the familiar name string // for the given named, familiarizing if needed. func FamiliarName(ref Named) string { - if nn, ok := ref.(NormalizedNamed); ok { + if nn, ok := ref.(normalizedNamed); ok { return nn.Familiar().Name() } return ref.Name() @@ -23,8 +25,18 @@ func FamiliarName(ref Named) string { // FamiliarString returns the familiar string representation // for the given reference, familiarizing if needed. func FamiliarString(ref Reference) string { - if nn, ok := ref.(NormalizedNamed); ok { + if nn, ok := ref.(normalizedNamed); ok { return nn.Familiar().String() } return ref.String() } + +// FamiliarMatch reports whether ref matches the specified pattern. +// See https://godoc.org/path#Match for supported patterns. +func FamiliarMatch(pattern string, ref Reference) (bool, error) { + matched, err := path.Match(pattern, FamiliarString(ref)) + if namedRef, isNamed := ref.(Named); isNamed && !matched { + matched, _ = path.Match(pattern, FamiliarName(namedRef)) + } + return matched, err +} diff --git a/vendor/github.com/docker/distribution/reference/normalize.go b/vendor/github.com/docker/distribution/reference/normalize.go index da797b705..2d71fc5e9 100644 --- a/vendor/github.com/docker/distribution/reference/normalize.go +++ b/vendor/github.com/docker/distribution/reference/normalize.go @@ -12,16 +12,16 @@ import ( var ( legacyDefaultDomain = "index.docker.io" defaultDomain = "docker.io" - defaultRepoPrefix = "library/" + officialRepoName = "library" defaultTag = "latest" ) -// NormalizedNamed represents a name which has been +// normalizedNamed represents a name which has been // normalized and has a familiar form. A familiar name // is what is used in Docker UI. An example normalized // name is "docker.io/library/ubuntu" and corresponding // familiar name of "ubuntu". -type NormalizedNamed interface { +type normalizedNamed interface { Named Familiar() Named } @@ -30,7 +30,7 @@ type NormalizedNamed interface { // transforming a familiar name from Docker UI to a fully // qualified reference. If the value may be an identifier // use ParseAnyReference. -func ParseNormalizedNamed(s string) (NormalizedNamed, error) { +func ParseNormalizedNamed(s string) (Named, error) { if ok := anchoredIdentifierRegexp.MatchString(s); ok { return nil, fmt.Errorf("invalid repository name (%s), cannot specify 64-byte hexadecimal strings", s) } @@ -49,7 +49,7 @@ func ParseNormalizedNamed(s string) (NormalizedNamed, error) { if err != nil { return nil, err } - named, isNamed := ref.(NormalizedNamed) + named, isNamed := ref.(Named) if !isNamed { return nil, fmt.Errorf("reference %s has no name", ref.String()) } @@ -70,7 +70,7 @@ func splitDockerDomain(name string) (domain, remainder string) { domain = defaultDomain } if domain == defaultDomain && !strings.ContainsRune(remainder, '/') { - remainder = defaultRepoPrefix + remainder + remainder = officialRepoName + "/" + remainder } return } @@ -89,7 +89,10 @@ func familiarizeName(named namedRepository) repository { if repo.domain == defaultDomain { repo.domain = "" - repo.path = strings.TrimPrefix(repo.path, defaultRepoPrefix) + // Handle official repositories which have the pattern "library/" + if split := strings.Split(repo.path, "/"); len(split) == 2 && split[0] == officialRepoName { + repo.path = split[1] + } } return repo } @@ -120,11 +123,10 @@ func (c canonicalReference) Familiar() Named { } } -// EnsureTagged adds the default tag "latest" to a reference if it only has +// TagNameOnly adds the default tag "latest" to a reference if it only has // a repo name. -func EnsureTagged(ref Named) NamedTagged { - namedTagged, ok := ref.(NamedTagged) - if !ok { +func TagNameOnly(ref Named) Named { + if IsNameOnly(ref) { namedTagged, err := WithTag(ref, defaultTag) if err != nil { // Default tag must be valid, to create a NamedTagged @@ -134,7 +136,7 @@ func EnsureTagged(ref Named) NamedTagged { } return namedTagged } - return namedTagged + return ref } // ParseAnyReference parses a reference string as a possible identifier, diff --git a/vendor/github.com/docker/distribution/reference/reference.go b/vendor/github.com/docker/distribution/reference/reference.go index 888e9b6d3..2f66cca87 100644 --- a/vendor/github.com/docker/distribution/reference/reference.go +++ b/vendor/github.com/docker/distribution/reference/reference.go @@ -15,19 +15,18 @@ // tag := /[\w][\w.-]{0,127}/ // // digest := digest-algorithm ":" digest-hex -// digest-algorithm := digest-algorithm-component [ digest-algorithm-separator digest-algorithm-component ] +// digest-algorithm := digest-algorithm-component [ digest-algorithm-separator digest-algorithm-component ]* // digest-algorithm-separator := /[+.-_]/ // digest-algorithm-component := /[A-Za-z][A-Za-z0-9]*/ // digest-hex := /[0-9a-fA-F]{32,}/ ; At least 128 bit digest value // -// identifier := /[a-f0-9]{64}/ -// short-identifier := /[a-f0-9]{6,64}/ +// identifier := /[a-f0-9]{64}/ +// short-identifier := /[a-f0-9]{6,64}/ package reference import ( "errors" "fmt" - "path" "strings" "github.com/opencontainers/go-digest" @@ -56,6 +55,9 @@ var ( // ErrNameTooLong is returned when a repository name is longer than NameTotalLengthMax. ErrNameTooLong = fmt.Errorf("repository name must not be more than %v characters", NameTotalLengthMax) + + // ErrNameNotCanonical is returned when a name is not canonical. + ErrNameNotCanonical = errors.New("repository name must be canonical") ) // Reference is an opaque object reference identifier that may include @@ -232,18 +234,17 @@ func Parse(s string) (Reference, error) { } // ParseNamed parses s and returns a syntactically valid reference implementing -// the Named interface. The reference must have a name, otherwise an error is -// returned. +// the Named interface. The reference must have a name and be in the canonical +// form, otherwise an error is returned. // If an error was encountered it is returned, along with a nil Reference. // NOTE: ParseNamed will not handle short digests. func ParseNamed(s string) (Named, error) { - ref, err := Parse(s) + named, err := ParseNormalizedNamed(s) if err != nil { return nil, err } - named, isNamed := ref.(Named) - if !isNamed { - return nil, fmt.Errorf("reference %s has no name", ref.String()) + if named.String() != s { + return nil, ErrNameNotCanonical } return named, nil } @@ -317,16 +318,6 @@ func WithDigest(name Named, digest digest.Digest) (Canonical, error) { }, nil } -// Match reports whether ref matches the specified pattern. -// See https://godoc.org/path#Match for supported patterns. -func Match(pattern string, ref Reference) (bool, error) { - matched, err := path.Match(pattern, ref.String()) - if namedRef, isNamed := ref.(Named); isNamed && !matched { - matched, _ = path.Match(pattern, namedRef.Name()) - } - return matched, err -} - // TrimNamed removes any tag or digest from the named reference. func TrimNamed(ref Named) Named { domain, path := SplitHostname(ref) @@ -408,7 +399,7 @@ func (r repository) Path() string { type digestReference digest.Digest func (d digestReference) String() string { - return d.String() + return digest.Digest(d).String() } func (d digestReference) Digest() digest.Digest { diff --git a/vendor/github.com/docker/distribution/reference/regexp.go b/vendor/github.com/docker/distribution/reference/regexp.go index 405e995db..786034932 100644 --- a/vendor/github.com/docker/distribution/reference/regexp.go +++ b/vendor/github.com/docker/distribution/reference/regexp.go @@ -20,15 +20,15 @@ var ( optional(repeated(separatorRegexp, alphaNumericRegexp))) // domainComponentRegexp restricts the registry domain component of a - // repository name to start with a component as defined by domainRegexp + // repository name to start with a component as defined by DomainRegexp // and followed by an optional port. domainComponentRegexp = match(`(?:[a-zA-Z0-9]|[a-zA-Z0-9][a-zA-Z0-9-]*[a-zA-Z0-9])`) - // domainRegexp defines the structure of potential domain components + // DomainRegexp defines the structure of potential domain components // that may be part of image names. This is purposely a subset of what is // allowed by DNS to ensure backwards compatibility with Docker image // names. - domainRegexp = expression( + DomainRegexp = expression( domainComponentRegexp, optional(repeated(literal(`.`), domainComponentRegexp)), optional(literal(`:`), match(`[0-9]+`))) @@ -51,14 +51,14 @@ var ( // regexp has capturing groups for the domain and name part omitting // the separating forward slash from either. NameRegexp = expression( - optional(domainRegexp, literal(`/`)), + optional(DomainRegexp, literal(`/`)), nameComponentRegexp, optional(repeated(literal(`/`), nameComponentRegexp))) // anchoredNameRegexp is used to parse a name value, capturing the // domain and trailing components. anchoredNameRegexp = anchored( - optional(capture(domainRegexp), literal(`/`)), + optional(capture(DomainRegexp), literal(`/`)), capture(nameComponentRegexp, optional(repeated(literal(`/`), nameComponentRegexp)))) diff --git a/vendor/github.com/docker/distribution/registry.go b/vendor/github.com/docker/distribution/registry.go new file mode 100644 index 000000000..6c3210989 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry.go @@ -0,0 +1,118 @@ +package distribution + +import ( + "context" + + "github.com/docker/distribution/reference" +) + +// Scope defines the set of items that match a namespace. +type Scope interface { + // Contains returns true if the name belongs to the namespace. + Contains(name string) bool +} + +type fullScope struct{} + +func (f fullScope) Contains(string) bool { + return true +} + +// GlobalScope represents the full namespace scope which contains +// all other scopes. +var GlobalScope = Scope(fullScope{}) + +// Namespace represents a collection of repositories, addressable by name. +// Generally, a namespace is backed by a set of one or more services, +// providing facilities such as registry access, trust, and indexing. +type Namespace interface { + // Scope describes the names that can be used with this Namespace. The + // global namespace will have a scope that matches all names. The scope + // effectively provides an identity for the namespace. + Scope() Scope + + // Repository should return a reference to the named repository. The + // registry may or may not have the repository but should always return a + // reference. + Repository(ctx context.Context, name reference.Named) (Repository, error) + + // Repositories fills 'repos' with a lexicographically sorted catalog of repositories + // up to the size of 'repos' and returns the value 'n' for the number of entries + // which were filled. 'last' contains an offset in the catalog, and 'err' will be + // set to io.EOF if there are no more entries to obtain. + Repositories(ctx context.Context, repos []string, last string) (n int, err error) + + // Blobs returns a blob enumerator to access all blobs + Blobs() BlobEnumerator + + // BlobStatter returns a BlobStatter to control + BlobStatter() BlobStatter +} + +// RepositoryEnumerator describes an operation to enumerate repositories +type RepositoryEnumerator interface { + Enumerate(ctx context.Context, ingester func(string) error) error +} + +// RepositoryRemover removes given repository +type RepositoryRemover interface { + Remove(ctx context.Context, name reference.Named) error +} + +// ManifestServiceOption is a function argument for Manifest Service methods +type ManifestServiceOption interface { + Apply(ManifestService) error +} + +// WithTag allows a tag to be passed into Put +func WithTag(tag string) ManifestServiceOption { + return WithTagOption{tag} +} + +// WithTagOption holds a tag +type WithTagOption struct{ Tag string } + +// Apply conforms to the ManifestServiceOption interface +func (o WithTagOption) Apply(m ManifestService) error { + // no implementation + return nil +} + +// WithManifestMediaTypes lists the media types the client wishes +// the server to provide. +func WithManifestMediaTypes(mediaTypes []string) ManifestServiceOption { + return WithManifestMediaTypesOption{mediaTypes} +} + +// WithManifestMediaTypesOption holds a list of accepted media types +type WithManifestMediaTypesOption struct{ MediaTypes []string } + +// Apply conforms to the ManifestServiceOption interface +func (o WithManifestMediaTypesOption) Apply(m ManifestService) error { + // no implementation + return nil +} + +// Repository is a named collection of manifests and layers. +type Repository interface { + // Named returns the name of the repository. + Named() reference.Named + + // Manifests returns a reference to this repository's manifest service. + // with the supplied options applied. + Manifests(ctx context.Context, options ...ManifestServiceOption) (ManifestService, error) + + // Blobs returns a reference to this repository's blob service. + Blobs(ctx context.Context) BlobStore + + // TODO(stevvooe): The above BlobStore return can probably be relaxed to + // be a BlobService for use with clients. This will allow such + // implementations to avoid implementing ServeBlob. + + // Tags returns a reference to this repositories tag service + Tags(ctx context.Context) TagService +} + +// TODO(stevvooe): Must add close methods to all these. May want to change the +// way instances are created to better reflect internal dependency +// relationships. diff --git a/vendor/github.com/docker/distribution/registry/api/errcode/errors.go b/vendor/github.com/docker/distribution/registry/api/errcode/errors.go new file mode 100644 index 000000000..6d9bb4b62 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/api/errcode/errors.go @@ -0,0 +1,267 @@ +package errcode + +import ( + "encoding/json" + "fmt" + "strings" +) + +// ErrorCoder is the base interface for ErrorCode and Error allowing +// users of each to just call ErrorCode to get the real ID of each +type ErrorCoder interface { + ErrorCode() ErrorCode +} + +// ErrorCode represents the error type. The errors are serialized via strings +// and the integer format may change and should *never* be exported. +type ErrorCode int + +var _ error = ErrorCode(0) + +// ErrorCode just returns itself +func (ec ErrorCode) ErrorCode() ErrorCode { + return ec +} + +// Error returns the ID/Value +func (ec ErrorCode) Error() string { + // NOTE(stevvooe): Cannot use message here since it may have unpopulated args. + return strings.ToLower(strings.Replace(ec.String(), "_", " ", -1)) +} + +// Descriptor returns the descriptor for the error code. +func (ec ErrorCode) Descriptor() ErrorDescriptor { + d, ok := errorCodeToDescriptors[ec] + + if !ok { + return ErrorCodeUnknown.Descriptor() + } + + return d +} + +// String returns the canonical identifier for this error code. +func (ec ErrorCode) String() string { + return ec.Descriptor().Value +} + +// Message returned the human-readable error message for this error code. +func (ec ErrorCode) Message() string { + return ec.Descriptor().Message +} + +// MarshalText encodes the receiver into UTF-8-encoded text and returns the +// result. +func (ec ErrorCode) MarshalText() (text []byte, err error) { + return []byte(ec.String()), nil +} + +// UnmarshalText decodes the form generated by MarshalText. +func (ec *ErrorCode) UnmarshalText(text []byte) error { + desc, ok := idToDescriptors[string(text)] + + if !ok { + desc = ErrorCodeUnknown.Descriptor() + } + + *ec = desc.Code + + return nil +} + +// WithMessage creates a new Error struct based on the passed-in info and +// overrides the Message property. +func (ec ErrorCode) WithMessage(message string) Error { + return Error{ + Code: ec, + Message: message, + } +} + +// WithDetail creates a new Error struct based on the passed-in info and +// set the Detail property appropriately +func (ec ErrorCode) WithDetail(detail interface{}) Error { + return Error{ + Code: ec, + Message: ec.Message(), + }.WithDetail(detail) +} + +// WithArgs creates a new Error struct and sets the Args slice +func (ec ErrorCode) WithArgs(args ...interface{}) Error { + return Error{ + Code: ec, + Message: ec.Message(), + }.WithArgs(args...) +} + +// Error provides a wrapper around ErrorCode with extra Details provided. +type Error struct { + Code ErrorCode `json:"code"` + Message string `json:"message"` + Detail interface{} `json:"detail,omitempty"` + + // TODO(duglin): See if we need an "args" property so we can do the + // variable substitution right before showing the message to the user +} + +var _ error = Error{} + +// ErrorCode returns the ID/Value of this Error +func (e Error) ErrorCode() ErrorCode { + return e.Code +} + +// Error returns a human readable representation of the error. +func (e Error) Error() string { + return fmt.Sprintf("%s: %s", e.Code.Error(), e.Message) +} + +// WithDetail will return a new Error, based on the current one, but with +// some Detail info added +func (e Error) WithDetail(detail interface{}) Error { + return Error{ + Code: e.Code, + Message: e.Message, + Detail: detail, + } +} + +// WithArgs uses the passed-in list of interface{} as the substitution +// variables in the Error's Message string, but returns a new Error +func (e Error) WithArgs(args ...interface{}) Error { + return Error{ + Code: e.Code, + Message: fmt.Sprintf(e.Code.Message(), args...), + Detail: e.Detail, + } +} + +// ErrorDescriptor provides relevant information about a given error code. +type ErrorDescriptor struct { + // Code is the error code that this descriptor describes. + Code ErrorCode + + // Value provides a unique, string key, often captilized with + // underscores, to identify the error code. This value is used as the + // keyed value when serializing api errors. + Value string + + // Message is a short, human readable decription of the error condition + // included in API responses. + Message string + + // Description provides a complete account of the errors purpose, suitable + // for use in documentation. + Description string + + // HTTPStatusCode provides the http status code that is associated with + // this error condition. + HTTPStatusCode int +} + +// ParseErrorCode returns the value by the string error code. +// `ErrorCodeUnknown` will be returned if the error is not known. +func ParseErrorCode(value string) ErrorCode { + ed, ok := idToDescriptors[value] + if ok { + return ed.Code + } + + return ErrorCodeUnknown +} + +// Errors provides the envelope for multiple errors and a few sugar methods +// for use within the application. +type Errors []error + +var _ error = Errors{} + +func (errs Errors) Error() string { + switch len(errs) { + case 0: + return "" + case 1: + return errs[0].Error() + default: + msg := "errors:\n" + for _, err := range errs { + msg += err.Error() + "\n" + } + return msg + } +} + +// Len returns the current number of errors. +func (errs Errors) Len() int { + return len(errs) +} + +// MarshalJSON converts slice of error, ErrorCode or Error into a +// slice of Error - then serializes +func (errs Errors) MarshalJSON() ([]byte, error) { + var tmpErrs struct { + Errors []Error `json:"errors,omitempty"` + } + + for _, daErr := range errs { + var err Error + + switch daErr.(type) { + case ErrorCode: + err = daErr.(ErrorCode).WithDetail(nil) + case Error: + err = daErr.(Error) + default: + err = ErrorCodeUnknown.WithDetail(daErr) + + } + + // If the Error struct was setup and they forgot to set the + // Message field (meaning its "") then grab it from the ErrCode + msg := err.Message + if msg == "" { + msg = err.Code.Message() + } + + tmpErrs.Errors = append(tmpErrs.Errors, Error{ + Code: err.Code, + Message: msg, + Detail: err.Detail, + }) + } + + return json.Marshal(tmpErrs) +} + +// UnmarshalJSON deserializes []Error and then converts it into slice of +// Error or ErrorCode +func (errs *Errors) UnmarshalJSON(data []byte) error { + var tmpErrs struct { + Errors []Error + } + + if err := json.Unmarshal(data, &tmpErrs); err != nil { + return err + } + + var newErrs Errors + for _, daErr := range tmpErrs.Errors { + // If Message is empty or exactly matches the Code's message string + // then just use the Code, no need for a full Error struct + if daErr.Detail == nil && (daErr.Message == "" || daErr.Message == daErr.Code.Message()) { + // Error's w/o details get converted to ErrorCode + newErrs = append(newErrs, daErr.Code) + } else { + // Error's w/ details are untouched + newErrs = append(newErrs, Error{ + Code: daErr.Code, + Message: daErr.Message, + Detail: daErr.Detail, + }) + } + } + + *errs = newErrs + return nil +} diff --git a/vendor/github.com/docker/distribution/registry/api/errcode/handler.go b/vendor/github.com/docker/distribution/registry/api/errcode/handler.go new file mode 100644 index 000000000..d77e70473 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/api/errcode/handler.go @@ -0,0 +1,40 @@ +package errcode + +import ( + "encoding/json" + "net/http" +) + +// ServeJSON attempts to serve the errcode in a JSON envelope. It marshals err +// and sets the content-type header to 'application/json'. It will handle +// ErrorCoder and Errors, and if necessary will create an envelope. +func ServeJSON(w http.ResponseWriter, err error) error { + w.Header().Set("Content-Type", "application/json; charset=utf-8") + var sc int + + switch errs := err.(type) { + case Errors: + if len(errs) < 1 { + break + } + + if err, ok := errs[0].(ErrorCoder); ok { + sc = err.ErrorCode().Descriptor().HTTPStatusCode + } + case ErrorCoder: + sc = errs.ErrorCode().Descriptor().HTTPStatusCode + err = Errors{err} // create an envelope. + default: + // We just have an unhandled error type, so just place in an envelope + // and move along. + err = Errors{err} + } + + if sc == 0 { + sc = http.StatusInternalServerError + } + + w.WriteHeader(sc) + + return json.NewEncoder(w).Encode(err) +} diff --git a/vendor/github.com/docker/distribution/registry/api/errcode/register.go b/vendor/github.com/docker/distribution/registry/api/errcode/register.go new file mode 100644 index 000000000..d1e8826c6 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/api/errcode/register.go @@ -0,0 +1,138 @@ +package errcode + +import ( + "fmt" + "net/http" + "sort" + "sync" +) + +var ( + errorCodeToDescriptors = map[ErrorCode]ErrorDescriptor{} + idToDescriptors = map[string]ErrorDescriptor{} + groupToDescriptors = map[string][]ErrorDescriptor{} +) + +var ( + // ErrorCodeUnknown is a generic error that can be used as a last + // resort if there is no situation-specific error message that can be used + ErrorCodeUnknown = Register("errcode", ErrorDescriptor{ + Value: "UNKNOWN", + Message: "unknown error", + Description: `Generic error returned when the error does not have an + API classification.`, + HTTPStatusCode: http.StatusInternalServerError, + }) + + // ErrorCodeUnsupported is returned when an operation is not supported. + ErrorCodeUnsupported = Register("errcode", ErrorDescriptor{ + Value: "UNSUPPORTED", + Message: "The operation is unsupported.", + Description: `The operation was unsupported due to a missing + implementation or invalid set of parameters.`, + HTTPStatusCode: http.StatusMethodNotAllowed, + }) + + // ErrorCodeUnauthorized is returned if a request requires + // authentication. + ErrorCodeUnauthorized = Register("errcode", ErrorDescriptor{ + Value: "UNAUTHORIZED", + Message: "authentication required", + Description: `The access controller was unable to authenticate + the client. Often this will be accompanied by a + Www-Authenticate HTTP response header indicating how to + authenticate.`, + HTTPStatusCode: http.StatusUnauthorized, + }) + + // ErrorCodeDenied is returned if a client does not have sufficient + // permission to perform an action. + ErrorCodeDenied = Register("errcode", ErrorDescriptor{ + Value: "DENIED", + Message: "requested access to the resource is denied", + Description: `The access controller denied access for the + operation on a resource.`, + HTTPStatusCode: http.StatusForbidden, + }) + + // ErrorCodeUnavailable provides a common error to report unavailability + // of a service or endpoint. + ErrorCodeUnavailable = Register("errcode", ErrorDescriptor{ + Value: "UNAVAILABLE", + Message: "service unavailable", + Description: "Returned when a service is not available", + HTTPStatusCode: http.StatusServiceUnavailable, + }) + + // ErrorCodeTooManyRequests is returned if a client attempts too many + // times to contact a service endpoint. + ErrorCodeTooManyRequests = Register("errcode", ErrorDescriptor{ + Value: "TOOMANYREQUESTS", + Message: "too many requests", + Description: `Returned when a client attempts to contact a + service too many times`, + HTTPStatusCode: http.StatusTooManyRequests, + }) +) + +var nextCode = 1000 +var registerLock sync.Mutex + +// Register will make the passed-in error known to the environment and +// return a new ErrorCode +func Register(group string, descriptor ErrorDescriptor) ErrorCode { + registerLock.Lock() + defer registerLock.Unlock() + + descriptor.Code = ErrorCode(nextCode) + + if _, ok := idToDescriptors[descriptor.Value]; ok { + panic(fmt.Sprintf("ErrorValue %q is already registered", descriptor.Value)) + } + if _, ok := errorCodeToDescriptors[descriptor.Code]; ok { + panic(fmt.Sprintf("ErrorCode %v is already registered", descriptor.Code)) + } + + groupToDescriptors[group] = append(groupToDescriptors[group], descriptor) + errorCodeToDescriptors[descriptor.Code] = descriptor + idToDescriptors[descriptor.Value] = descriptor + + nextCode++ + return descriptor.Code +} + +type byValue []ErrorDescriptor + +func (a byValue) Len() int { return len(a) } +func (a byValue) Swap(i, j int) { a[i], a[j] = a[j], a[i] } +func (a byValue) Less(i, j int) bool { return a[i].Value < a[j].Value } + +// GetGroupNames returns the list of Error group names that are registered +func GetGroupNames() []string { + keys := []string{} + + for k := range groupToDescriptors { + keys = append(keys, k) + } + sort.Strings(keys) + return keys +} + +// GetErrorCodeGroup returns the named group of error descriptors +func GetErrorCodeGroup(name string) []ErrorDescriptor { + desc := groupToDescriptors[name] + sort.Sort(byValue(desc)) + return desc +} + +// GetErrorAllDescriptors returns a slice of all ErrorDescriptors that are +// registered, irrespective of what group they're in +func GetErrorAllDescriptors() []ErrorDescriptor { + result := []ErrorDescriptor{} + + for _, group := range GetGroupNames() { + result = append(result, GetErrorCodeGroup(group)...) + } + sort.Sort(byValue(result)) + return result +} diff --git a/vendor/github.com/docker/distribution/registry/api/v2/descriptors.go b/vendor/github.com/docker/distribution/registry/api/v2/descriptors.go new file mode 100644 index 000000000..a9616c58a --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/api/v2/descriptors.go @@ -0,0 +1,1596 @@ +package v2 + +import ( + "net/http" + "regexp" + + "github.com/docker/distribution/reference" + "github.com/docker/distribution/registry/api/errcode" + "github.com/opencontainers/go-digest" +) + +var ( + nameParameterDescriptor = ParameterDescriptor{ + Name: "name", + Type: "string", + Format: reference.NameRegexp.String(), + Required: true, + Description: `Name of the target repository.`, + } + + referenceParameterDescriptor = ParameterDescriptor{ + Name: "reference", + Type: "string", + Format: reference.TagRegexp.String(), + Required: true, + Description: `Tag or digest of the target manifest.`, + } + + uuidParameterDescriptor = ParameterDescriptor{ + Name: "uuid", + Type: "opaque", + Required: true, + Description: "A uuid identifying the upload. This field can accept characters that match `[a-zA-Z0-9-_.=]+`.", + } + + digestPathParameter = ParameterDescriptor{ + Name: "digest", + Type: "path", + Required: true, + Format: digest.DigestRegexp.String(), + Description: `Digest of desired blob.`, + } + + hostHeader = ParameterDescriptor{ + Name: "Host", + Type: "string", + Description: "Standard HTTP Host Header. Should be set to the registry host.", + Format: "", + Examples: []string{"registry-1.docker.io"}, + } + + authHeader = ParameterDescriptor{ + Name: "Authorization", + Type: "string", + Description: "An RFC7235 compliant authorization header.", + Format: " ", + Examples: []string{"Bearer dGhpcyBpcyBhIGZha2UgYmVhcmVyIHRva2VuIQ=="}, + } + + authChallengeHeader = ParameterDescriptor{ + Name: "WWW-Authenticate", + Type: "string", + Description: "An RFC7235 compliant authentication challenge header.", + Format: ` realm="", ..."`, + Examples: []string{ + `Bearer realm="https://auth.docker.com/", service="registry.docker.com", scopes="repository:library/ubuntu:pull"`, + }, + } + + contentLengthZeroHeader = ParameterDescriptor{ + Name: "Content-Length", + Description: "The `Content-Length` header must be zero and the body must be empty.", + Type: "integer", + Format: "0", + } + + dockerUploadUUIDHeader = ParameterDescriptor{ + Name: "Docker-Upload-UUID", + Description: "Identifies the docker upload uuid for the current request.", + Type: "uuid", + Format: "", + } + + digestHeader = ParameterDescriptor{ + Name: "Docker-Content-Digest", + Description: "Digest of the targeted content for the request.", + Type: "digest", + Format: "", + } + + linkHeader = ParameterDescriptor{ + Name: "Link", + Type: "link", + Description: "RFC5988 compliant rel='next' with URL to next result set, if available", + Format: `<?n=&last=>; rel="next"`, + } + + paginationParameters = []ParameterDescriptor{ + { + Name: "n", + Type: "integer", + Description: "Limit the number of entries in each response. It not present, all entries will be returned.", + Format: "", + Required: false, + }, + { + Name: "last", + Type: "string", + Description: "Result set will include values lexically after last.", + Format: "", + Required: false, + }, + } + + unauthorizedResponseDescriptor = ResponseDescriptor{ + Name: "Authentication Required", + StatusCode: http.StatusUnauthorized, + Description: "The client is not authenticated.", + Headers: []ParameterDescriptor{ + authChallengeHeader, + { + Name: "Content-Length", + Type: "integer", + Description: "Length of the JSON response body.", + Format: "", + }, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + ErrorCodes: []errcode.ErrorCode{ + errcode.ErrorCodeUnauthorized, + }, + } + + repositoryNotFoundResponseDescriptor = ResponseDescriptor{ + Name: "No Such Repository Error", + StatusCode: http.StatusNotFound, + Description: "The repository is not known to the registry.", + Headers: []ParameterDescriptor{ + { + Name: "Content-Length", + Type: "integer", + Description: "Length of the JSON response body.", + Format: "", + }, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeNameUnknown, + }, + } + + deniedResponseDescriptor = ResponseDescriptor{ + Name: "Access Denied", + StatusCode: http.StatusForbidden, + Description: "The client does not have required access to the repository.", + Headers: []ParameterDescriptor{ + { + Name: "Content-Length", + Type: "integer", + Description: "Length of the JSON response body.", + Format: "", + }, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + ErrorCodes: []errcode.ErrorCode{ + errcode.ErrorCodeDenied, + }, + } + + tooManyRequestsDescriptor = ResponseDescriptor{ + Name: "Too Many Requests", + StatusCode: http.StatusTooManyRequests, + Description: "The client made too many requests within a time interval.", + Headers: []ParameterDescriptor{ + { + Name: "Content-Length", + Type: "integer", + Description: "Length of the JSON response body.", + Format: "", + }, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + ErrorCodes: []errcode.ErrorCode{ + errcode.ErrorCodeTooManyRequests, + }, + } +) + +const ( + manifestBody = `{ + "name": , + "tag": , + "fsLayers": [ + { + "blobSum": "" + }, + ... + ] + ], + "history": , + "signature": +}` + + errorsBody = `{ + "errors:" [ + { + "code": , + "message": "", + "detail": ... + }, + ... + ] +}` +) + +// APIDescriptor exports descriptions of the layout of the v2 registry API. +var APIDescriptor = struct { + // RouteDescriptors provides a list of the routes available in the API. + RouteDescriptors []RouteDescriptor +}{ + RouteDescriptors: routeDescriptors, +} + +// RouteDescriptor describes a route specified by name. +type RouteDescriptor struct { + // Name is the name of the route, as specified in RouteNameXXX exports. + // These names a should be considered a unique reference for a route. If + // the route is registered with gorilla, this is the name that will be + // used. + Name string + + // Path is a gorilla/mux-compatible regexp that can be used to match the + // route. For any incoming method and path, only one route descriptor + // should match. + Path string + + // Entity should be a short, human-readalbe description of the object + // targeted by the endpoint. + Entity string + + // Description should provide an accurate overview of the functionality + // provided by the route. + Description string + + // Methods should describe the various HTTP methods that may be used on + // this route, including request and response formats. + Methods []MethodDescriptor +} + +// MethodDescriptor provides a description of the requests that may be +// conducted with the target method. +type MethodDescriptor struct { + + // Method is an HTTP method, such as GET, PUT or POST. + Method string + + // Description should provide an overview of the functionality provided by + // the covered method, suitable for use in documentation. Use of markdown + // here is encouraged. + Description string + + // Requests is a slice of request descriptors enumerating how this + // endpoint may be used. + Requests []RequestDescriptor +} + +// RequestDescriptor covers a particular set of headers and parameters that +// can be carried out with the parent method. Its most helpful to have one +// RequestDescriptor per API use case. +type RequestDescriptor struct { + // Name provides a short identifier for the request, usable as a title or + // to provide quick context for the particular request. + Name string + + // Description should cover the requests purpose, covering any details for + // this particular use case. + Description string + + // Headers describes headers that must be used with the HTTP request. + Headers []ParameterDescriptor + + // PathParameters enumerate the parameterized path components for the + // given request, as defined in the route's regular expression. + PathParameters []ParameterDescriptor + + // QueryParameters provides a list of query parameters for the given + // request. + QueryParameters []ParameterDescriptor + + // Body describes the format of the request body. + Body BodyDescriptor + + // Successes enumerates the possible responses that are considered to be + // the result of a successful request. + Successes []ResponseDescriptor + + // Failures covers the possible failures from this particular request. + Failures []ResponseDescriptor +} + +// ResponseDescriptor describes the components of an API response. +type ResponseDescriptor struct { + // Name provides a short identifier for the response, usable as a title or + // to provide quick context for the particular response. + Name string + + // Description should provide a brief overview of the role of the + // response. + Description string + + // StatusCode specifies the status received by this particular response. + StatusCode int + + // Headers covers any headers that may be returned from the response. + Headers []ParameterDescriptor + + // Fields describes any fields that may be present in the response. + Fields []ParameterDescriptor + + // ErrorCodes enumerates the error codes that may be returned along with + // the response. + ErrorCodes []errcode.ErrorCode + + // Body describes the body of the response, if any. + Body BodyDescriptor +} + +// BodyDescriptor describes a request body and its expected content type. For +// the most part, it should be example json or some placeholder for body +// data in documentation. +type BodyDescriptor struct { + ContentType string + Format string +} + +// ParameterDescriptor describes the format of a request parameter, which may +// be a header, path parameter or query parameter. +type ParameterDescriptor struct { + // Name is the name of the parameter, either of the path component or + // query parameter. + Name string + + // Type specifies the type of the parameter, such as string, integer, etc. + Type string + + // Description provides a human-readable description of the parameter. + Description string + + // Required means the field is required when set. + Required bool + + // Format is a specifying the string format accepted by this parameter. + Format string + + // Regexp is a compiled regular expression that can be used to validate + // the contents of the parameter. + Regexp *regexp.Regexp + + // Examples provides multiple examples for the values that might be valid + // for this parameter. + Examples []string +} + +var routeDescriptors = []RouteDescriptor{ + { + Name: RouteNameBase, + Path: "/v2/", + Entity: "Base", + Description: `Base V2 API route. Typically, this can be used for lightweight version checks and to validate registry authentication.`, + Methods: []MethodDescriptor{ + { + Method: "GET", + Description: "Check that the endpoint implements Docker Registry API V2.", + Requests: []RequestDescriptor{ + { + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + }, + Successes: []ResponseDescriptor{ + { + Description: "The API implements V2 protocol and is accessible.", + StatusCode: http.StatusOK, + }, + }, + Failures: []ResponseDescriptor{ + { + Description: "The registry does not implement the V2 API.", + StatusCode: http.StatusNotFound, + }, + unauthorizedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + }, + }, + }, + }, + { + Name: RouteNameTags, + Path: "/v2/{name:" + reference.NameRegexp.String() + "}/tags/list", + Entity: "Tags", + Description: "Retrieve information about tags.", + Methods: []MethodDescriptor{ + { + Method: "GET", + Description: "Fetch the tags under the repository identified by `name`.", + Requests: []RequestDescriptor{ + { + Name: "Tags", + Description: "Return all tags for the repository", + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + }, + Successes: []ResponseDescriptor{ + { + StatusCode: http.StatusOK, + Description: "A list of tags for the named repository.", + Headers: []ParameterDescriptor{ + { + Name: "Content-Length", + Type: "integer", + Description: "Length of the JSON response body.", + Format: "", + }, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: `{ + "name": , + "tags": [ + , + ... + ] +}`, + }, + }, + }, + Failures: []ResponseDescriptor{ + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + { + Name: "Tags Paginated", + Description: "Return a portion of the tags for the specified repository.", + PathParameters: []ParameterDescriptor{nameParameterDescriptor}, + QueryParameters: paginationParameters, + Successes: []ResponseDescriptor{ + { + StatusCode: http.StatusOK, + Description: "A list of tags for the named repository.", + Headers: []ParameterDescriptor{ + { + Name: "Content-Length", + Type: "integer", + Description: "Length of the JSON response body.", + Format: "", + }, + linkHeader, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: `{ + "name": , + "tags": [ + , + ... + ], +}`, + }, + }, + }, + Failures: []ResponseDescriptor{ + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + }, + }, + }, + }, + { + Name: RouteNameManifest, + Path: "/v2/{name:" + reference.NameRegexp.String() + "}/manifests/{reference:" + reference.TagRegexp.String() + "|" + digest.DigestRegexp.String() + "}", + Entity: "Manifest", + Description: "Create, update, delete and retrieve manifests.", + Methods: []MethodDescriptor{ + { + Method: "GET", + Description: "Fetch the manifest identified by `name` and `reference` where `reference` can be a tag or digest. A `HEAD` request can also be issued to this endpoint to obtain resource information without receiving all data.", + Requests: []RequestDescriptor{ + { + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + referenceParameterDescriptor, + }, + Successes: []ResponseDescriptor{ + { + Description: "The manifest identified by `name` and `reference`. The contents can be used to identify and resolve resources required to run the specified image.", + StatusCode: http.StatusOK, + Headers: []ParameterDescriptor{ + digestHeader, + }, + Body: BodyDescriptor{ + ContentType: "", + Format: manifestBody, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Description: "The name or reference was invalid.", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeNameInvalid, + ErrorCodeTagInvalid, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + }, + }, + { + Method: "PUT", + Description: "Put the manifest identified by `name` and `reference` where `reference` can be a tag or digest.", + Requests: []RequestDescriptor{ + { + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + referenceParameterDescriptor, + }, + Body: BodyDescriptor{ + ContentType: "", + Format: manifestBody, + }, + Successes: []ResponseDescriptor{ + { + Description: "The manifest has been accepted by the registry and is stored under the specified `name` and `tag`.", + StatusCode: http.StatusCreated, + Headers: []ParameterDescriptor{ + { + Name: "Location", + Type: "url", + Description: "The canonical location url of the uploaded manifest.", + Format: "", + }, + contentLengthZeroHeader, + digestHeader, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Name: "Invalid Manifest", + Description: "The received manifest was invalid in some way, as described by the error codes. The client should resolve the issue and retry the request.", + StatusCode: http.StatusBadRequest, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeNameInvalid, + ErrorCodeTagInvalid, + ErrorCodeManifestInvalid, + ErrorCodeManifestUnverified, + ErrorCodeBlobUnknown, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + { + Name: "Missing Layer(s)", + Description: "One or more layers may be missing during a manifest upload. If so, the missing layers will be enumerated in the error response.", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeBlobUnknown, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: `{ + "errors:" [{ + "code": "BLOB_UNKNOWN", + "message": "blob unknown to registry", + "detail": { + "digest": "" + } + }, + ... + ] +}`, + }, + }, + { + Name: "Not allowed", + Description: "Manifest put is not allowed because the registry is configured as a pull-through cache or for some other reason", + StatusCode: http.StatusMethodNotAllowed, + ErrorCodes: []errcode.ErrorCode{ + errcode.ErrorCodeUnsupported, + }, + }, + }, + }, + }, + }, + { + Method: "DELETE", + Description: "Delete the manifest identified by `name` and `reference`. Note that a manifest can _only_ be deleted by `digest`.", + Requests: []RequestDescriptor{ + { + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + referenceParameterDescriptor, + }, + Successes: []ResponseDescriptor{ + { + StatusCode: http.StatusAccepted, + }, + }, + Failures: []ResponseDescriptor{ + { + Name: "Invalid Name or Reference", + Description: "The specified `name` or `reference` were invalid and the delete was unable to proceed.", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeNameInvalid, + ErrorCodeTagInvalid, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + { + Name: "Unknown Manifest", + Description: "The specified `name` or `reference` are unknown to the registry and the delete was unable to proceed. Clients can assume the manifest was already deleted if this response is returned.", + StatusCode: http.StatusNotFound, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeNameUnknown, + ErrorCodeManifestUnknown, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + { + Name: "Not allowed", + Description: "Manifest delete is not allowed because the registry is configured as a pull-through cache or `delete` has been disabled.", + StatusCode: http.StatusMethodNotAllowed, + ErrorCodes: []errcode.ErrorCode{ + errcode.ErrorCodeUnsupported, + }, + }, + }, + }, + }, + }, + }, + }, + + { + Name: RouteNameBlob, + Path: "/v2/{name:" + reference.NameRegexp.String() + "}/blobs/{digest:" + digest.DigestRegexp.String() + "}", + Entity: "Blob", + Description: "Operations on blobs identified by `name` and `digest`. Used to fetch or delete layers by digest.", + Methods: []MethodDescriptor{ + { + Method: "GET", + Description: "Retrieve the blob from the registry identified by `digest`. A `HEAD` request can also be issued to this endpoint to obtain resource information without receiving all data.", + Requests: []RequestDescriptor{ + { + Name: "Fetch Blob", + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + digestPathParameter, + }, + Successes: []ResponseDescriptor{ + { + Description: "The blob identified by `digest` is available. The blob content will be present in the body of the request.", + StatusCode: http.StatusOK, + Headers: []ParameterDescriptor{ + { + Name: "Content-Length", + Type: "integer", + Description: "The length of the requested blob content.", + Format: "", + }, + digestHeader, + }, + Body: BodyDescriptor{ + ContentType: "application/octet-stream", + Format: "", + }, + }, + { + Description: "The blob identified by `digest` is available at the provided location.", + StatusCode: http.StatusTemporaryRedirect, + Headers: []ParameterDescriptor{ + { + Name: "Location", + Type: "url", + Description: "The location where the layer should be accessible.", + Format: "", + }, + digestHeader, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Description: "There was a problem with the request that needs to be addressed by the client, such as an invalid `name` or `tag`.", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeNameInvalid, + ErrorCodeDigestInvalid, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + { + Description: "The blob, identified by `name` and `digest`, is unknown to the registry.", + StatusCode: http.StatusNotFound, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeNameUnknown, + ErrorCodeBlobUnknown, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + { + Name: "Fetch Blob Part", + Description: "This endpoint may also support RFC7233 compliant range requests. Support can be detected by issuing a HEAD request. If the header `Accept-Range: bytes` is returned, range requests can be used to fetch partial content.", + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + { + Name: "Range", + Type: "string", + Description: "HTTP Range header specifying blob chunk.", + Format: "bytes=-", + }, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + digestPathParameter, + }, + Successes: []ResponseDescriptor{ + { + Description: "The blob identified by `digest` is available. The specified chunk of blob content will be present in the body of the request.", + StatusCode: http.StatusPartialContent, + Headers: []ParameterDescriptor{ + { + Name: "Content-Length", + Type: "integer", + Description: "The length of the requested blob chunk.", + Format: "", + }, + { + Name: "Content-Range", + Type: "byte range", + Description: "Content range of blob chunk.", + Format: "bytes -/", + }, + }, + Body: BodyDescriptor{ + ContentType: "application/octet-stream", + Format: "", + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Description: "There was a problem with the request that needs to be addressed by the client, such as an invalid `name` or `tag`.", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeNameInvalid, + ErrorCodeDigestInvalid, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + { + StatusCode: http.StatusNotFound, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeNameUnknown, + ErrorCodeBlobUnknown, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + { + Description: "The range specification cannot be satisfied for the requested content. This can happen when the range is not formatted correctly or if the range is outside of the valid size of the content.", + StatusCode: http.StatusRequestedRangeNotSatisfiable, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + }, + }, + { + Method: "DELETE", + Description: "Delete the blob identified by `name` and `digest`", + Requests: []RequestDescriptor{ + { + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + digestPathParameter, + }, + Successes: []ResponseDescriptor{ + { + StatusCode: http.StatusAccepted, + Headers: []ParameterDescriptor{ + { + Name: "Content-Length", + Type: "integer", + Description: "0", + Format: "0", + }, + digestHeader, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Name: "Invalid Name or Digest", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeDigestInvalid, + ErrorCodeNameInvalid, + }, + }, + { + Description: "The blob, identified by `name` and `digest`, is unknown to the registry.", + StatusCode: http.StatusNotFound, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeNameUnknown, + ErrorCodeBlobUnknown, + }, + }, + { + Description: "Blob delete is not allowed because the registry is configured as a pull-through cache or `delete` has been disabled", + StatusCode: http.StatusMethodNotAllowed, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + ErrorCodes: []errcode.ErrorCode{ + errcode.ErrorCodeUnsupported, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + }, + }, + + // TODO(stevvooe): We may want to add a PUT request here to + // kickoff an upload of a blob, integrated with the blob upload + // API. + }, + }, + + { + Name: RouteNameBlobUpload, + Path: "/v2/{name:" + reference.NameRegexp.String() + "}/blobs/uploads/", + Entity: "Initiate Blob Upload", + Description: "Initiate a blob upload. This endpoint can be used to create resumable uploads or monolithic uploads.", + Methods: []MethodDescriptor{ + { + Method: "POST", + Description: "Initiate a resumable blob upload. If successful, an upload location will be provided to complete the upload. Optionally, if the `digest` parameter is present, the request body will be used to complete the upload in a single request.", + Requests: []RequestDescriptor{ + { + Name: "Initiate Monolithic Blob Upload", + Description: "Upload a blob identified by the `digest` parameter in single request. This upload will not be resumable unless a recoverable error is returned.", + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + { + Name: "Content-Length", + Type: "integer", + Format: "", + }, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + }, + QueryParameters: []ParameterDescriptor{ + { + Name: "digest", + Type: "query", + Format: "", + Regexp: digest.DigestRegexp, + Description: `Digest of uploaded blob. If present, the upload will be completed, in a single request, with contents of the request body as the resulting blob.`, + }, + }, + Body: BodyDescriptor{ + ContentType: "application/octect-stream", + Format: "", + }, + Successes: []ResponseDescriptor{ + { + Description: "The blob has been created in the registry and is available at the provided location.", + StatusCode: http.StatusCreated, + Headers: []ParameterDescriptor{ + { + Name: "Location", + Type: "url", + Format: "", + }, + contentLengthZeroHeader, + dockerUploadUUIDHeader, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Name: "Invalid Name or Digest", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeDigestInvalid, + ErrorCodeNameInvalid, + }, + }, + { + Name: "Not allowed", + Description: "Blob upload is not allowed because the registry is configured as a pull-through cache or for some other reason", + StatusCode: http.StatusMethodNotAllowed, + ErrorCodes: []errcode.ErrorCode{ + errcode.ErrorCodeUnsupported, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + { + Name: "Initiate Resumable Blob Upload", + Description: "Initiate a resumable blob upload with an empty request body.", + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + contentLengthZeroHeader, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + }, + Successes: []ResponseDescriptor{ + { + Description: "The upload has been created. The `Location` header must be used to complete the upload. The response should be identical to a `GET` request on the contents of the returned `Location` header.", + StatusCode: http.StatusAccepted, + Headers: []ParameterDescriptor{ + contentLengthZeroHeader, + { + Name: "Location", + Type: "url", + Format: "/v2//blobs/uploads/", + Description: "The location of the created upload. Clients should use the contents verbatim to complete the upload, adding parameters where required.", + }, + { + Name: "Range", + Format: "0-0", + Description: "Range header indicating the progress of the upload. When starting an upload, it will return an empty range, since no content has been received.", + }, + dockerUploadUUIDHeader, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Name: "Invalid Name or Digest", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeDigestInvalid, + ErrorCodeNameInvalid, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + { + Name: "Mount Blob", + Description: "Mount a blob identified by the `mount` parameter from another repository.", + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + contentLengthZeroHeader, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + }, + QueryParameters: []ParameterDescriptor{ + { + Name: "mount", + Type: "query", + Format: "", + Regexp: digest.DigestRegexp, + Description: `Digest of blob to mount from the source repository.`, + }, + { + Name: "from", + Type: "query", + Format: "", + Regexp: reference.NameRegexp, + Description: `Name of the source repository.`, + }, + }, + Successes: []ResponseDescriptor{ + { + Description: "The blob has been mounted in the repository and is available at the provided location.", + StatusCode: http.StatusCreated, + Headers: []ParameterDescriptor{ + { + Name: "Location", + Type: "url", + Format: "", + }, + contentLengthZeroHeader, + dockerUploadUUIDHeader, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Name: "Invalid Name or Digest", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeDigestInvalid, + ErrorCodeNameInvalid, + }, + }, + { + Name: "Not allowed", + Description: "Blob mount is not allowed because the registry is configured as a pull-through cache or for some other reason", + StatusCode: http.StatusMethodNotAllowed, + ErrorCodes: []errcode.ErrorCode{ + errcode.ErrorCodeUnsupported, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + }, + }, + }, + }, + + { + Name: RouteNameBlobUploadChunk, + Path: "/v2/{name:" + reference.NameRegexp.String() + "}/blobs/uploads/{uuid:[a-zA-Z0-9-_.=]+}", + Entity: "Blob Upload", + Description: "Interact with blob uploads. Clients should never assemble URLs for this endpoint and should only take it through the `Location` header on related API requests. The `Location` header and its parameters should be preserved by clients, using the latest value returned via upload related API calls.", + Methods: []MethodDescriptor{ + { + Method: "GET", + Description: "Retrieve status of upload identified by `uuid`. The primary purpose of this endpoint is to resolve the current status of a resumable upload.", + Requests: []RequestDescriptor{ + { + Description: "Retrieve the progress of the current upload, as reported by the `Range` header.", + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + uuidParameterDescriptor, + }, + Successes: []ResponseDescriptor{ + { + Name: "Upload Progress", + Description: "The upload is known and in progress. The last received offset is available in the `Range` header.", + StatusCode: http.StatusNoContent, + Headers: []ParameterDescriptor{ + { + Name: "Range", + Type: "header", + Format: "0-", + Description: "Range indicating the current progress of the upload.", + }, + contentLengthZeroHeader, + dockerUploadUUIDHeader, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Description: "There was an error processing the upload and it must be restarted.", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeDigestInvalid, + ErrorCodeNameInvalid, + ErrorCodeBlobUploadInvalid, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + { + Description: "The upload is unknown to the registry. The upload must be restarted.", + StatusCode: http.StatusNotFound, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeBlobUploadUnknown, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + }, + }, + { + Method: "PATCH", + Description: "Upload a chunk of data for the specified upload.", + Requests: []RequestDescriptor{ + { + Name: "Stream upload", + Description: "Upload a stream of data to upload without completing the upload.", + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + uuidParameterDescriptor, + }, + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + }, + Body: BodyDescriptor{ + ContentType: "application/octet-stream", + Format: "", + }, + Successes: []ResponseDescriptor{ + { + Name: "Data Accepted", + Description: "The stream of data has been accepted and the current progress is available in the range header. The updated upload location is available in the `Location` header.", + StatusCode: http.StatusNoContent, + Headers: []ParameterDescriptor{ + { + Name: "Location", + Type: "url", + Format: "/v2//blobs/uploads/", + Description: "The location of the upload. Clients should assume this changes after each request. Clients should use the contents verbatim to complete the upload, adding parameters where required.", + }, + { + Name: "Range", + Type: "header", + Format: "0-", + Description: "Range indicating the current progress of the upload.", + }, + contentLengthZeroHeader, + dockerUploadUUIDHeader, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Description: "There was an error processing the upload and it must be restarted.", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeDigestInvalid, + ErrorCodeNameInvalid, + ErrorCodeBlobUploadInvalid, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + { + Description: "The upload is unknown to the registry. The upload must be restarted.", + StatusCode: http.StatusNotFound, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeBlobUploadUnknown, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + { + Name: "Chunked upload", + Description: "Upload a chunk of data to specified upload without completing the upload. The data will be uploaded to the specified Content Range.", + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + uuidParameterDescriptor, + }, + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + { + Name: "Content-Range", + Type: "header", + Format: "-", + Required: true, + Description: "Range of bytes identifying the desired block of content represented by the body. Start must the end offset retrieved via status check plus one. Note that this is a non-standard use of the `Content-Range` header.", + }, + { + Name: "Content-Length", + Type: "integer", + Format: "", + Description: "Length of the chunk being uploaded, corresponding the length of the request body.", + }, + }, + Body: BodyDescriptor{ + ContentType: "application/octet-stream", + Format: "", + }, + Successes: []ResponseDescriptor{ + { + Name: "Chunk Accepted", + Description: "The chunk of data has been accepted and the current progress is available in the range header. The updated upload location is available in the `Location` header.", + StatusCode: http.StatusNoContent, + Headers: []ParameterDescriptor{ + { + Name: "Location", + Type: "url", + Format: "/v2//blobs/uploads/", + Description: "The location of the upload. Clients should assume this changes after each request. Clients should use the contents verbatim to complete the upload, adding parameters where required.", + }, + { + Name: "Range", + Type: "header", + Format: "0-", + Description: "Range indicating the current progress of the upload.", + }, + contentLengthZeroHeader, + dockerUploadUUIDHeader, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Description: "There was an error processing the upload and it must be restarted.", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeDigestInvalid, + ErrorCodeNameInvalid, + ErrorCodeBlobUploadInvalid, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + { + Description: "The upload is unknown to the registry. The upload must be restarted.", + StatusCode: http.StatusNotFound, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeBlobUploadUnknown, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + { + Description: "The `Content-Range` specification cannot be accepted, either because it does not overlap with the current progress or it is invalid.", + StatusCode: http.StatusRequestedRangeNotSatisfiable, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + }, + }, + { + Method: "PUT", + Description: "Complete the upload specified by `uuid`, optionally appending the body as the final chunk.", + Requests: []RequestDescriptor{ + { + Description: "Complete the upload, providing all the data in the body, if necessary. A request without a body will just complete the upload with previously uploaded content.", + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + { + Name: "Content-Length", + Type: "integer", + Format: "", + Description: "Length of the data being uploaded, corresponding to the length of the request body. May be zero if no data is provided.", + }, + }, + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + uuidParameterDescriptor, + }, + QueryParameters: []ParameterDescriptor{ + { + Name: "digest", + Type: "string", + Format: "", + Regexp: digest.DigestRegexp, + Required: true, + Description: `Digest of uploaded blob.`, + }, + }, + Body: BodyDescriptor{ + ContentType: "application/octet-stream", + Format: "", + }, + Successes: []ResponseDescriptor{ + { + Name: "Upload Complete", + Description: "The upload has been completed and accepted by the registry. The canonical location will be available in the `Location` header.", + StatusCode: http.StatusNoContent, + Headers: []ParameterDescriptor{ + { + Name: "Location", + Type: "url", + Format: "", + Description: "The canonical location of the blob for retrieval", + }, + { + Name: "Content-Range", + Type: "header", + Format: "-", + Description: "Range of bytes identifying the desired block of content represented by the body. Start must match the end of offset retrieved via status check. Note that this is a non-standard use of the `Content-Range` header.", + }, + contentLengthZeroHeader, + digestHeader, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Description: "There was an error processing the upload and it must be restarted.", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeDigestInvalid, + ErrorCodeNameInvalid, + ErrorCodeBlobUploadInvalid, + errcode.ErrorCodeUnsupported, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + { + Description: "The upload is unknown to the registry. The upload must be restarted.", + StatusCode: http.StatusNotFound, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeBlobUploadUnknown, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + }, + }, + { + Method: "DELETE", + Description: "Cancel outstanding upload processes, releasing associated resources. If this is not called, the unfinished uploads will eventually timeout.", + Requests: []RequestDescriptor{ + { + Description: "Cancel the upload specified by `uuid`.", + PathParameters: []ParameterDescriptor{ + nameParameterDescriptor, + uuidParameterDescriptor, + }, + Headers: []ParameterDescriptor{ + hostHeader, + authHeader, + contentLengthZeroHeader, + }, + Successes: []ResponseDescriptor{ + { + Name: "Upload Deleted", + Description: "The upload has been successfully deleted.", + StatusCode: http.StatusNoContent, + Headers: []ParameterDescriptor{ + contentLengthZeroHeader, + }, + }, + }, + Failures: []ResponseDescriptor{ + { + Description: "An error was encountered processing the delete. The client may ignore this error.", + StatusCode: http.StatusBadRequest, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeNameInvalid, + ErrorCodeBlobUploadInvalid, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + { + Description: "The upload is unknown to the registry. The client may ignore this error and assume the upload has been deleted.", + StatusCode: http.StatusNotFound, + ErrorCodes: []errcode.ErrorCode{ + ErrorCodeBlobUploadUnknown, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: errorsBody, + }, + }, + unauthorizedResponseDescriptor, + repositoryNotFoundResponseDescriptor, + deniedResponseDescriptor, + tooManyRequestsDescriptor, + }, + }, + }, + }, + }, + }, + { + Name: RouteNameCatalog, + Path: "/v2/_catalog", + Entity: "Catalog", + Description: "List a set of available repositories in the local registry cluster. Does not provide any indication of what may be available upstream. Applications can only determine if a repository is available but not if it is not available.", + Methods: []MethodDescriptor{ + { + Method: "GET", + Description: "Retrieve a sorted, json list of repositories available in the registry.", + Requests: []RequestDescriptor{ + { + Name: "Catalog Fetch", + Description: "Request an unabridged list of repositories available. The implementation may impose a maximum limit and return a partial set with pagination links.", + Successes: []ResponseDescriptor{ + { + Description: "Returns the unabridged list of repositories as a json response.", + StatusCode: http.StatusOK, + Headers: []ParameterDescriptor{ + { + Name: "Content-Length", + Type: "integer", + Description: "Length of the JSON response body.", + Format: "", + }, + }, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: `{ + "repositories": [ + , + ... + ] +}`, + }, + }, + }, + }, + { + Name: "Catalog Fetch Paginated", + Description: "Return the specified portion of repositories.", + QueryParameters: paginationParameters, + Successes: []ResponseDescriptor{ + { + StatusCode: http.StatusOK, + Body: BodyDescriptor{ + ContentType: "application/json; charset=utf-8", + Format: `{ + "repositories": [ + , + ... + ] + "next": "?last=&n=" +}`, + }, + Headers: []ParameterDescriptor{ + { + Name: "Content-Length", + Type: "integer", + Description: "Length of the JSON response body.", + Format: "", + }, + linkHeader, + }, + }, + }, + }, + }, + }, + }, + }, +} + +var routeDescriptorsMap map[string]RouteDescriptor + +func init() { + routeDescriptorsMap = make(map[string]RouteDescriptor, len(routeDescriptors)) + + for _, descriptor := range routeDescriptors { + routeDescriptorsMap[descriptor.Name] = descriptor + } +} diff --git a/vendor/github.com/docker/distribution/registry/api/v2/doc.go b/vendor/github.com/docker/distribution/registry/api/v2/doc.go new file mode 100644 index 000000000..cde011959 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/api/v2/doc.go @@ -0,0 +1,9 @@ +// Package v2 describes routes, urls and the error codes used in the Docker +// Registry JSON HTTP API V2. In addition to declarations, descriptors are +// provided for routes and error codes that can be used for implementation and +// automatically generating documentation. +// +// Definitions here are considered to be locked down for the V2 registry api. +// Any changes must be considered carefully and should not proceed without a +// change proposal in docker core. +package v2 diff --git a/vendor/github.com/docker/distribution/registry/api/v2/errors.go b/vendor/github.com/docker/distribution/registry/api/v2/errors.go new file mode 100644 index 000000000..2883c51e1 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/api/v2/errors.go @@ -0,0 +1,136 @@ +package v2 + +import ( + "net/http" + + "github.com/docker/distribution/registry/api/errcode" +) + +const errGroup = "registry.api.v2" + +var ( + // ErrorCodeDigestInvalid is returned when uploading a blob if the + // provided digest does not match the blob contents. + ErrorCodeDigestInvalid = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "DIGEST_INVALID", + Message: "provided digest did not match uploaded content", + Description: `When a blob is uploaded, the registry will check that + the content matches the digest provided by the client. The error may + include a detail structure with the key "digest", including the + invalid digest string. This error may also be returned when a manifest + includes an invalid layer digest.`, + HTTPStatusCode: http.StatusBadRequest, + }) + + // ErrorCodeSizeInvalid is returned when uploading a blob if the provided + ErrorCodeSizeInvalid = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "SIZE_INVALID", + Message: "provided length did not match content length", + Description: `When a layer is uploaded, the provided size will be + checked against the uploaded content. If they do not match, this error + will be returned.`, + HTTPStatusCode: http.StatusBadRequest, + }) + + // ErrorCodeNameInvalid is returned when the name in the manifest does not + // match the provided name. + ErrorCodeNameInvalid = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "NAME_INVALID", + Message: "invalid repository name", + Description: `Invalid repository name encountered either during + manifest validation or any API operation.`, + HTTPStatusCode: http.StatusBadRequest, + }) + + // ErrorCodeTagInvalid is returned when the tag in the manifest does not + // match the provided tag. + ErrorCodeTagInvalid = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "TAG_INVALID", + Message: "manifest tag did not match URI", + Description: `During a manifest upload, if the tag in the manifest + does not match the uri tag, this error will be returned.`, + HTTPStatusCode: http.StatusBadRequest, + }) + + // ErrorCodeNameUnknown when the repository name is not known. + ErrorCodeNameUnknown = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "NAME_UNKNOWN", + Message: "repository name not known to registry", + Description: `This is returned if the name used during an operation is + unknown to the registry.`, + HTTPStatusCode: http.StatusNotFound, + }) + + // ErrorCodeManifestUnknown returned when image manifest is unknown. + ErrorCodeManifestUnknown = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "MANIFEST_UNKNOWN", + Message: "manifest unknown", + Description: `This error is returned when the manifest, identified by + name and tag is unknown to the repository.`, + HTTPStatusCode: http.StatusNotFound, + }) + + // ErrorCodeManifestInvalid returned when an image manifest is invalid, + // typically during a PUT operation. This error encompasses all errors + // encountered during manifest validation that aren't signature errors. + ErrorCodeManifestInvalid = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "MANIFEST_INVALID", + Message: "manifest invalid", + Description: `During upload, manifests undergo several checks ensuring + validity. If those checks fail, this error may be returned, unless a + more specific error is included. The detail will contain information + the failed validation.`, + HTTPStatusCode: http.StatusBadRequest, + }) + + // ErrorCodeManifestUnverified is returned when the manifest fails + // signature verification. + ErrorCodeManifestUnverified = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "MANIFEST_UNVERIFIED", + Message: "manifest failed signature verification", + Description: `During manifest upload, if the manifest fails signature + verification, this error will be returned.`, + HTTPStatusCode: http.StatusBadRequest, + }) + + // ErrorCodeManifestBlobUnknown is returned when a manifest blob is + // unknown to the registry. + ErrorCodeManifestBlobUnknown = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "MANIFEST_BLOB_UNKNOWN", + Message: "blob unknown to registry", + Description: `This error may be returned when a manifest blob is + unknown to the registry.`, + HTTPStatusCode: http.StatusBadRequest, + }) + + // ErrorCodeBlobUnknown is returned when a blob is unknown to the + // registry. This can happen when the manifest references a nonexistent + // layer or the result is not found by a blob fetch. + ErrorCodeBlobUnknown = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "BLOB_UNKNOWN", + Message: "blob unknown to registry", + Description: `This error may be returned when a blob is unknown to the + registry in a specified repository. This can be returned with a + standard get or if a manifest references an unknown layer during + upload.`, + HTTPStatusCode: http.StatusNotFound, + }) + + // ErrorCodeBlobUploadUnknown is returned when an upload is unknown. + ErrorCodeBlobUploadUnknown = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "BLOB_UPLOAD_UNKNOWN", + Message: "blob upload unknown to registry", + Description: `If a blob upload has been cancelled or was never + started, this error code may be returned.`, + HTTPStatusCode: http.StatusNotFound, + }) + + // ErrorCodeBlobUploadInvalid is returned when an upload is invalid. + ErrorCodeBlobUploadInvalid = errcode.Register(errGroup, errcode.ErrorDescriptor{ + Value: "BLOB_UPLOAD_INVALID", + Message: "blob upload invalid", + Description: `The blob upload encountered an error and can no + longer proceed.`, + HTTPStatusCode: http.StatusNotFound, + }) +) diff --git a/vendor/github.com/docker/distribution/registry/api/v2/headerparser.go b/vendor/github.com/docker/distribution/registry/api/v2/headerparser.go new file mode 100644 index 000000000..9bc41a3a6 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/api/v2/headerparser.go @@ -0,0 +1,161 @@ +package v2 + +import ( + "fmt" + "regexp" + "strings" + "unicode" +) + +var ( + // according to rfc7230 + reToken = regexp.MustCompile(`^[^"(),/:;<=>?@[\]{}[:space:][:cntrl:]]+`) + reQuotedValue = regexp.MustCompile(`^[^\\"]+`) + reEscapedCharacter = regexp.MustCompile(`^[[:blank:][:graph:]]`) +) + +// parseForwardedHeader is a benevolent parser of Forwarded header defined in rfc7239. The header contains +// a comma-separated list of forwarding key-value pairs. Each list element is set by single proxy. The +// function parses only the first element of the list, which is set by the very first proxy. It returns a map +// of corresponding key-value pairs and an unparsed slice of the input string. +// +// Examples of Forwarded header values: +// +// 1. Forwarded: For=192.0.2.43; Proto=https,For="[2001:db8:cafe::17]",For=unknown +// 2. Forwarded: for="192.0.2.43:443"; host="registry.example.org", for="10.10.05.40:80" +// +// The first will be parsed into {"for": "192.0.2.43", "proto": "https"} while the second into +// {"for": "192.0.2.43:443", "host": "registry.example.org"}. +func parseForwardedHeader(forwarded string) (map[string]string, string, error) { + // Following are states of forwarded header parser. Any state could transition to a failure. + const ( + // terminating state; can transition to Parameter + stateElement = iota + // terminating state; can transition to KeyValueDelimiter + stateParameter + // can transition to Value + stateKeyValueDelimiter + // can transition to one of { QuotedValue, PairEnd } + stateValue + // can transition to one of { EscapedCharacter, PairEnd } + stateQuotedValue + // can transition to one of { QuotedValue } + stateEscapedCharacter + // terminating state; can transition to one of { Parameter, Element } + statePairEnd + ) + + var ( + parameter string + value string + parse = forwarded[:] + res = map[string]string{} + state = stateElement + ) + +Loop: + for { + // skip spaces unless in quoted value + if state != stateQuotedValue && state != stateEscapedCharacter { + parse = strings.TrimLeftFunc(parse, unicode.IsSpace) + } + + if len(parse) == 0 { + if state != stateElement && state != statePairEnd && state != stateParameter { + return nil, parse, fmt.Errorf("unexpected end of input") + } + // terminating + break + } + + switch state { + // terminate at list element delimiter + case stateElement: + if parse[0] == ',' { + parse = parse[1:] + break Loop + } + state = stateParameter + + // parse parameter (the key of key-value pair) + case stateParameter: + match := reToken.FindString(parse) + if len(match) == 0 { + return nil, parse, fmt.Errorf("failed to parse token at position %d", len(forwarded)-len(parse)) + } + parameter = strings.ToLower(match) + parse = parse[len(match):] + state = stateKeyValueDelimiter + + // parse '=' + case stateKeyValueDelimiter: + if parse[0] != '=' { + return nil, parse, fmt.Errorf("expected '=', not '%c' at position %d", parse[0], len(forwarded)-len(parse)) + } + parse = parse[1:] + state = stateValue + + // parse value or quoted value + case stateValue: + if parse[0] == '"' { + parse = parse[1:] + state = stateQuotedValue + } else { + value = reToken.FindString(parse) + if len(value) == 0 { + return nil, parse, fmt.Errorf("failed to parse value at position %d", len(forwarded)-len(parse)) + } + if _, exists := res[parameter]; exists { + return nil, parse, fmt.Errorf("duplicate parameter %q at position %d", parameter, len(forwarded)-len(parse)) + } + res[parameter] = value + parse = parse[len(value):] + value = "" + state = statePairEnd + } + + // parse a part of quoted value until the first backslash + case stateQuotedValue: + match := reQuotedValue.FindString(parse) + value += match + parse = parse[len(match):] + switch { + case len(parse) == 0: + return nil, parse, fmt.Errorf("unterminated quoted string") + case parse[0] == '"': + res[parameter] = value + value = "" + parse = parse[1:] + state = statePairEnd + case parse[0] == '\\': + parse = parse[1:] + state = stateEscapedCharacter + } + + // parse escaped character in a quoted string, ignore the backslash + // transition back to QuotedValue state + case stateEscapedCharacter: + c := reEscapedCharacter.FindString(parse) + if len(c) == 0 { + return nil, parse, fmt.Errorf("invalid escape sequence at position %d", len(forwarded)-len(parse)-1) + } + value += c + parse = parse[1:] + state = stateQuotedValue + + // expect either a new key-value pair, new list or end of input + case statePairEnd: + switch parse[0] { + case ';': + parse = parse[1:] + state = stateParameter + case ',': + state = stateElement + default: + return nil, parse, fmt.Errorf("expected ',' or ';', not %c at position %d", parse[0], len(forwarded)-len(parse)) + } + } + } + + return res, parse, nil +} diff --git a/vendor/github.com/docker/distribution/registry/api/v2/routes.go b/vendor/github.com/docker/distribution/registry/api/v2/routes.go new file mode 100644 index 000000000..9612ac2e5 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/api/v2/routes.go @@ -0,0 +1,40 @@ +package v2 + +import "github.com/gorilla/mux" + +// The following are definitions of the name under which all V2 routes are +// registered. These symbols can be used to look up a route based on the name. +const ( + RouteNameBase = "base" + RouteNameManifest = "manifest" + RouteNameTags = "tags" + RouteNameBlob = "blob" + RouteNameBlobUpload = "blob-upload" + RouteNameBlobUploadChunk = "blob-upload-chunk" + RouteNameCatalog = "catalog" +) + +// Router builds a gorilla router with named routes for the various API +// methods. This can be used directly by both server implementations and +// clients. +func Router() *mux.Router { + return RouterWithPrefix("") +} + +// RouterWithPrefix builds a gorilla router with a configured prefix +// on all routes. +func RouterWithPrefix(prefix string) *mux.Router { + rootRouter := mux.NewRouter() + router := rootRouter + if prefix != "" { + router = router.PathPrefix(prefix).Subrouter() + } + + router.StrictSlash(true) + + for _, descriptor := range routeDescriptors { + router.Path(descriptor.Path).Name(descriptor.Name) + } + + return rootRouter +} diff --git a/vendor/github.com/docker/distribution/registry/api/v2/urls.go b/vendor/github.com/docker/distribution/registry/api/v2/urls.go new file mode 100644 index 000000000..1337bdb12 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/api/v2/urls.go @@ -0,0 +1,266 @@ +package v2 + +import ( + "fmt" + "net/http" + "net/url" + "strings" + + "github.com/docker/distribution/reference" + "github.com/gorilla/mux" +) + +// URLBuilder creates registry API urls from a single base endpoint. It can be +// used to create urls for use in a registry client or server. +// +// All urls will be created from the given base, including the api version. +// For example, if a root of "/foo/" is provided, urls generated will be fall +// under "/foo/v2/...". Most application will only provide a schema, host and +// port, such as "https://localhost:5000/". +type URLBuilder struct { + root *url.URL // url root (ie http://localhost/) + router *mux.Router + relative bool +} + +// NewURLBuilder creates a URLBuilder with provided root url object. +func NewURLBuilder(root *url.URL, relative bool) *URLBuilder { + return &URLBuilder{ + root: root, + router: Router(), + relative: relative, + } +} + +// NewURLBuilderFromString workes identically to NewURLBuilder except it takes +// a string argument for the root, returning an error if it is not a valid +// url. +func NewURLBuilderFromString(root string, relative bool) (*URLBuilder, error) { + u, err := url.Parse(root) + if err != nil { + return nil, err + } + + return NewURLBuilder(u, relative), nil +} + +// NewURLBuilderFromRequest uses information from an *http.Request to +// construct the root url. +func NewURLBuilderFromRequest(r *http.Request, relative bool) *URLBuilder { + var ( + scheme = "http" + host = r.Host + ) + + if r.TLS != nil { + scheme = "https" + } else if len(r.URL.Scheme) > 0 { + scheme = r.URL.Scheme + } + + // Handle fowarded headers + // Prefer "Forwarded" header as defined by rfc7239 if given + // see https://tools.ietf.org/html/rfc7239 + if forwarded := r.Header.Get("Forwarded"); len(forwarded) > 0 { + forwardedHeader, _, err := parseForwardedHeader(forwarded) + if err == nil { + if fproto := forwardedHeader["proto"]; len(fproto) > 0 { + scheme = fproto + } + if fhost := forwardedHeader["host"]; len(fhost) > 0 { + host = fhost + } + } + } else { + if forwardedProto := r.Header.Get("X-Forwarded-Proto"); len(forwardedProto) > 0 { + scheme = forwardedProto + } + if forwardedHost := r.Header.Get("X-Forwarded-Host"); len(forwardedHost) > 0 { + // According to the Apache mod_proxy docs, X-Forwarded-Host can be a + // comma-separated list of hosts, to which each proxy appends the + // requested host. We want to grab the first from this comma-separated + // list. + hosts := strings.SplitN(forwardedHost, ",", 2) + host = strings.TrimSpace(hosts[0]) + } + } + + basePath := routeDescriptorsMap[RouteNameBase].Path + + requestPath := r.URL.Path + index := strings.Index(requestPath, basePath) + + u := &url.URL{ + Scheme: scheme, + Host: host, + } + + if index > 0 { + // N.B. index+1 is important because we want to include the trailing / + u.Path = requestPath[0 : index+1] + } + + return NewURLBuilder(u, relative) +} + +// BuildBaseURL constructs a base url for the API, typically just "/v2/". +func (ub *URLBuilder) BuildBaseURL() (string, error) { + route := ub.cloneRoute(RouteNameBase) + + baseURL, err := route.URL() + if err != nil { + return "", err + } + + return baseURL.String(), nil +} + +// BuildCatalogURL constructs a url get a catalog of repositories +func (ub *URLBuilder) BuildCatalogURL(values ...url.Values) (string, error) { + route := ub.cloneRoute(RouteNameCatalog) + + catalogURL, err := route.URL() + if err != nil { + return "", err + } + + return appendValuesURL(catalogURL, values...).String(), nil +} + +// BuildTagsURL constructs a url to list the tags in the named repository. +func (ub *URLBuilder) BuildTagsURL(name reference.Named) (string, error) { + route := ub.cloneRoute(RouteNameTags) + + tagsURL, err := route.URL("name", name.Name()) + if err != nil { + return "", err + } + + return tagsURL.String(), nil +} + +// BuildManifestURL constructs a url for the manifest identified by name and +// reference. The argument reference may be either a tag or digest. +func (ub *URLBuilder) BuildManifestURL(ref reference.Named) (string, error) { + route := ub.cloneRoute(RouteNameManifest) + + tagOrDigest := "" + switch v := ref.(type) { + case reference.Tagged: + tagOrDigest = v.Tag() + case reference.Digested: + tagOrDigest = v.Digest().String() + default: + return "", fmt.Errorf("reference must have a tag or digest") + } + + manifestURL, err := route.URL("name", ref.Name(), "reference", tagOrDigest) + if err != nil { + return "", err + } + + return manifestURL.String(), nil +} + +// BuildBlobURL constructs the url for the blob identified by name and dgst. +func (ub *URLBuilder) BuildBlobURL(ref reference.Canonical) (string, error) { + route := ub.cloneRoute(RouteNameBlob) + + layerURL, err := route.URL("name", ref.Name(), "digest", ref.Digest().String()) + if err != nil { + return "", err + } + + return layerURL.String(), nil +} + +// BuildBlobUploadURL constructs a url to begin a blob upload in the +// repository identified by name. +func (ub *URLBuilder) BuildBlobUploadURL(name reference.Named, values ...url.Values) (string, error) { + route := ub.cloneRoute(RouteNameBlobUpload) + + uploadURL, err := route.URL("name", name.Name()) + if err != nil { + return "", err + } + + return appendValuesURL(uploadURL, values...).String(), nil +} + +// BuildBlobUploadChunkURL constructs a url for the upload identified by uuid, +// including any url values. This should generally not be used by clients, as +// this url is provided by server implementations during the blob upload +// process. +func (ub *URLBuilder) BuildBlobUploadChunkURL(name reference.Named, uuid string, values ...url.Values) (string, error) { + route := ub.cloneRoute(RouteNameBlobUploadChunk) + + uploadURL, err := route.URL("name", name.Name(), "uuid", uuid) + if err != nil { + return "", err + } + + return appendValuesURL(uploadURL, values...).String(), nil +} + +// clondedRoute returns a clone of the named route from the router. Routes +// must be cloned to avoid modifying them during url generation. +func (ub *URLBuilder) cloneRoute(name string) clonedRoute { + route := new(mux.Route) + root := new(url.URL) + + *route = *ub.router.GetRoute(name) // clone the route + *root = *ub.root + + return clonedRoute{Route: route, root: root, relative: ub.relative} +} + +type clonedRoute struct { + *mux.Route + root *url.URL + relative bool +} + +func (cr clonedRoute) URL(pairs ...string) (*url.URL, error) { + routeURL, err := cr.Route.URL(pairs...) + if err != nil { + return nil, err + } + + if cr.relative { + return routeURL, nil + } + + if routeURL.Scheme == "" && routeURL.User == nil && routeURL.Host == "" { + routeURL.Path = routeURL.Path[1:] + } + + url := cr.root.ResolveReference(routeURL) + url.Scheme = cr.root.Scheme + return url, nil +} + +// appendValuesURL appends the parameters to the url. +func appendValuesURL(u *url.URL, values ...url.Values) *url.URL { + merged := u.Query() + + for _, v := range values { + for k, vv := range v { + merged[k] = append(merged[k], vv...) + } + } + + u.RawQuery = merged.Encode() + return u +} + +// appendValues appends the parameters to the url. Panics if the string is not +// a url. +func appendValues(u string, values ...url.Values) string { + up, err := url.Parse(u) + + if err != nil { + panic(err) // should never happen + } + + return appendValuesURL(up, values...).String() +} diff --git a/vendor/github.com/docker/distribution/registry/client/auth/api_version.go b/vendor/github.com/docker/distribution/registry/client/auth/api_version.go new file mode 100644 index 000000000..7d8f1d957 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/client/auth/api_version.go @@ -0,0 +1,58 @@ +package auth + +import ( + "net/http" + "strings" +) + +// APIVersion represents a version of an API including its +// type and version number. +type APIVersion struct { + // Type refers to the name of a specific API specification + // such as "registry" + Type string + + // Version is the version of the API specification implemented, + // This may omit the revision number and only include + // the major and minor version, such as "2.0" + Version string +} + +// String returns the string formatted API Version +func (v APIVersion) String() string { + return v.Type + "/" + v.Version +} + +// APIVersions gets the API versions out of an HTTP response using the provided +// version header as the key for the HTTP header. +func APIVersions(resp *http.Response, versionHeader string) []APIVersion { + versions := []APIVersion{} + if versionHeader != "" { + for _, supportedVersions := range resp.Header[http.CanonicalHeaderKey(versionHeader)] { + for _, version := range strings.Fields(supportedVersions) { + versions = append(versions, ParseAPIVersion(version)) + } + } + } + return versions +} + +// ParseAPIVersion parses an API version string into an APIVersion +// Format (Expected, not enforced): +// API version string = '/' +// API type = [a-z][a-z0-9]* +// API version = [0-9]+(\.[0-9]+)? +// TODO(dmcgowan): Enforce format, add error condition, remove unknown type +func ParseAPIVersion(versionStr string) APIVersion { + idx := strings.IndexRune(versionStr, '/') + if idx == -1 { + return APIVersion{ + Type: "unknown", + Version: versionStr, + } + } + return APIVersion{ + Type: strings.ToLower(versionStr[:idx]), + Version: versionStr[idx+1:], + } +} diff --git a/vendor/github.com/docker/distribution/registry/client/auth/challenge/addr.go b/vendor/github.com/docker/distribution/registry/client/auth/challenge/addr.go new file mode 100644 index 000000000..2c3ebe165 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/client/auth/challenge/addr.go @@ -0,0 +1,27 @@ +package challenge + +import ( + "net/url" + "strings" +) + +// FROM: https://golang.org/src/net/http/http.go +// Given a string of the form "host", "host:port", or "[ipv6::address]:port", +// return true if the string includes a port. +func hasPort(s string) bool { return strings.LastIndex(s, ":") > strings.LastIndex(s, "]") } + +// FROM: http://golang.org/src/net/http/transport.go +var portMap = map[string]string{ + "http": "80", + "https": "443", +} + +// canonicalAddr returns url.Host but always with a ":port" suffix +// FROM: http://golang.org/src/net/http/transport.go +func canonicalAddr(url *url.URL) string { + addr := url.Host + if !hasPort(addr) { + return addr + ":" + portMap[url.Scheme] + } + return addr +} diff --git a/vendor/github.com/docker/distribution/registry/client/auth/challenge/authchallenge.go b/vendor/github.com/docker/distribution/registry/client/auth/challenge/authchallenge.go new file mode 100644 index 000000000..6e3f1ccc4 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/client/auth/challenge/authchallenge.go @@ -0,0 +1,237 @@ +package challenge + +import ( + "fmt" + "net/http" + "net/url" + "strings" + "sync" +) + +// Challenge carries information from a WWW-Authenticate response header. +// See RFC 2617. +type Challenge struct { + // Scheme is the auth-scheme according to RFC 2617 + Scheme string + + // Parameters are the auth-params according to RFC 2617 + Parameters map[string]string +} + +// Manager manages the challenges for endpoints. +// The challenges are pulled out of HTTP responses. Only +// responses which expect challenges should be added to +// the manager, since a non-unauthorized request will be +// viewed as not requiring challenges. +type Manager interface { + // GetChallenges returns the challenges for the given + // endpoint URL. + GetChallenges(endpoint url.URL) ([]Challenge, error) + + // AddResponse adds the response to the challenge + // manager. The challenges will be parsed out of + // the WWW-Authenicate headers and added to the + // URL which was produced the response. If the + // response was authorized, any challenges for the + // endpoint will be cleared. + AddResponse(resp *http.Response) error +} + +// NewSimpleManager returns an instance of +// Manger which only maps endpoints to challenges +// based on the responses which have been added the +// manager. The simple manager will make no attempt to +// perform requests on the endpoints or cache the responses +// to a backend. +func NewSimpleManager() Manager { + return &simpleManager{ + Challenges: make(map[string][]Challenge), + } +} + +type simpleManager struct { + sync.RWMutex + Challenges map[string][]Challenge +} + +func normalizeURL(endpoint *url.URL) { + endpoint.Host = strings.ToLower(endpoint.Host) + endpoint.Host = canonicalAddr(endpoint) +} + +func (m *simpleManager) GetChallenges(endpoint url.URL) ([]Challenge, error) { + normalizeURL(&endpoint) + + m.RLock() + defer m.RUnlock() + challenges := m.Challenges[endpoint.String()] + return challenges, nil +} + +func (m *simpleManager) AddResponse(resp *http.Response) error { + challenges := ResponseChallenges(resp) + if resp.Request == nil { + return fmt.Errorf("missing request reference") + } + urlCopy := url.URL{ + Path: resp.Request.URL.Path, + Host: resp.Request.URL.Host, + Scheme: resp.Request.URL.Scheme, + } + normalizeURL(&urlCopy) + + m.Lock() + defer m.Unlock() + m.Challenges[urlCopy.String()] = challenges + return nil +} + +// Octet types from RFC 2616. +type octetType byte + +var octetTypes [256]octetType + +const ( + isToken octetType = 1 << iota + isSpace +) + +func init() { + // OCTET = + // CHAR = + // CTL = + // CR = + // LF = + // SP = + // HT = + // <"> = + // CRLF = CR LF + // LWS = [CRLF] 1*( SP | HT ) + // TEXT = + // separators = "(" | ")" | "<" | ">" | "@" | "," | ";" | ":" | "\" | <"> + // | "/" | "[" | "]" | "?" | "=" | "{" | "}" | SP | HT + // token = 1* + // qdtext = > + + for c := 0; c < 256; c++ { + var t octetType + isCtl := c <= 31 || c == 127 + isChar := 0 <= c && c <= 127 + isSeparator := strings.IndexRune(" \t\"(),/:;<=>?@[]\\{}", rune(c)) >= 0 + if strings.IndexRune(" \t\r\n", rune(c)) >= 0 { + t |= isSpace + } + if isChar && !isCtl && !isSeparator { + t |= isToken + } + octetTypes[c] = t + } +} + +// ResponseChallenges returns a list of authorization challenges +// for the given http Response. Challenges are only checked if +// the response status code was a 401. +func ResponseChallenges(resp *http.Response) []Challenge { + if resp.StatusCode == http.StatusUnauthorized { + // Parse the WWW-Authenticate Header and store the challenges + // on this endpoint object. + return parseAuthHeader(resp.Header) + } + + return nil +} + +func parseAuthHeader(header http.Header) []Challenge { + challenges := []Challenge{} + for _, h := range header[http.CanonicalHeaderKey("WWW-Authenticate")] { + v, p := parseValueAndParams(h) + if v != "" { + challenges = append(challenges, Challenge{Scheme: v, Parameters: p}) + } + } + return challenges +} + +func parseValueAndParams(header string) (value string, params map[string]string) { + params = make(map[string]string) + value, s := expectToken(header) + if value == "" { + return + } + value = strings.ToLower(value) + s = "," + skipSpace(s) + for strings.HasPrefix(s, ",") { + var pkey string + pkey, s = expectToken(skipSpace(s[1:])) + if pkey == "" { + return + } + if !strings.HasPrefix(s, "=") { + return + } + var pvalue string + pvalue, s = expectTokenOrQuoted(s[1:]) + if pvalue == "" { + return + } + pkey = strings.ToLower(pkey) + params[pkey] = pvalue + s = skipSpace(s) + } + return +} + +func skipSpace(s string) (rest string) { + i := 0 + for ; i < len(s); i++ { + if octetTypes[s[i]]&isSpace == 0 { + break + } + } + return s[i:] +} + +func expectToken(s string) (token, rest string) { + i := 0 + for ; i < len(s); i++ { + if octetTypes[s[i]]&isToken == 0 { + break + } + } + return s[:i], s[i:] +} + +func expectTokenOrQuoted(s string) (value string, rest string) { + if !strings.HasPrefix(s, "\"") { + return expectToken(s) + } + s = s[1:] + for i := 0; i < len(s); i++ { + switch s[i] { + case '"': + return s[:i], s[i+1:] + case '\\': + p := make([]byte, len(s)-1) + j := copy(p, s[:i]) + escape := true + for i = i + 1; i < len(s); i++ { + b := s[i] + switch { + case escape: + escape = false + p[j] = b + j++ + case b == '\\': + escape = true + case b == '"': + return string(p[:j]), s[i+1:] + default: + p[j] = b + j++ + } + } + return "", "" + } + } + return "", "" +} diff --git a/vendor/github.com/docker/distribution/registry/client/auth/session.go b/vendor/github.com/docker/distribution/registry/client/auth/session.go new file mode 100644 index 000000000..db86c9b06 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/client/auth/session.go @@ -0,0 +1,532 @@ +package auth + +import ( + "encoding/json" + "errors" + "fmt" + "net/http" + "net/url" + "strings" + "sync" + "time" + + "github.com/docker/distribution/registry/client" + "github.com/docker/distribution/registry/client/auth/challenge" + "github.com/docker/distribution/registry/client/transport" +) + +var ( + // ErrNoBasicAuthCredentials is returned if a request can't be authorized with + // basic auth due to lack of credentials. + ErrNoBasicAuthCredentials = errors.New("no basic auth credentials") + + // ErrNoToken is returned if a request is successful but the body does not + // contain an authorization token. + ErrNoToken = errors.New("authorization server did not include a token in the response") +) + +const defaultClientID = "registry-client" + +// AuthenticationHandler is an interface for authorizing a request from +// params from a "WWW-Authenicate" header for a single scheme. +type AuthenticationHandler interface { + // Scheme returns the scheme as expected from the "WWW-Authenicate" header. + Scheme() string + + // AuthorizeRequest adds the authorization header to a request (if needed) + // using the parameters from "WWW-Authenticate" method. The parameters + // values depend on the scheme. + AuthorizeRequest(req *http.Request, params map[string]string) error +} + +// CredentialStore is an interface for getting credentials for +// a given URL +type CredentialStore interface { + // Basic returns basic auth for the given URL + Basic(*url.URL) (string, string) + + // RefreshToken returns a refresh token for the + // given URL and service + RefreshToken(*url.URL, string) string + + // SetRefreshToken sets the refresh token if none + // is provided for the given url and service + SetRefreshToken(realm *url.URL, service, token string) +} + +// NewAuthorizer creates an authorizer which can handle multiple authentication +// schemes. The handlers are tried in order, the higher priority authentication +// methods should be first. The challengeMap holds a list of challenges for +// a given root API endpoint (for example "https://registry-1.docker.io/v2/"). +func NewAuthorizer(manager challenge.Manager, handlers ...AuthenticationHandler) transport.RequestModifier { + return &endpointAuthorizer{ + challenges: manager, + handlers: handlers, + } +} + +type endpointAuthorizer struct { + challenges challenge.Manager + handlers []AuthenticationHandler + transport http.RoundTripper +} + +func (ea *endpointAuthorizer) ModifyRequest(req *http.Request) error { + pingPath := req.URL.Path + if v2Root := strings.Index(req.URL.Path, "/v2/"); v2Root != -1 { + pingPath = pingPath[:v2Root+4] + } else if v1Root := strings.Index(req.URL.Path, "/v1/"); v1Root != -1 { + pingPath = pingPath[:v1Root] + "/v2/" + } else { + return nil + } + + ping := url.URL{ + Host: req.URL.Host, + Scheme: req.URL.Scheme, + Path: pingPath, + } + + challenges, err := ea.challenges.GetChallenges(ping) + if err != nil { + return err + } + + if len(challenges) > 0 { + for _, handler := range ea.handlers { + for _, c := range challenges { + if c.Scheme != handler.Scheme() { + continue + } + if err := handler.AuthorizeRequest(req, c.Parameters); err != nil { + return err + } + } + } + } + + return nil +} + +// This is the minimum duration a token can last (in seconds). +// A token must not live less than 60 seconds because older versions +// of the Docker client didn't read their expiration from the token +// response and assumed 60 seconds. So to remain compatible with +// those implementations, a token must live at least this long. +const minimumTokenLifetimeSeconds = 60 + +// Private interface for time used by this package to enable tests to provide their own implementation. +type clock interface { + Now() time.Time +} + +type tokenHandler struct { + header http.Header + creds CredentialStore + transport http.RoundTripper + clock clock + + offlineAccess bool + forceOAuth bool + clientID string + scopes []Scope + + tokenLock sync.Mutex + tokenCache string + tokenExpiration time.Time + + logger Logger +} + +// Scope is a type which is serializable to a string +// using the allow scope grammar. +type Scope interface { + String() string +} + +// RepositoryScope represents a token scope for access +// to a repository. +type RepositoryScope struct { + Repository string + Class string + Actions []string +} + +// String returns the string representation of the repository +// using the scope grammar +func (rs RepositoryScope) String() string { + repoType := "repository" + // Keep existing format for image class to maintain backwards compatibility + // with authorization servers which do not support the expanded grammar. + if rs.Class != "" && rs.Class != "image" { + repoType = fmt.Sprintf("%s(%s)", repoType, rs.Class) + } + return fmt.Sprintf("%s:%s:%s", repoType, rs.Repository, strings.Join(rs.Actions, ",")) +} + +// RegistryScope represents a token scope for access +// to resources in the registry. +type RegistryScope struct { + Name string + Actions []string +} + +// String returns the string representation of the user +// using the scope grammar +func (rs RegistryScope) String() string { + return fmt.Sprintf("registry:%s:%s", rs.Name, strings.Join(rs.Actions, ",")) +} + +// Logger defines the injectable logging interface, used on TokenHandlers. +type Logger interface { + Debugf(format string, args ...interface{}) +} + +func logDebugf(logger Logger, format string, args ...interface{}) { + if logger == nil { + return + } + logger.Debugf(format, args...) +} + +// TokenHandlerOptions is used to configure a new token handler +type TokenHandlerOptions struct { + Transport http.RoundTripper + Credentials CredentialStore + + OfflineAccess bool + ForceOAuth bool + ClientID string + Scopes []Scope + Logger Logger +} + +// An implementation of clock for providing real time data. +type realClock struct{} + +// Now implements clock +func (realClock) Now() time.Time { return time.Now() } + +// NewTokenHandler creates a new AuthenicationHandler which supports +// fetching tokens from a remote token server. +func NewTokenHandler(transport http.RoundTripper, creds CredentialStore, scope string, actions ...string) AuthenticationHandler { + // Create options... + return NewTokenHandlerWithOptions(TokenHandlerOptions{ + Transport: transport, + Credentials: creds, + Scopes: []Scope{ + RepositoryScope{ + Repository: scope, + Actions: actions, + }, + }, + }) +} + +// NewTokenHandlerWithOptions creates a new token handler using the provided +// options structure. +func NewTokenHandlerWithOptions(options TokenHandlerOptions) AuthenticationHandler { + handler := &tokenHandler{ + transport: options.Transport, + creds: options.Credentials, + offlineAccess: options.OfflineAccess, + forceOAuth: options.ForceOAuth, + clientID: options.ClientID, + scopes: options.Scopes, + clock: realClock{}, + logger: options.Logger, + } + + return handler +} + +func (th *tokenHandler) client() *http.Client { + return &http.Client{ + Transport: th.transport, + Timeout: 15 * time.Second, + } +} + +func (th *tokenHandler) Scheme() string { + return "bearer" +} + +func (th *tokenHandler) AuthorizeRequest(req *http.Request, params map[string]string) error { + var additionalScopes []string + if fromParam := req.URL.Query().Get("from"); fromParam != "" { + additionalScopes = append(additionalScopes, RepositoryScope{ + Repository: fromParam, + Actions: []string{"pull"}, + }.String()) + } + + token, err := th.getToken(params, additionalScopes...) + if err != nil { + return err + } + + req.Header.Set("Authorization", fmt.Sprintf("Bearer %s", token)) + + return nil +} + +func (th *tokenHandler) getToken(params map[string]string, additionalScopes ...string) (string, error) { + th.tokenLock.Lock() + defer th.tokenLock.Unlock() + scopes := make([]string, 0, len(th.scopes)+len(additionalScopes)) + for _, scope := range th.scopes { + scopes = append(scopes, scope.String()) + } + var addedScopes bool + for _, scope := range additionalScopes { + if hasScope(scopes, scope) { + continue + } + scopes = append(scopes, scope) + addedScopes = true + } + + now := th.clock.Now() + if now.After(th.tokenExpiration) || addedScopes { + token, expiration, err := th.fetchToken(params, scopes) + if err != nil { + return "", err + } + + // do not update cache for added scope tokens + if !addedScopes { + th.tokenCache = token + th.tokenExpiration = expiration + } + + return token, nil + } + + return th.tokenCache, nil +} + +func hasScope(scopes []string, scope string) bool { + for _, s := range scopes { + if s == scope { + return true + } + } + return false +} + +type postTokenResponse struct { + AccessToken string `json:"access_token"` + RefreshToken string `json:"refresh_token"` + ExpiresIn int `json:"expires_in"` + IssuedAt time.Time `json:"issued_at"` + Scope string `json:"scope"` +} + +func (th *tokenHandler) fetchTokenWithOAuth(realm *url.URL, refreshToken, service string, scopes []string) (token string, expiration time.Time, err error) { + form := url.Values{} + form.Set("scope", strings.Join(scopes, " ")) + form.Set("service", service) + + clientID := th.clientID + if clientID == "" { + // Use default client, this is a required field + clientID = defaultClientID + } + form.Set("client_id", clientID) + + if refreshToken != "" { + form.Set("grant_type", "refresh_token") + form.Set("refresh_token", refreshToken) + } else if th.creds != nil { + form.Set("grant_type", "password") + username, password := th.creds.Basic(realm) + form.Set("username", username) + form.Set("password", password) + + // attempt to get a refresh token + form.Set("access_type", "offline") + } else { + // refuse to do oauth without a grant type + return "", time.Time{}, fmt.Errorf("no supported grant type") + } + + resp, err := th.client().PostForm(realm.String(), form) + if err != nil { + return "", time.Time{}, err + } + defer resp.Body.Close() + + if !client.SuccessStatus(resp.StatusCode) { + err := client.HandleErrorResponse(resp) + return "", time.Time{}, err + } + + decoder := json.NewDecoder(resp.Body) + + var tr postTokenResponse + if err = decoder.Decode(&tr); err != nil { + return "", time.Time{}, fmt.Errorf("unable to decode token response: %s", err) + } + + if tr.RefreshToken != "" && tr.RefreshToken != refreshToken { + th.creds.SetRefreshToken(realm, service, tr.RefreshToken) + } + + if tr.ExpiresIn < minimumTokenLifetimeSeconds { + // The default/minimum lifetime. + tr.ExpiresIn = minimumTokenLifetimeSeconds + logDebugf(th.logger, "Increasing token expiration to: %d seconds", tr.ExpiresIn) + } + + if tr.IssuedAt.IsZero() { + // issued_at is optional in the token response. + tr.IssuedAt = th.clock.Now().UTC() + } + + return tr.AccessToken, tr.IssuedAt.Add(time.Duration(tr.ExpiresIn) * time.Second), nil +} + +type getTokenResponse struct { + Token string `json:"token"` + AccessToken string `json:"access_token"` + ExpiresIn int `json:"expires_in"` + IssuedAt time.Time `json:"issued_at"` + RefreshToken string `json:"refresh_token"` +} + +func (th *tokenHandler) fetchTokenWithBasicAuth(realm *url.URL, service string, scopes []string) (token string, expiration time.Time, err error) { + + req, err := http.NewRequest("GET", realm.String(), nil) + if err != nil { + return "", time.Time{}, err + } + + reqParams := req.URL.Query() + + if service != "" { + reqParams.Add("service", service) + } + + for _, scope := range scopes { + reqParams.Add("scope", scope) + } + + if th.offlineAccess { + reqParams.Add("offline_token", "true") + clientID := th.clientID + if clientID == "" { + clientID = defaultClientID + } + reqParams.Add("client_id", clientID) + } + + if th.creds != nil { + username, password := th.creds.Basic(realm) + if username != "" && password != "" { + reqParams.Add("account", username) + req.SetBasicAuth(username, password) + } + } + + req.URL.RawQuery = reqParams.Encode() + + resp, err := th.client().Do(req) + if err != nil { + return "", time.Time{}, err + } + defer resp.Body.Close() + + if !client.SuccessStatus(resp.StatusCode) { + err := client.HandleErrorResponse(resp) + return "", time.Time{}, err + } + + decoder := json.NewDecoder(resp.Body) + + var tr getTokenResponse + if err = decoder.Decode(&tr); err != nil { + return "", time.Time{}, fmt.Errorf("unable to decode token response: %s", err) + } + + if tr.RefreshToken != "" && th.creds != nil { + th.creds.SetRefreshToken(realm, service, tr.RefreshToken) + } + + // `access_token` is equivalent to `token` and if both are specified + // the choice is undefined. Canonicalize `access_token` by sticking + // things in `token`. + if tr.AccessToken != "" { + tr.Token = tr.AccessToken + } + + if tr.Token == "" { + return "", time.Time{}, ErrNoToken + } + + if tr.ExpiresIn < minimumTokenLifetimeSeconds { + // The default/minimum lifetime. + tr.ExpiresIn = minimumTokenLifetimeSeconds + logDebugf(th.logger, "Increasing token expiration to: %d seconds", tr.ExpiresIn) + } + + if tr.IssuedAt.IsZero() { + // issued_at is optional in the token response. + tr.IssuedAt = th.clock.Now().UTC() + } + + return tr.Token, tr.IssuedAt.Add(time.Duration(tr.ExpiresIn) * time.Second), nil +} + +func (th *tokenHandler) fetchToken(params map[string]string, scopes []string) (token string, expiration time.Time, err error) { + realm, ok := params["realm"] + if !ok { + return "", time.Time{}, errors.New("no realm specified for token auth challenge") + } + + // TODO(dmcgowan): Handle empty scheme and relative realm + realmURL, err := url.Parse(realm) + if err != nil { + return "", time.Time{}, fmt.Errorf("invalid token auth challenge realm: %s", err) + } + + service := params["service"] + + var refreshToken string + + if th.creds != nil { + refreshToken = th.creds.RefreshToken(realmURL, service) + } + + if refreshToken != "" || th.forceOAuth { + return th.fetchTokenWithOAuth(realmURL, refreshToken, service, scopes) + } + + return th.fetchTokenWithBasicAuth(realmURL, service, scopes) +} + +type basicHandler struct { + creds CredentialStore +} + +// NewBasicHandler creaters a new authentiation handler which adds +// basic authentication credentials to a request. +func NewBasicHandler(creds CredentialStore) AuthenticationHandler { + return &basicHandler{ + creds: creds, + } +} + +func (*basicHandler) Scheme() string { + return "basic" +} + +func (bh *basicHandler) AuthorizeRequest(req *http.Request, params map[string]string) error { + if bh.creds != nil { + username, password := bh.creds.Basic(req.URL) + if username != "" && password != "" { + req.SetBasicAuth(username, password) + return nil + } + } + return ErrNoBasicAuthCredentials +} diff --git a/vendor/github.com/docker/distribution/registry/client/blob_writer.go b/vendor/github.com/docker/distribution/registry/client/blob_writer.go new file mode 100644 index 000000000..695bf852f --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/client/blob_writer.go @@ -0,0 +1,162 @@ +package client + +import ( + "bytes" + "context" + "fmt" + "io" + "io/ioutil" + "net/http" + "time" + + "github.com/docker/distribution" +) + +type httpBlobUpload struct { + statter distribution.BlobStatter + client *http.Client + + uuid string + startedAt time.Time + + location string // always the last value of the location header. + offset int64 + closed bool +} + +func (hbu *httpBlobUpload) Reader() (io.ReadCloser, error) { + panic("Not implemented") +} + +func (hbu *httpBlobUpload) handleErrorResponse(resp *http.Response) error { + if resp.StatusCode == http.StatusNotFound { + return distribution.ErrBlobUploadUnknown + } + return HandleErrorResponse(resp) +} + +func (hbu *httpBlobUpload) ReadFrom(r io.Reader) (n int64, err error) { + req, err := http.NewRequest("PATCH", hbu.location, ioutil.NopCloser(r)) + if err != nil { + return 0, err + } + defer req.Body.Close() + + resp, err := hbu.client.Do(req) + if err != nil { + return 0, err + } + + if !SuccessStatus(resp.StatusCode) { + return 0, hbu.handleErrorResponse(resp) + } + + hbu.uuid = resp.Header.Get("Docker-Upload-UUID") + hbu.location, err = sanitizeLocation(resp.Header.Get("Location"), hbu.location) + if err != nil { + return 0, err + } + rng := resp.Header.Get("Range") + var start, end int64 + if n, err := fmt.Sscanf(rng, "%d-%d", &start, &end); err != nil { + return 0, err + } else if n != 2 || end < start { + return 0, fmt.Errorf("bad range format: %s", rng) + } + + return (end - start + 1), nil + +} + +func (hbu *httpBlobUpload) Write(p []byte) (n int, err error) { + req, err := http.NewRequest("PATCH", hbu.location, bytes.NewReader(p)) + if err != nil { + return 0, err + } + req.Header.Set("Content-Range", fmt.Sprintf("%d-%d", hbu.offset, hbu.offset+int64(len(p)-1))) + req.Header.Set("Content-Length", fmt.Sprintf("%d", len(p))) + req.Header.Set("Content-Type", "application/octet-stream") + + resp, err := hbu.client.Do(req) + if err != nil { + return 0, err + } + + if !SuccessStatus(resp.StatusCode) { + return 0, hbu.handleErrorResponse(resp) + } + + hbu.uuid = resp.Header.Get("Docker-Upload-UUID") + hbu.location, err = sanitizeLocation(resp.Header.Get("Location"), hbu.location) + if err != nil { + return 0, err + } + rng := resp.Header.Get("Range") + var start, end int + if n, err := fmt.Sscanf(rng, "%d-%d", &start, &end); err != nil { + return 0, err + } else if n != 2 || end < start { + return 0, fmt.Errorf("bad range format: %s", rng) + } + + return (end - start + 1), nil + +} + +func (hbu *httpBlobUpload) Size() int64 { + return hbu.offset +} + +func (hbu *httpBlobUpload) ID() string { + return hbu.uuid +} + +func (hbu *httpBlobUpload) StartedAt() time.Time { + return hbu.startedAt +} + +func (hbu *httpBlobUpload) Commit(ctx context.Context, desc distribution.Descriptor) (distribution.Descriptor, error) { + // TODO(dmcgowan): Check if already finished, if so just fetch + req, err := http.NewRequest("PUT", hbu.location, nil) + if err != nil { + return distribution.Descriptor{}, err + } + + values := req.URL.Query() + values.Set("digest", desc.Digest.String()) + req.URL.RawQuery = values.Encode() + + resp, err := hbu.client.Do(req) + if err != nil { + return distribution.Descriptor{}, err + } + defer resp.Body.Close() + + if !SuccessStatus(resp.StatusCode) { + return distribution.Descriptor{}, hbu.handleErrorResponse(resp) + } + + return hbu.statter.Stat(ctx, desc.Digest) +} + +func (hbu *httpBlobUpload) Cancel(ctx context.Context) error { + req, err := http.NewRequest("DELETE", hbu.location, nil) + if err != nil { + return err + } + resp, err := hbu.client.Do(req) + if err != nil { + return err + } + defer resp.Body.Close() + + if resp.StatusCode == http.StatusNotFound || SuccessStatus(resp.StatusCode) { + return nil + } + return hbu.handleErrorResponse(resp) +} + +func (hbu *httpBlobUpload) Close() error { + hbu.closed = true + return nil +} diff --git a/vendor/github.com/docker/distribution/registry/client/errors.go b/vendor/github.com/docker/distribution/registry/client/errors.go new file mode 100644 index 000000000..52d49d5d2 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/client/errors.go @@ -0,0 +1,139 @@ +package client + +import ( + "encoding/json" + "errors" + "fmt" + "io" + "io/ioutil" + "net/http" + + "github.com/docker/distribution/registry/api/errcode" + "github.com/docker/distribution/registry/client/auth/challenge" +) + +// ErrNoErrorsInBody is returned when an HTTP response body parses to an empty +// errcode.Errors slice. +var ErrNoErrorsInBody = errors.New("no error details found in HTTP response body") + +// UnexpectedHTTPStatusError is returned when an unexpected HTTP status is +// returned when making a registry api call. +type UnexpectedHTTPStatusError struct { + Status string +} + +func (e *UnexpectedHTTPStatusError) Error() string { + return fmt.Sprintf("received unexpected HTTP status: %s", e.Status) +} + +// UnexpectedHTTPResponseError is returned when an expected HTTP status code +// is returned, but the content was unexpected and failed to be parsed. +type UnexpectedHTTPResponseError struct { + ParseErr error + StatusCode int + Response []byte +} + +func (e *UnexpectedHTTPResponseError) Error() string { + return fmt.Sprintf("error parsing HTTP %d response body: %s: %q", e.StatusCode, e.ParseErr.Error(), string(e.Response)) +} + +func parseHTTPErrorResponse(statusCode int, r io.Reader) error { + var errors errcode.Errors + body, err := ioutil.ReadAll(r) + if err != nil { + return err + } + + // For backward compatibility, handle irregularly formatted + // messages that contain a "details" field. + var detailsErr struct { + Details string `json:"details"` + } + err = json.Unmarshal(body, &detailsErr) + if err == nil && detailsErr.Details != "" { + switch statusCode { + case http.StatusUnauthorized: + return errcode.ErrorCodeUnauthorized.WithMessage(detailsErr.Details) + case http.StatusTooManyRequests: + return errcode.ErrorCodeTooManyRequests.WithMessage(detailsErr.Details) + default: + return errcode.ErrorCodeUnknown.WithMessage(detailsErr.Details) + } + } + + if err := json.Unmarshal(body, &errors); err != nil { + return &UnexpectedHTTPResponseError{ + ParseErr: err, + StatusCode: statusCode, + Response: body, + } + } + + if len(errors) == 0 { + // If there was no error specified in the body, return + // UnexpectedHTTPResponseError. + return &UnexpectedHTTPResponseError{ + ParseErr: ErrNoErrorsInBody, + StatusCode: statusCode, + Response: body, + } + } + + return errors +} + +func makeErrorList(err error) []error { + if errL, ok := err.(errcode.Errors); ok { + return []error(errL) + } + return []error{err} +} + +func mergeErrors(err1, err2 error) error { + return errcode.Errors(append(makeErrorList(err1), makeErrorList(err2)...)) +} + +// HandleErrorResponse returns error parsed from HTTP response for an +// unsuccessful HTTP response code (in the range 400 - 499 inclusive). An +// UnexpectedHTTPStatusError returned for response code outside of expected +// range. +func HandleErrorResponse(resp *http.Response) error { + if resp.StatusCode >= 400 && resp.StatusCode < 500 { + // Check for OAuth errors within the `WWW-Authenticate` header first + // See https://tools.ietf.org/html/rfc6750#section-3 + for _, c := range challenge.ResponseChallenges(resp) { + if c.Scheme == "bearer" { + var err errcode.Error + // codes defined at https://tools.ietf.org/html/rfc6750#section-3.1 + switch c.Parameters["error"] { + case "invalid_token": + err.Code = errcode.ErrorCodeUnauthorized + case "insufficient_scope": + err.Code = errcode.ErrorCodeDenied + default: + continue + } + if description := c.Parameters["error_description"]; description != "" { + err.Message = description + } else { + err.Message = err.Code.Message() + } + + return mergeErrors(err, parseHTTPErrorResponse(resp.StatusCode, resp.Body)) + } + } + err := parseHTTPErrorResponse(resp.StatusCode, resp.Body) + if uErr, ok := err.(*UnexpectedHTTPResponseError); ok && resp.StatusCode == 401 { + return errcode.ErrorCodeUnauthorized.WithDetail(uErr.Response) + } + return err + } + return &UnexpectedHTTPStatusError{Status: resp.Status} +} + +// SuccessStatus returns true if the argument is a successful HTTP response +// code (in the range 200 - 399 inclusive). +func SuccessStatus(status int) bool { + return status >= 200 && status <= 399 +} diff --git a/vendor/github.com/docker/distribution/registry/client/repository.go b/vendor/github.com/docker/distribution/registry/client/repository.go new file mode 100644 index 000000000..aa442e654 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/client/repository.go @@ -0,0 +1,867 @@ +package client + +import ( + "bytes" + "context" + "encoding/json" + "errors" + "fmt" + "io" + "io/ioutil" + "net/http" + "net/url" + "strconv" + "strings" + "time" + + "github.com/docker/distribution" + "github.com/docker/distribution/reference" + "github.com/docker/distribution/registry/api/v2" + "github.com/docker/distribution/registry/client/transport" + "github.com/docker/distribution/registry/storage/cache" + "github.com/docker/distribution/registry/storage/cache/memory" + "github.com/opencontainers/go-digest" +) + +// Registry provides an interface for calling Repositories, which returns a catalog of repositories. +type Registry interface { + Repositories(ctx context.Context, repos []string, last string) (n int, err error) +} + +// checkHTTPRedirect is a callback that can manipulate redirected HTTP +// requests. It is used to preserve Accept and Range headers. +func checkHTTPRedirect(req *http.Request, via []*http.Request) error { + if len(via) >= 10 { + return errors.New("stopped after 10 redirects") + } + + if len(via) > 0 { + for headerName, headerVals := range via[0].Header { + if headerName != "Accept" && headerName != "Range" { + continue + } + for _, val := range headerVals { + // Don't add to redirected request if redirected + // request already has a header with the same + // name and value. + hasValue := false + for _, existingVal := range req.Header[headerName] { + if existingVal == val { + hasValue = true + break + } + } + if !hasValue { + req.Header.Add(headerName, val) + } + } + } + } + + return nil +} + +// NewRegistry creates a registry namespace which can be used to get a listing of repositories +func NewRegistry(baseURL string, transport http.RoundTripper) (Registry, error) { + ub, err := v2.NewURLBuilderFromString(baseURL, false) + if err != nil { + return nil, err + } + + client := &http.Client{ + Transport: transport, + Timeout: 1 * time.Minute, + CheckRedirect: checkHTTPRedirect, + } + + return ®istry{ + client: client, + ub: ub, + }, nil +} + +type registry struct { + client *http.Client + ub *v2.URLBuilder +} + +// Repositories returns a lexigraphically sorted catalog given a base URL. The 'entries' slice will be filled up to the size +// of the slice, starting at the value provided in 'last'. The number of entries will be returned along with io.EOF if there +// are no more entries +func (r *registry) Repositories(ctx context.Context, entries []string, last string) (int, error) { + var numFilled int + var returnErr error + + values := buildCatalogValues(len(entries), last) + u, err := r.ub.BuildCatalogURL(values) + if err != nil { + return 0, err + } + + resp, err := r.client.Get(u) + if err != nil { + return 0, err + } + defer resp.Body.Close() + + if SuccessStatus(resp.StatusCode) { + var ctlg struct { + Repositories []string `json:"repositories"` + } + decoder := json.NewDecoder(resp.Body) + + if err := decoder.Decode(&ctlg); err != nil { + return 0, err + } + + for cnt := range ctlg.Repositories { + entries[cnt] = ctlg.Repositories[cnt] + } + numFilled = len(ctlg.Repositories) + + link := resp.Header.Get("Link") + if link == "" { + returnErr = io.EOF + } + } else { + return 0, HandleErrorResponse(resp) + } + + return numFilled, returnErr +} + +// NewRepository creates a new Repository for the given repository name and base URL. +func NewRepository(name reference.Named, baseURL string, transport http.RoundTripper) (distribution.Repository, error) { + ub, err := v2.NewURLBuilderFromString(baseURL, false) + if err != nil { + return nil, err + } + + client := &http.Client{ + Transport: transport, + CheckRedirect: checkHTTPRedirect, + // TODO(dmcgowan): create cookie jar + } + + return &repository{ + client: client, + ub: ub, + name: name, + }, nil +} + +type repository struct { + client *http.Client + ub *v2.URLBuilder + name reference.Named +} + +func (r *repository) Named() reference.Named { + return r.name +} + +func (r *repository) Blobs(ctx context.Context) distribution.BlobStore { + statter := &blobStatter{ + name: r.name, + ub: r.ub, + client: r.client, + } + return &blobs{ + name: r.name, + ub: r.ub, + client: r.client, + statter: cache.NewCachedBlobStatter(memory.NewInMemoryBlobDescriptorCacheProvider(), statter), + } +} + +func (r *repository) Manifests(ctx context.Context, options ...distribution.ManifestServiceOption) (distribution.ManifestService, error) { + // todo(richardscothern): options should be sent over the wire + return &manifests{ + name: r.name, + ub: r.ub, + client: r.client, + etags: make(map[string]string), + }, nil +} + +func (r *repository) Tags(ctx context.Context) distribution.TagService { + return &tags{ + client: r.client, + ub: r.ub, + name: r.Named(), + } +} + +// tags implements remote tagging operations. +type tags struct { + client *http.Client + ub *v2.URLBuilder + name reference.Named +} + +// All returns all tags +func (t *tags) All(ctx context.Context) ([]string, error) { + var tags []string + + listURLStr, err := t.ub.BuildTagsURL(t.name) + if err != nil { + return tags, err + } + + listURL, err := url.Parse(listURLStr) + if err != nil { + return tags, err + } + + for { + resp, err := t.client.Get(listURL.String()) + if err != nil { + return tags, err + } + defer resp.Body.Close() + + if SuccessStatus(resp.StatusCode) { + b, err := ioutil.ReadAll(resp.Body) + if err != nil { + return tags, err + } + + tagsResponse := struct { + Tags []string `json:"tags"` + }{} + if err := json.Unmarshal(b, &tagsResponse); err != nil { + return tags, err + } + tags = append(tags, tagsResponse.Tags...) + if link := resp.Header.Get("Link"); link != "" { + linkURLStr := strings.Trim(strings.Split(link, ";")[0], "<>") + linkURL, err := url.Parse(linkURLStr) + if err != nil { + return tags, err + } + + listURL = listURL.ResolveReference(linkURL) + } else { + return tags, nil + } + } else { + return tags, HandleErrorResponse(resp) + } + } +} + +func descriptorFromResponse(response *http.Response) (distribution.Descriptor, error) { + desc := distribution.Descriptor{} + headers := response.Header + + ctHeader := headers.Get("Content-Type") + if ctHeader == "" { + return distribution.Descriptor{}, errors.New("missing or empty Content-Type header") + } + desc.MediaType = ctHeader + + digestHeader := headers.Get("Docker-Content-Digest") + if digestHeader == "" { + bytes, err := ioutil.ReadAll(response.Body) + if err != nil { + return distribution.Descriptor{}, err + } + _, desc, err := distribution.UnmarshalManifest(ctHeader, bytes) + if err != nil { + return distribution.Descriptor{}, err + } + return desc, nil + } + + dgst, err := digest.Parse(digestHeader) + if err != nil { + return distribution.Descriptor{}, err + } + desc.Digest = dgst + + lengthHeader := headers.Get("Content-Length") + if lengthHeader == "" { + return distribution.Descriptor{}, errors.New("missing or empty Content-Length header") + } + length, err := strconv.ParseInt(lengthHeader, 10, 64) + if err != nil { + return distribution.Descriptor{}, err + } + desc.Size = length + + return desc, nil + +} + +// Get issues a HEAD request for a Manifest against its named endpoint in order +// to construct a descriptor for the tag. If the registry doesn't support HEADing +// a manifest, fallback to GET. +func (t *tags) Get(ctx context.Context, tag string) (distribution.Descriptor, error) { + ref, err := reference.WithTag(t.name, tag) + if err != nil { + return distribution.Descriptor{}, err + } + u, err := t.ub.BuildManifestURL(ref) + if err != nil { + return distribution.Descriptor{}, err + } + + newRequest := func(method string) (*http.Response, error) { + req, err := http.NewRequest(method, u, nil) + if err != nil { + return nil, err + } + + for _, t := range distribution.ManifestMediaTypes() { + req.Header.Add("Accept", t) + } + resp, err := t.client.Do(req) + return resp, err + } + + resp, err := newRequest("HEAD") + if err != nil { + return distribution.Descriptor{}, err + } + defer resp.Body.Close() + + switch { + case resp.StatusCode >= 200 && resp.StatusCode < 400 && len(resp.Header.Get("Docker-Content-Digest")) > 0: + // if the response is a success AND a Docker-Content-Digest can be retrieved from the headers + return descriptorFromResponse(resp) + default: + // if the response is an error - there will be no body to decode. + // Issue a GET request: + // - for data from a server that does not handle HEAD + // - to get error details in case of a failure + resp, err = newRequest("GET") + if err != nil { + return distribution.Descriptor{}, err + } + defer resp.Body.Close() + + if resp.StatusCode >= 200 && resp.StatusCode < 400 { + return descriptorFromResponse(resp) + } + return distribution.Descriptor{}, HandleErrorResponse(resp) + } +} + +func (t *tags) Lookup(ctx context.Context, digest distribution.Descriptor) ([]string, error) { + panic("not implemented") +} + +func (t *tags) Tag(ctx context.Context, tag string, desc distribution.Descriptor) error { + panic("not implemented") +} + +func (t *tags) Untag(ctx context.Context, tag string) error { + panic("not implemented") +} + +type manifests struct { + name reference.Named + ub *v2.URLBuilder + client *http.Client + etags map[string]string +} + +func (ms *manifests) Exists(ctx context.Context, dgst digest.Digest) (bool, error) { + ref, err := reference.WithDigest(ms.name, dgst) + if err != nil { + return false, err + } + u, err := ms.ub.BuildManifestURL(ref) + if err != nil { + return false, err + } + + resp, err := ms.client.Head(u) + if err != nil { + return false, err + } + + if SuccessStatus(resp.StatusCode) { + return true, nil + } else if resp.StatusCode == http.StatusNotFound { + return false, nil + } + return false, HandleErrorResponse(resp) +} + +// AddEtagToTag allows a client to supply an eTag to Get which will be +// used for a conditional HTTP request. If the eTag matches, a nil manifest +// and ErrManifestNotModified error will be returned. etag is automatically +// quoted when added to this map. +func AddEtagToTag(tag, etag string) distribution.ManifestServiceOption { + return etagOption{tag, etag} +} + +type etagOption struct{ tag, etag string } + +func (o etagOption) Apply(ms distribution.ManifestService) error { + if ms, ok := ms.(*manifests); ok { + ms.etags[o.tag] = fmt.Sprintf(`"%s"`, o.etag) + return nil + } + return fmt.Errorf("etag options is a client-only option") +} + +// ReturnContentDigest allows a client to set a the content digest on +// a successful request from the 'Docker-Content-Digest' header. This +// returned digest is represents the digest which the registry uses +// to refer to the content and can be used to delete the content. +func ReturnContentDigest(dgst *digest.Digest) distribution.ManifestServiceOption { + return contentDigestOption{dgst} +} + +type contentDigestOption struct{ digest *digest.Digest } + +func (o contentDigestOption) Apply(ms distribution.ManifestService) error { + return nil +} + +func (ms *manifests) Get(ctx context.Context, dgst digest.Digest, options ...distribution.ManifestServiceOption) (distribution.Manifest, error) { + var ( + digestOrTag string + ref reference.Named + err error + contentDgst *digest.Digest + mediaTypes []string + ) + + for _, option := range options { + switch opt := option.(type) { + case distribution.WithTagOption: + digestOrTag = opt.Tag + ref, err = reference.WithTag(ms.name, opt.Tag) + if err != nil { + return nil, err + } + case contentDigestOption: + contentDgst = opt.digest + case distribution.WithManifestMediaTypesOption: + mediaTypes = opt.MediaTypes + default: + err := option.Apply(ms) + if err != nil { + return nil, err + } + } + } + + if digestOrTag == "" { + digestOrTag = dgst.String() + ref, err = reference.WithDigest(ms.name, dgst) + if err != nil { + return nil, err + } + } + + if len(mediaTypes) == 0 { + mediaTypes = distribution.ManifestMediaTypes() + } + + u, err := ms.ub.BuildManifestURL(ref) + if err != nil { + return nil, err + } + + req, err := http.NewRequest("GET", u, nil) + if err != nil { + return nil, err + } + + for _, t := range mediaTypes { + req.Header.Add("Accept", t) + } + + if _, ok := ms.etags[digestOrTag]; ok { + req.Header.Set("If-None-Match", ms.etags[digestOrTag]) + } + + resp, err := ms.client.Do(req) + if err != nil { + return nil, err + } + defer resp.Body.Close() + if resp.StatusCode == http.StatusNotModified { + return nil, distribution.ErrManifestNotModified + } else if SuccessStatus(resp.StatusCode) { + if contentDgst != nil { + dgst, err := digest.Parse(resp.Header.Get("Docker-Content-Digest")) + if err == nil { + *contentDgst = dgst + } + } + mt := resp.Header.Get("Content-Type") + body, err := ioutil.ReadAll(resp.Body) + + if err != nil { + return nil, err + } + m, _, err := distribution.UnmarshalManifest(mt, body) + if err != nil { + return nil, err + } + return m, nil + } + return nil, HandleErrorResponse(resp) +} + +// Put puts a manifest. A tag can be specified using an options parameter which uses some shared state to hold the +// tag name in order to build the correct upload URL. +func (ms *manifests) Put(ctx context.Context, m distribution.Manifest, options ...distribution.ManifestServiceOption) (digest.Digest, error) { + ref := ms.name + var tagged bool + + for _, option := range options { + if opt, ok := option.(distribution.WithTagOption); ok { + var err error + ref, err = reference.WithTag(ref, opt.Tag) + if err != nil { + return "", err + } + tagged = true + } else { + err := option.Apply(ms) + if err != nil { + return "", err + } + } + } + mediaType, p, err := m.Payload() + if err != nil { + return "", err + } + + if !tagged { + // generate a canonical digest and Put by digest + _, d, err := distribution.UnmarshalManifest(mediaType, p) + if err != nil { + return "", err + } + ref, err = reference.WithDigest(ref, d.Digest) + if err != nil { + return "", err + } + } + + manifestURL, err := ms.ub.BuildManifestURL(ref) + if err != nil { + return "", err + } + + putRequest, err := http.NewRequest("PUT", manifestURL, bytes.NewReader(p)) + if err != nil { + return "", err + } + + putRequest.Header.Set("Content-Type", mediaType) + + resp, err := ms.client.Do(putRequest) + if err != nil { + return "", err + } + defer resp.Body.Close() + + if SuccessStatus(resp.StatusCode) { + dgstHeader := resp.Header.Get("Docker-Content-Digest") + dgst, err := digest.Parse(dgstHeader) + if err != nil { + return "", err + } + + return dgst, nil + } + + return "", HandleErrorResponse(resp) +} + +func (ms *manifests) Delete(ctx context.Context, dgst digest.Digest) error { + ref, err := reference.WithDigest(ms.name, dgst) + if err != nil { + return err + } + u, err := ms.ub.BuildManifestURL(ref) + if err != nil { + return err + } + req, err := http.NewRequest("DELETE", u, nil) + if err != nil { + return err + } + + resp, err := ms.client.Do(req) + if err != nil { + return err + } + defer resp.Body.Close() + + if SuccessStatus(resp.StatusCode) { + return nil + } + return HandleErrorResponse(resp) +} + +// todo(richardscothern): Restore interface and implementation with merge of #1050 +/*func (ms *manifests) Enumerate(ctx context.Context, manifests []distribution.Manifest, last distribution.Manifest) (n int, err error) { + panic("not supported") +}*/ + +type blobs struct { + name reference.Named + ub *v2.URLBuilder + client *http.Client + + statter distribution.BlobDescriptorService + distribution.BlobDeleter +} + +func sanitizeLocation(location, base string) (string, error) { + baseURL, err := url.Parse(base) + if err != nil { + return "", err + } + + locationURL, err := url.Parse(location) + if err != nil { + return "", err + } + + return baseURL.ResolveReference(locationURL).String(), nil +} + +func (bs *blobs) Stat(ctx context.Context, dgst digest.Digest) (distribution.Descriptor, error) { + return bs.statter.Stat(ctx, dgst) + +} + +func (bs *blobs) Get(ctx context.Context, dgst digest.Digest) ([]byte, error) { + reader, err := bs.Open(ctx, dgst) + if err != nil { + return nil, err + } + defer reader.Close() + + return ioutil.ReadAll(reader) +} + +func (bs *blobs) Open(ctx context.Context, dgst digest.Digest) (distribution.ReadSeekCloser, error) { + ref, err := reference.WithDigest(bs.name, dgst) + if err != nil { + return nil, err + } + blobURL, err := bs.ub.BuildBlobURL(ref) + if err != nil { + return nil, err + } + + return transport.NewHTTPReadSeeker(bs.client, blobURL, + func(resp *http.Response) error { + if resp.StatusCode == http.StatusNotFound { + return distribution.ErrBlobUnknown + } + return HandleErrorResponse(resp) + }), nil +} + +func (bs *blobs) ServeBlob(ctx context.Context, w http.ResponseWriter, r *http.Request, dgst digest.Digest) error { + panic("not implemented") +} + +func (bs *blobs) Put(ctx context.Context, mediaType string, p []byte) (distribution.Descriptor, error) { + writer, err := bs.Create(ctx) + if err != nil { + return distribution.Descriptor{}, err + } + dgstr := digest.Canonical.Digester() + n, err := io.Copy(writer, io.TeeReader(bytes.NewReader(p), dgstr.Hash())) + if err != nil { + return distribution.Descriptor{}, err + } + if n < int64(len(p)) { + return distribution.Descriptor{}, fmt.Errorf("short copy: wrote %d of %d", n, len(p)) + } + + desc := distribution.Descriptor{ + MediaType: mediaType, + Size: int64(len(p)), + Digest: dgstr.Digest(), + } + + return writer.Commit(ctx, desc) +} + +type optionFunc func(interface{}) error + +func (f optionFunc) Apply(v interface{}) error { + return f(v) +} + +// WithMountFrom returns a BlobCreateOption which designates that the blob should be +// mounted from the given canonical reference. +func WithMountFrom(ref reference.Canonical) distribution.BlobCreateOption { + return optionFunc(func(v interface{}) error { + opts, ok := v.(*distribution.CreateOptions) + if !ok { + return fmt.Errorf("unexpected options type: %T", v) + } + + opts.Mount.ShouldMount = true + opts.Mount.From = ref + + return nil + }) +} + +func (bs *blobs) Create(ctx context.Context, options ...distribution.BlobCreateOption) (distribution.BlobWriter, error) { + var opts distribution.CreateOptions + + for _, option := range options { + err := option.Apply(&opts) + if err != nil { + return nil, err + } + } + + var values []url.Values + + if opts.Mount.ShouldMount { + values = append(values, url.Values{"from": {opts.Mount.From.Name()}, "mount": {opts.Mount.From.Digest().String()}}) + } + + u, err := bs.ub.BuildBlobUploadURL(bs.name, values...) + if err != nil { + return nil, err + } + + resp, err := bs.client.Post(u, "", nil) + if err != nil { + return nil, err + } + defer resp.Body.Close() + + switch resp.StatusCode { + case http.StatusCreated: + desc, err := bs.statter.Stat(ctx, opts.Mount.From.Digest()) + if err != nil { + return nil, err + } + return nil, distribution.ErrBlobMounted{From: opts.Mount.From, Descriptor: desc} + case http.StatusAccepted: + // TODO(dmcgowan): Check for invalid UUID + uuid := resp.Header.Get("Docker-Upload-UUID") + location, err := sanitizeLocation(resp.Header.Get("Location"), u) + if err != nil { + return nil, err + } + + return &httpBlobUpload{ + statter: bs.statter, + client: bs.client, + uuid: uuid, + startedAt: time.Now(), + location: location, + }, nil + default: + return nil, HandleErrorResponse(resp) + } +} + +func (bs *blobs) Resume(ctx context.Context, id string) (distribution.BlobWriter, error) { + panic("not implemented") +} + +func (bs *blobs) Delete(ctx context.Context, dgst digest.Digest) error { + return bs.statter.Clear(ctx, dgst) +} + +type blobStatter struct { + name reference.Named + ub *v2.URLBuilder + client *http.Client +} + +func (bs *blobStatter) Stat(ctx context.Context, dgst digest.Digest) (distribution.Descriptor, error) { + ref, err := reference.WithDigest(bs.name, dgst) + if err != nil { + return distribution.Descriptor{}, err + } + u, err := bs.ub.BuildBlobURL(ref) + if err != nil { + return distribution.Descriptor{}, err + } + + resp, err := bs.client.Head(u) + if err != nil { + return distribution.Descriptor{}, err + } + defer resp.Body.Close() + + if SuccessStatus(resp.StatusCode) { + lengthHeader := resp.Header.Get("Content-Length") + if lengthHeader == "" { + return distribution.Descriptor{}, fmt.Errorf("missing content-length header for request: %s", u) + } + + length, err := strconv.ParseInt(lengthHeader, 10, 64) + if err != nil { + return distribution.Descriptor{}, fmt.Errorf("error parsing content-length: %v", err) + } + + return distribution.Descriptor{ + MediaType: resp.Header.Get("Content-Type"), + Size: length, + Digest: dgst, + }, nil + } else if resp.StatusCode == http.StatusNotFound { + return distribution.Descriptor{}, distribution.ErrBlobUnknown + } + return distribution.Descriptor{}, HandleErrorResponse(resp) +} + +func buildCatalogValues(maxEntries int, last string) url.Values { + values := url.Values{} + + if maxEntries > 0 { + values.Add("n", strconv.Itoa(maxEntries)) + } + + if last != "" { + values.Add("last", last) + } + + return values +} + +func (bs *blobStatter) Clear(ctx context.Context, dgst digest.Digest) error { + ref, err := reference.WithDigest(bs.name, dgst) + if err != nil { + return err + } + blobURL, err := bs.ub.BuildBlobURL(ref) + if err != nil { + return err + } + + req, err := http.NewRequest("DELETE", blobURL, nil) + if err != nil { + return err + } + + resp, err := bs.client.Do(req) + if err != nil { + return err + } + defer resp.Body.Close() + + if SuccessStatus(resp.StatusCode) { + return nil + } + return HandleErrorResponse(resp) +} + +func (bs *blobStatter) SetDescriptor(ctx context.Context, dgst digest.Digest, desc distribution.Descriptor) error { + return nil +} diff --git a/vendor/github.com/docker/distribution/registry/client/transport/http_reader.go b/vendor/github.com/docker/distribution/registry/client/transport/http_reader.go new file mode 100644 index 000000000..e5ff09d75 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/client/transport/http_reader.go @@ -0,0 +1,251 @@ +package transport + +import ( + "errors" + "fmt" + "io" + "net/http" + "os" + "regexp" + "strconv" +) + +var ( + contentRangeRegexp = regexp.MustCompile(`bytes ([0-9]+)-([0-9]+)/([0-9]+|\\*)`) + + // ErrWrongCodeForByteRange is returned if the client sends a request + // with a Range header but the server returns a 2xx or 3xx code other + // than 206 Partial Content. + ErrWrongCodeForByteRange = errors.New("expected HTTP 206 from byte range request") +) + +// ReadSeekCloser combines io.ReadSeeker with io.Closer. +type ReadSeekCloser interface { + io.ReadSeeker + io.Closer +} + +// NewHTTPReadSeeker handles reading from an HTTP endpoint using a GET +// request. When seeking and starting a read from a non-zero offset +// the a "Range" header will be added which sets the offset. +// TODO(dmcgowan): Move this into a separate utility package +func NewHTTPReadSeeker(client *http.Client, url string, errorHandler func(*http.Response) error) ReadSeekCloser { + return &httpReadSeeker{ + client: client, + url: url, + errorHandler: errorHandler, + } +} + +type httpReadSeeker struct { + client *http.Client + url string + + // errorHandler creates an error from an unsuccessful HTTP response. + // This allows the error to be created with the HTTP response body + // without leaking the body through a returned error. + errorHandler func(*http.Response) error + + size int64 + + // rc is the remote read closer. + rc io.ReadCloser + // readerOffset tracks the offset as of the last read. + readerOffset int64 + // seekOffset allows Seek to override the offset. Seek changes + // seekOffset instead of changing readOffset directly so that + // connection resets can be delayed and possibly avoided if the + // seek is undone (i.e. seeking to the end and then back to the + // beginning). + seekOffset int64 + err error +} + +func (hrs *httpReadSeeker) Read(p []byte) (n int, err error) { + if hrs.err != nil { + return 0, hrs.err + } + + // If we sought to a different position, we need to reset the + // connection. This logic is here instead of Seek so that if + // a seek is undone before the next read, the connection doesn't + // need to be closed and reopened. A common example of this is + // seeking to the end to determine the length, and then seeking + // back to the original position. + if hrs.readerOffset != hrs.seekOffset { + hrs.reset() + } + + hrs.readerOffset = hrs.seekOffset + + rd, err := hrs.reader() + if err != nil { + return 0, err + } + + n, err = rd.Read(p) + hrs.seekOffset += int64(n) + hrs.readerOffset += int64(n) + + return n, err +} + +func (hrs *httpReadSeeker) Seek(offset int64, whence int) (int64, error) { + if hrs.err != nil { + return 0, hrs.err + } + + lastReaderOffset := hrs.readerOffset + + if whence == os.SEEK_SET && hrs.rc == nil { + // If no request has been made yet, and we are seeking to an + // absolute position, set the read offset as well to avoid an + // unnecessary request. + hrs.readerOffset = offset + } + + _, err := hrs.reader() + if err != nil { + hrs.readerOffset = lastReaderOffset + return 0, err + } + + newOffset := hrs.seekOffset + + switch whence { + case os.SEEK_CUR: + newOffset += offset + case os.SEEK_END: + if hrs.size < 0 { + return 0, errors.New("content length not known") + } + newOffset = hrs.size + offset + case os.SEEK_SET: + newOffset = offset + } + + if newOffset < 0 { + err = errors.New("cannot seek to negative position") + } else { + hrs.seekOffset = newOffset + } + + return hrs.seekOffset, err +} + +func (hrs *httpReadSeeker) Close() error { + if hrs.err != nil { + return hrs.err + } + + // close and release reader chain + if hrs.rc != nil { + hrs.rc.Close() + } + + hrs.rc = nil + + hrs.err = errors.New("httpLayer: closed") + + return nil +} + +func (hrs *httpReadSeeker) reset() { + if hrs.err != nil { + return + } + if hrs.rc != nil { + hrs.rc.Close() + hrs.rc = nil + } +} + +func (hrs *httpReadSeeker) reader() (io.Reader, error) { + if hrs.err != nil { + return nil, hrs.err + } + + if hrs.rc != nil { + return hrs.rc, nil + } + + req, err := http.NewRequest("GET", hrs.url, nil) + if err != nil { + return nil, err + } + + if hrs.readerOffset > 0 { + // If we are at different offset, issue a range request from there. + req.Header.Add("Range", fmt.Sprintf("bytes=%d-", hrs.readerOffset)) + // TODO: get context in here + // context.GetLogger(hrs.context).Infof("Range: %s", req.Header.Get("Range")) + } + + req.Header.Add("Accept-Encoding", "identity") + resp, err := hrs.client.Do(req) + if err != nil { + return nil, err + } + + // Normally would use client.SuccessStatus, but that would be a cyclic + // import + if resp.StatusCode >= 200 && resp.StatusCode <= 399 { + if hrs.readerOffset > 0 { + if resp.StatusCode != http.StatusPartialContent { + return nil, ErrWrongCodeForByteRange + } + + contentRange := resp.Header.Get("Content-Range") + if contentRange == "" { + return nil, errors.New("no Content-Range header found in HTTP 206 response") + } + + submatches := contentRangeRegexp.FindStringSubmatch(contentRange) + if len(submatches) < 4 { + return nil, fmt.Errorf("could not parse Content-Range header: %s", contentRange) + } + + startByte, err := strconv.ParseUint(submatches[1], 10, 64) + if err != nil { + return nil, fmt.Errorf("could not parse start of range in Content-Range header: %s", contentRange) + } + + if startByte != uint64(hrs.readerOffset) { + return nil, fmt.Errorf("received Content-Range starting at offset %d instead of requested %d", startByte, hrs.readerOffset) + } + + endByte, err := strconv.ParseUint(submatches[2], 10, 64) + if err != nil { + return nil, fmt.Errorf("could not parse end of range in Content-Range header: %s", contentRange) + } + + if submatches[3] == "*" { + hrs.size = -1 + } else { + size, err := strconv.ParseUint(submatches[3], 10, 64) + if err != nil { + return nil, fmt.Errorf("could not parse total size in Content-Range header: %s", contentRange) + } + + if endByte+1 != size { + return nil, fmt.Errorf("range in Content-Range stops before the end of the content: %s", contentRange) + } + + hrs.size = int64(size) + } + } else if resp.StatusCode == http.StatusOK { + hrs.size = resp.ContentLength + } else { + hrs.size = -1 + } + hrs.rc = resp.Body + } else { + defer resp.Body.Close() + if hrs.errorHandler != nil { + return nil, hrs.errorHandler(resp) + } + return nil, fmt.Errorf("unexpected status resolving reader: %v", resp.Status) + } + + return hrs.rc, nil +} diff --git a/vendor/github.com/docker/distribution/registry/client/transport/transport.go b/vendor/github.com/docker/distribution/registry/client/transport/transport.go new file mode 100644 index 000000000..30e45fab0 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/client/transport/transport.go @@ -0,0 +1,147 @@ +package transport + +import ( + "io" + "net/http" + "sync" +) + +// RequestModifier represents an object which will do an inplace +// modification of an HTTP request. +type RequestModifier interface { + ModifyRequest(*http.Request) error +} + +type headerModifier http.Header + +// NewHeaderRequestModifier returns a new RequestModifier which will +// add the given headers to a request. +func NewHeaderRequestModifier(header http.Header) RequestModifier { + return headerModifier(header) +} + +func (h headerModifier) ModifyRequest(req *http.Request) error { + for k, s := range http.Header(h) { + req.Header[k] = append(req.Header[k], s...) + } + + return nil +} + +// NewTransport creates a new transport which will apply modifiers to +// the request on a RoundTrip call. +func NewTransport(base http.RoundTripper, modifiers ...RequestModifier) http.RoundTripper { + return &transport{ + Modifiers: modifiers, + Base: base, + } +} + +// transport is an http.RoundTripper that makes HTTP requests after +// copying and modifying the request +type transport struct { + Modifiers []RequestModifier + Base http.RoundTripper + + mu sync.Mutex // guards modReq + modReq map[*http.Request]*http.Request // original -> modified +} + +// RoundTrip authorizes and authenticates the request with an +// access token. If no token exists or token is expired, +// tries to refresh/fetch a new token. +func (t *transport) RoundTrip(req *http.Request) (*http.Response, error) { + req2 := cloneRequest(req) + for _, modifier := range t.Modifiers { + if err := modifier.ModifyRequest(req2); err != nil { + return nil, err + } + } + + t.setModReq(req, req2) + res, err := t.base().RoundTrip(req2) + if err != nil { + t.setModReq(req, nil) + return nil, err + } + res.Body = &onEOFReader{ + rc: res.Body, + fn: func() { t.setModReq(req, nil) }, + } + return res, nil +} + +// CancelRequest cancels an in-flight request by closing its connection. +func (t *transport) CancelRequest(req *http.Request) { + type canceler interface { + CancelRequest(*http.Request) + } + if cr, ok := t.base().(canceler); ok { + t.mu.Lock() + modReq := t.modReq[req] + delete(t.modReq, req) + t.mu.Unlock() + cr.CancelRequest(modReq) + } +} + +func (t *transport) base() http.RoundTripper { + if t.Base != nil { + return t.Base + } + return http.DefaultTransport +} + +func (t *transport) setModReq(orig, mod *http.Request) { + t.mu.Lock() + defer t.mu.Unlock() + if t.modReq == nil { + t.modReq = make(map[*http.Request]*http.Request) + } + if mod == nil { + delete(t.modReq, orig) + } else { + t.modReq[orig] = mod + } +} + +// cloneRequest returns a clone of the provided *http.Request. +// The clone is a shallow copy of the struct and its Header map. +func cloneRequest(r *http.Request) *http.Request { + // shallow copy of the struct + r2 := new(http.Request) + *r2 = *r + // deep copy of the Header + r2.Header = make(http.Header, len(r.Header)) + for k, s := range r.Header { + r2.Header[k] = append([]string(nil), s...) + } + + return r2 +} + +type onEOFReader struct { + rc io.ReadCloser + fn func() +} + +func (r *onEOFReader) Read(p []byte) (n int, err error) { + n, err = r.rc.Read(p) + if err == io.EOF { + r.runFunc() + } + return +} + +func (r *onEOFReader) Close() error { + err := r.rc.Close() + r.runFunc() + return err +} + +func (r *onEOFReader) runFunc() { + if fn := r.fn; fn != nil { + fn() + r.fn = nil + } +} diff --git a/vendor/github.com/docker/distribution/registry/storage/cache/cache.go b/vendor/github.com/docker/distribution/registry/storage/cache/cache.go new file mode 100644 index 000000000..10a390919 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/storage/cache/cache.go @@ -0,0 +1,35 @@ +// Package cache provides facilities to speed up access to the storage +// backend. +package cache + +import ( + "fmt" + + "github.com/docker/distribution" +) + +// BlobDescriptorCacheProvider provides repository scoped +// BlobDescriptorService cache instances and a global descriptor cache. +type BlobDescriptorCacheProvider interface { + distribution.BlobDescriptorService + + RepositoryScoped(repo string) (distribution.BlobDescriptorService, error) +} + +// ValidateDescriptor provides a helper function to ensure that caches have +// common criteria for admitting descriptors. +func ValidateDescriptor(desc distribution.Descriptor) error { + if err := desc.Digest.Validate(); err != nil { + return err + } + + if desc.Size < 0 { + return fmt.Errorf("cache: invalid length in descriptor: %v < 0", desc.Size) + } + + if desc.MediaType == "" { + return fmt.Errorf("cache: empty mediatype on descriptor: %v", desc) + } + + return nil +} diff --git a/vendor/github.com/docker/distribution/registry/storage/cache/cachedblobdescriptorstore.go b/vendor/github.com/docker/distribution/registry/storage/cache/cachedblobdescriptorstore.go new file mode 100644 index 000000000..ac4c45211 --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/storage/cache/cachedblobdescriptorstore.go @@ -0,0 +1,129 @@ +package cache + +import ( + "context" + + "github.com/docker/distribution" + prometheus "github.com/docker/distribution/metrics" + "github.com/opencontainers/go-digest" +) + +// Metrics is used to hold metric counters +// related to the number of times a cache was +// hit or missed. +type Metrics struct { + Requests uint64 + Hits uint64 + Misses uint64 +} + +// Logger can be provided on the MetricsTracker to log errors. +// +// Usually, this is just a proxy to dcontext.GetLogger. +type Logger interface { + Errorf(format string, args ...interface{}) +} + +// MetricsTracker represents a metric tracker +// which simply counts the number of hits and misses. +type MetricsTracker interface { + Hit() + Miss() + Metrics() Metrics + Logger(context.Context) Logger +} + +type cachedBlobStatter struct { + cache distribution.BlobDescriptorService + backend distribution.BlobDescriptorService + tracker MetricsTracker +} + +var ( + // cacheCount is the number of total cache request received/hits/misses + cacheCount = prometheus.StorageNamespace.NewLabeledCounter("cache", "The number of cache request received", "type") +) + +// NewCachedBlobStatter creates a new statter which prefers a cache and +// falls back to a backend. +func NewCachedBlobStatter(cache distribution.BlobDescriptorService, backend distribution.BlobDescriptorService) distribution.BlobDescriptorService { + return &cachedBlobStatter{ + cache: cache, + backend: backend, + } +} + +// NewCachedBlobStatterWithMetrics creates a new statter which prefers a cache and +// falls back to a backend. Hits and misses will send to the tracker. +func NewCachedBlobStatterWithMetrics(cache distribution.BlobDescriptorService, backend distribution.BlobDescriptorService, tracker MetricsTracker) distribution.BlobStatter { + return &cachedBlobStatter{ + cache: cache, + backend: backend, + tracker: tracker, + } +} + +func (cbds *cachedBlobStatter) Stat(ctx context.Context, dgst digest.Digest) (distribution.Descriptor, error) { + cacheCount.WithValues("Request").Inc(1) + desc, err := cbds.cache.Stat(ctx, dgst) + if err != nil { + if err != distribution.ErrBlobUnknown { + logErrorf(ctx, cbds.tracker, "error retrieving descriptor from cache: %v", err) + } + + goto fallback + } + cacheCount.WithValues("Hit").Inc(1) + if cbds.tracker != nil { + cbds.tracker.Hit() + } + return desc, nil +fallback: + cacheCount.WithValues("Miss").Inc(1) + if cbds.tracker != nil { + cbds.tracker.Miss() + } + desc, err = cbds.backend.Stat(ctx, dgst) + if err != nil { + return desc, err + } + + if err := cbds.cache.SetDescriptor(ctx, dgst, desc); err != nil { + logErrorf(ctx, cbds.tracker, "error adding descriptor %v to cache: %v", desc.Digest, err) + } + + return desc, err + +} + +func (cbds *cachedBlobStatter) Clear(ctx context.Context, dgst digest.Digest) error { + err := cbds.cache.Clear(ctx, dgst) + if err != nil { + return err + } + + err = cbds.backend.Clear(ctx, dgst) + if err != nil { + return err + } + return nil +} + +func (cbds *cachedBlobStatter) SetDescriptor(ctx context.Context, dgst digest.Digest, desc distribution.Descriptor) error { + if err := cbds.cache.SetDescriptor(ctx, dgst, desc); err != nil { + logErrorf(ctx, cbds.tracker, "error adding descriptor %v to cache: %v", desc.Digest, err) + } + return nil +} + +func logErrorf(ctx context.Context, tracker MetricsTracker, format string, args ...interface{}) { + if tracker == nil { + return + } + + logger := tracker.Logger(ctx) + if logger == nil { + return + } + logger.Errorf(format, args...) +} diff --git a/vendor/github.com/docker/distribution/registry/storage/cache/memory/memory.go b/vendor/github.com/docker/distribution/registry/storage/cache/memory/memory.go new file mode 100644 index 000000000..42d94d9bd --- /dev/null +++ b/vendor/github.com/docker/distribution/registry/storage/cache/memory/memory.go @@ -0,0 +1,179 @@ +package memory + +import ( + "context" + "sync" + + "github.com/docker/distribution" + "github.com/docker/distribution/reference" + "github.com/docker/distribution/registry/storage/cache" + "github.com/opencontainers/go-digest" +) + +type inMemoryBlobDescriptorCacheProvider struct { + global *mapBlobDescriptorCache + repositories map[string]*mapBlobDescriptorCache + mu sync.RWMutex +} + +// NewInMemoryBlobDescriptorCacheProvider returns a new mapped-based cache for +// storing blob descriptor data. +func NewInMemoryBlobDescriptorCacheProvider() cache.BlobDescriptorCacheProvider { + return &inMemoryBlobDescriptorCacheProvider{ + global: newMapBlobDescriptorCache(), + repositories: make(map[string]*mapBlobDescriptorCache), + } +} + +func (imbdcp *inMemoryBlobDescriptorCacheProvider) RepositoryScoped(repo string) (distribution.BlobDescriptorService, error) { + if _, err := reference.ParseNormalizedNamed(repo); err != nil { + return nil, err + } + + imbdcp.mu.RLock() + defer imbdcp.mu.RUnlock() + + return &repositoryScopedInMemoryBlobDescriptorCache{ + repo: repo, + parent: imbdcp, + repository: imbdcp.repositories[repo], + }, nil +} + +func (imbdcp *inMemoryBlobDescriptorCacheProvider) Stat(ctx context.Context, dgst digest.Digest) (distribution.Descriptor, error) { + return imbdcp.global.Stat(ctx, dgst) +} + +func (imbdcp *inMemoryBlobDescriptorCacheProvider) Clear(ctx context.Context, dgst digest.Digest) error { + return imbdcp.global.Clear(ctx, dgst) +} + +func (imbdcp *inMemoryBlobDescriptorCacheProvider) SetDescriptor(ctx context.Context, dgst digest.Digest, desc distribution.Descriptor) error { + _, err := imbdcp.Stat(ctx, dgst) + if err == distribution.ErrBlobUnknown { + + if dgst.Algorithm() != desc.Digest.Algorithm() && dgst != desc.Digest { + // if the digests differ, set the other canonical mapping + if err := imbdcp.global.SetDescriptor(ctx, desc.Digest, desc); err != nil { + return err + } + } + + // unknown, just set it + return imbdcp.global.SetDescriptor(ctx, dgst, desc) + } + + // we already know it, do nothing + return err +} + +// repositoryScopedInMemoryBlobDescriptorCache provides the request scoped +// repository cache. Instances are not thread-safe but the delegated +// operations are. +type repositoryScopedInMemoryBlobDescriptorCache struct { + repo string + parent *inMemoryBlobDescriptorCacheProvider // allows lazy allocation of repo's map + repository *mapBlobDescriptorCache +} + +func (rsimbdcp *repositoryScopedInMemoryBlobDescriptorCache) Stat(ctx context.Context, dgst digest.Digest) (distribution.Descriptor, error) { + rsimbdcp.parent.mu.Lock() + repo := rsimbdcp.repository + rsimbdcp.parent.mu.Unlock() + + if repo == nil { + return distribution.Descriptor{}, distribution.ErrBlobUnknown + } + + return repo.Stat(ctx, dgst) +} + +func (rsimbdcp *repositoryScopedInMemoryBlobDescriptorCache) Clear(ctx context.Context, dgst digest.Digest) error { + rsimbdcp.parent.mu.Lock() + repo := rsimbdcp.repository + rsimbdcp.parent.mu.Unlock() + + if repo == nil { + return distribution.ErrBlobUnknown + } + + return repo.Clear(ctx, dgst) +} + +func (rsimbdcp *repositoryScopedInMemoryBlobDescriptorCache) SetDescriptor(ctx context.Context, dgst digest.Digest, desc distribution.Descriptor) error { + rsimbdcp.parent.mu.Lock() + repo := rsimbdcp.repository + if repo == nil { + // allocate map since we are setting it now. + var ok bool + // have to read back value since we may have allocated elsewhere. + repo, ok = rsimbdcp.parent.repositories[rsimbdcp.repo] + if !ok { + repo = newMapBlobDescriptorCache() + rsimbdcp.parent.repositories[rsimbdcp.repo] = repo + } + rsimbdcp.repository = repo + } + rsimbdcp.parent.mu.Unlock() + + if err := repo.SetDescriptor(ctx, dgst, desc); err != nil { + return err + } + + return rsimbdcp.parent.SetDescriptor(ctx, dgst, desc) +} + +// mapBlobDescriptorCache provides a simple map-based implementation of the +// descriptor cache. +type mapBlobDescriptorCache struct { + descriptors map[digest.Digest]distribution.Descriptor + mu sync.RWMutex +} + +var _ distribution.BlobDescriptorService = &mapBlobDescriptorCache{} + +func newMapBlobDescriptorCache() *mapBlobDescriptorCache { + return &mapBlobDescriptorCache{ + descriptors: make(map[digest.Digest]distribution.Descriptor), + } +} + +func (mbdc *mapBlobDescriptorCache) Stat(ctx context.Context, dgst digest.Digest) (distribution.Descriptor, error) { + if err := dgst.Validate(); err != nil { + return distribution.Descriptor{}, err + } + + mbdc.mu.RLock() + defer mbdc.mu.RUnlock() + + desc, ok := mbdc.descriptors[dgst] + if !ok { + return distribution.Descriptor{}, distribution.ErrBlobUnknown + } + + return desc, nil +} + +func (mbdc *mapBlobDescriptorCache) Clear(ctx context.Context, dgst digest.Digest) error { + mbdc.mu.Lock() + defer mbdc.mu.Unlock() + + delete(mbdc.descriptors, dgst) + return nil +} + +func (mbdc *mapBlobDescriptorCache) SetDescriptor(ctx context.Context, dgst digest.Digest, desc distribution.Descriptor) error { + if err := dgst.Validate(); err != nil { + return err + } + + if err := cache.ValidateDescriptor(desc); err != nil { + return err + } + + mbdc.mu.Lock() + defer mbdc.mu.Unlock() + + mbdc.descriptors[dgst] = desc + return nil +} diff --git a/vendor/github.com/docker/distribution/tags.go b/vendor/github.com/docker/distribution/tags.go new file mode 100644 index 000000000..f22df2b85 --- /dev/null +++ b/vendor/github.com/docker/distribution/tags.go @@ -0,0 +1,27 @@ +package distribution + +import ( + "context" +) + +// TagService provides access to information about tagged objects. +type TagService interface { + // Get retrieves the descriptor identified by the tag. Some + // implementations may differentiate between "trusted" tags and + // "untrusted" tags. If a tag is "untrusted", the mapping will be returned + // as an ErrTagUntrusted error, with the target descriptor. + Get(ctx context.Context, tag string) (Descriptor, error) + + // Tag associates the tag with the provided descriptor, updating the + // current association, if needed. + Tag(ctx context.Context, tag string, desc Descriptor) error + + // Untag removes the given tag association + Untag(ctx context.Context, tag string) error + + // All returns the set of tags managed by this tag service + All(ctx context.Context) ([]string, error) + + // Lookup returns the set of tags referencing the given digest. + Lookup(ctx context.Context, digest Descriptor) ([]string, error) +} diff --git a/vendor/github.com/docker/distribution/vendor.conf b/vendor/github.com/docker/distribution/vendor.conf new file mode 100644 index 000000000..31bb03422 --- /dev/null +++ b/vendor/github.com/docker/distribution/vendor.conf @@ -0,0 +1,51 @@ +github.com/Azure/azure-sdk-for-go 4650843026a7fdec254a8d9cf893693a254edd0b +github.com/Azure/go-autorest eaa7994b2278094c904d31993d26f56324db3052 +github.com/sirupsen/logrus 3d4380f53a34dcdc95f0c1db702615992b38d9a4 +github.com/aws/aws-sdk-go f831d5a0822a1ad72420ab18c6269bca1ddaf490 +github.com/bshuster-repo/logrus-logstash-hook d2c0ecc1836d91814e15e23bb5dc309c3ef51f4a +github.com/beorn7/perks 4c0e84591b9aa9e6dcfdf3e020114cd81f89d5f9 +github.com/bugsnag/bugsnag-go b1d153021fcd90ca3f080db36bec96dc690fb274 +github.com/bugsnag/osext 0dd3f918b21bec95ace9dc86c7e70266cfc5c702 +github.com/bugsnag/panicwrap e2c28503fcd0675329da73bf48b33404db873782 +github.com/denverdino/aliyungo afedced274aa9a7fcdd47ac97018f0f8db4e5de2 +github.com/dgrijalva/jwt-go a601269ab70c205d26370c16f7c81e9017c14e04 +github.com/docker/go-metrics 399ea8c73916000c64c2c76e8da00ca82f8387ab +github.com/docker/libtrust fa567046d9b14f6aa788882a950d69651d230b21 +github.com/garyburd/redigo 535138d7bcd717d6531c701ef5933d98b1866257 +github.com/go-ini/ini 2ba15ac2dc9cdf88c110ec2dc0ced7fa45f5678c +github.com/golang/protobuf 8d92cf5fc15a4382f8964b08e1f42a75c0591aa3 +github.com/gorilla/handlers 60c7bfde3e33c201519a200a4507a158cc03a17b +github.com/gorilla/mux 599cba5e7b6137d46ddf58fb1765f5d928e69604 +github.com/inconshreveable/mousetrap 76626ae9c91c4f2a10f34cad8ce83ea42c93bb75 +github.com/jmespath/go-jmespath bd40a432e4c76585ef6b72d3fd96fb9b6dc7b68d +github.com/marstr/guid 8bd9a64bf37eb297b492a4101fb28e80ac0b290f +github.com/satori/go.uuid f58768cc1a7a7e77a3bd49e98cdd21419399b6a3 +github.com/matttproud/golang_protobuf_extensions c12348ce28de40eed0136aa2b644d0ee0650e56c +github.com/miekg/dns 271c58e0c14f552178ea321a545ff9af38930f39 +github.com/mitchellh/mapstructure 482a9fd5fa83e8c4e7817413b80f3eb8feec03ef +github.com/ncw/swift b964f2ca856aac39885e258ad25aec08d5f64ee6 +github.com/prometheus/client_golang c332b6f63c0658a65eca15c0e5247ded801cf564 +github.com/prometheus/client_model 99fa1f4be8e564e8a6b613da7fa6f46c9edafc6c +github.com/prometheus/common 89604d197083d4781071d3c65855d24ecfb0a563 +github.com/prometheus/procfs cb4147076ac75738c9a7d279075a253c0cc5acbd +github.com/spf13/cobra 312092086bed4968099259622145a0c9ae280064 +github.com/spf13/pflag 5644820622454e71517561946e3d94b9f9db6842 +github.com/stevvooe/resumable 2aaf90b2ceea5072cb503ef2a620b08ff3119870 +github.com/xenolf/lego a9d8cec0e6563575e5868a005359ac97911b5985 +github.com/yvasiyarov/go-metrics 57bccd1ccd43f94bb17fdd8bf3007059b802f85e +github.com/yvasiyarov/gorelic a9bba5b9ab508a086f9a12b8c51fab68478e2128 +github.com/yvasiyarov/newrelic_platform_go b21fdbd4370f3717f3bbd2bf41c223bc273068e6 +golang.org/x/crypto c10c31b5e94b6f7a0283272dc2bb27163dcea24b +golang.org/x/net 4876518f9e71663000c348837735820161a42df7 +golang.org/x/oauth2 045497edb6234273d67dbc25da3f2ddbc4c4cacf +golang.org/x/time a4bde12657593d5e90d0533a3e4fd95e635124cb +google.golang.org/api 9bf6e6e569ff057f75d9604a46c52928f17d2b54 +google.golang.org/appengine 12d5545dc1cfa6047a286d5e853841b6471f4c19 +google.golang.org/cloud 975617b05ea8a58727e6c1a06b6161ff4185a9f2 +google.golang.org/grpc d3ddb4469d5a1b949fc7a7da7c1d6a0d1b6de994 +gopkg.in/check.v1 64131543e7896d5bcc6bd5a76287eb75ea96c673 +gopkg.in/square/go-jose.v1 40d457b439244b546f023d056628e5184136899b +gopkg.in/yaml.v2 v2.2.1 +rsc.io/letsencrypt e770c10b0f1a64775ae91d240407ce00d1a5bdeb https://github.com/dmcgowan/letsencrypt.git +github.com/opencontainers/go-digest a6d0ee40d4207ea02364bd3b9e8e77b9159ba1eb +github.com/opencontainers/image-spec ab7389ef9f50030c9b245bc16b981c7ddf192882 diff --git a/vendor/github.com/docker/docker-credential-helpers/LICENSE b/vendor/github.com/docker/docker-credential-helpers/LICENSE new file mode 100644 index 000000000..1ea555e2a --- /dev/null +++ b/vendor/github.com/docker/docker-credential-helpers/LICENSE @@ -0,0 +1,20 @@ +Copyright (c) 2016 David Calavera + +Permission is hereby granted, free of charge, to any person obtaining +a copy of this software and associated documentation files (the +"Software"), to deal in the Software without restriction, including +without limitation the rights to use, copy, modify, merge, publish, +distribute, sublicense, and/or sell copies of the Software, and to +permit persons to whom the Software is furnished to do so, subject to +the following conditions: + +The above copyright notice and this permission notice shall be +included in all copies or substantial portions of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, +EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF +MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. +IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY +CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, +TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE +SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. diff --git a/vendor/github.com/docker/docker-credential-helpers/client/client.go b/vendor/github.com/docker/docker-credential-helpers/client/client.go new file mode 100644 index 000000000..d1d0434cb --- /dev/null +++ b/vendor/github.com/docker/docker-credential-helpers/client/client.go @@ -0,0 +1,121 @@ +package client + +import ( + "bytes" + "encoding/json" + "fmt" + "strings" + + "github.com/docker/docker-credential-helpers/credentials" +) + +// isValidCredsMessage checks if 'msg' contains invalid credentials error message. +// It returns whether the logs are free of invalid credentials errors and the error if it isn't. +// error values can be errCredentialsMissingServerURL or errCredentialsMissingUsername. +func isValidCredsMessage(msg string) error { + if credentials.IsCredentialsMissingServerURLMessage(msg) { + return credentials.NewErrCredentialsMissingServerURL() + } + + if credentials.IsCredentialsMissingUsernameMessage(msg) { + return credentials.NewErrCredentialsMissingUsername() + } + + return nil +} + +// Store uses an external program to save credentials. +func Store(program ProgramFunc, creds *credentials.Credentials) error { + cmd := program("store") + + buffer := new(bytes.Buffer) + if err := json.NewEncoder(buffer).Encode(creds); err != nil { + return err + } + cmd.Input(buffer) + + out, err := cmd.Output() + if err != nil { + t := strings.TrimSpace(string(out)) + + if isValidErr := isValidCredsMessage(t); isValidErr != nil { + err = isValidErr + } + + return fmt.Errorf("error storing credentials - err: %v, out: `%s`", err, t) + } + + return nil +} + +// Get executes an external program to get the credentials from a native store. +func Get(program ProgramFunc, serverURL string) (*credentials.Credentials, error) { + cmd := program("get") + cmd.Input(strings.NewReader(serverURL)) + + out, err := cmd.Output() + if err != nil { + t := strings.TrimSpace(string(out)) + + if credentials.IsErrCredentialsNotFoundMessage(t) { + return nil, credentials.NewErrCredentialsNotFound() + } + + if isValidErr := isValidCredsMessage(t); isValidErr != nil { + err = isValidErr + } + + return nil, fmt.Errorf("error getting credentials - err: %v, out: `%s`", err, t) + } + + resp := &credentials.Credentials{ + ServerURL: serverURL, + } + + if err := json.NewDecoder(bytes.NewReader(out)).Decode(resp); err != nil { + return nil, err + } + + return resp, nil +} + +// Erase executes a program to remove the server credentials from the native store. +func Erase(program ProgramFunc, serverURL string) error { + cmd := program("erase") + cmd.Input(strings.NewReader(serverURL)) + out, err := cmd.Output() + if err != nil { + t := strings.TrimSpace(string(out)) + + if isValidErr := isValidCredsMessage(t); isValidErr != nil { + err = isValidErr + } + + return fmt.Errorf("error erasing credentials - err: %v, out: `%s`", err, t) + } + + return nil +} + +// List executes a program to list server credentials in the native store. +func List(program ProgramFunc) (map[string]string, error) { + cmd := program("list") + cmd.Input(strings.NewReader("unused")) + out, err := cmd.Output() + if err != nil { + t := strings.TrimSpace(string(out)) + + if isValidErr := isValidCredsMessage(t); isValidErr != nil { + err = isValidErr + } + + return nil, fmt.Errorf("error listing credentials - err: %v, out: `%s`", err, t) + } + + var resp map[string]string + if err = json.NewDecoder(bytes.NewReader(out)).Decode(&resp); err != nil { + return nil, err + } + + return resp, nil +} diff --git a/vendor/github.com/docker/docker-credential-helpers/client/command.go b/vendor/github.com/docker/docker-credential-helpers/client/command.go new file mode 100644 index 000000000..8da334306 --- /dev/null +++ b/vendor/github.com/docker/docker-credential-helpers/client/command.go @@ -0,0 +1,56 @@ +package client + +import ( + "fmt" + "io" + "os" + "os/exec" +) + +// Program is an interface to execute external programs. +type Program interface { + Output() ([]byte, error) + Input(in io.Reader) +} + +// ProgramFunc is a type of function that initializes programs based on arguments. +type ProgramFunc func(args ...string) Program + +// NewShellProgramFunc creates programs that are executed in a Shell. +func NewShellProgramFunc(name string) ProgramFunc { + return NewShellProgramFuncWithEnv(name, nil) +} + +// NewShellProgramFuncWithEnv creates programs that are executed in a Shell with environment variables +func NewShellProgramFuncWithEnv(name string, env *map[string]string) ProgramFunc { + return func(args ...string) Program { + return &Shell{cmd: createProgramCmdRedirectErr(name, args, env)} + } +} + +func createProgramCmdRedirectErr(commandName string, args []string, env *map[string]string) *exec.Cmd { + programCmd := exec.Command(commandName, args...) + programCmd.Env = os.Environ() + if env != nil { + for k, v := range *env { + programCmd.Env = append(programCmd.Env, fmt.Sprintf("%s=%s", k, v)) + } + } + programCmd.Stderr = os.Stderr + return programCmd +} + +// Shell invokes shell commands to talk with a remote credentials helper. +type Shell struct { + cmd *exec.Cmd +} + +// Output returns responses from the remote credentials helper. +func (s *Shell) Output() ([]byte, error) { + return s.cmd.Output() +} + +// Input sets the input to send to a remote credentials helper. +func (s *Shell) Input(in io.Reader) { + s.cmd.Stdin = in +} diff --git a/vendor/github.com/docker/docker-credential-helpers/credentials/credentials.go b/vendor/github.com/docker/docker-credential-helpers/credentials/credentials.go new file mode 100644 index 000000000..da8b594e7 --- /dev/null +++ b/vendor/github.com/docker/docker-credential-helpers/credentials/credentials.go @@ -0,0 +1,186 @@ +package credentials + +import ( + "bufio" + "bytes" + "encoding/json" + "fmt" + "io" + "os" + "strings" +) + +// Credentials holds the information shared between docker and the credentials store. +type Credentials struct { + ServerURL string + Username string + Secret string +} + +// isValid checks the integrity of Credentials object such that no credentials lack +// a server URL or a username. +// It returns whether the credentials are valid and the error if it isn't. +// error values can be errCredentialsMissingServerURL or errCredentialsMissingUsername +func (c *Credentials) isValid() (bool, error) { + if len(c.ServerURL) == 0 { + return false, NewErrCredentialsMissingServerURL() + } + + if len(c.Username) == 0 { + return false, NewErrCredentialsMissingUsername() + } + + return true, nil +} + +// CredsLabel holds the way Docker credentials should be labeled as such in credentials stores that allow labelling. +// That label allows to filter out non-Docker credentials too at lookup/search in macOS keychain, +// Windows credentials manager and Linux libsecret. Default value is "Docker Credentials" +var CredsLabel = "Docker Credentials" + +// SetCredsLabel is a simple setter for CredsLabel +func SetCredsLabel(label string) { + CredsLabel = label +} + +// Serve initializes the credentials helper and parses the action argument. +// This function is designed to be called from a command line interface. +// It uses os.Args[1] as the key for the action. +// It uses os.Stdin as input and os.Stdout as output. +// This function terminates the program with os.Exit(1) if there is an error. +func Serve(helper Helper) { + var err error + if len(os.Args) != 2 { + err = fmt.Errorf("Usage: %s ", os.Args[0]) + } + + if err == nil { + err = HandleCommand(helper, os.Args[1], os.Stdin, os.Stdout) + } + + if err != nil { + fmt.Fprintf(os.Stdout, "%v\n", err) + os.Exit(1) + } +} + +// HandleCommand uses a helper and a key to run a credential action. +func HandleCommand(helper Helper, key string, in io.Reader, out io.Writer) error { + switch key { + case "store": + return Store(helper, in) + case "get": + return Get(helper, in, out) + case "erase": + return Erase(helper, in) + case "list": + return List(helper, out) + case "version": + return PrintVersion(out) + } + return fmt.Errorf("Unknown credential action `%s`", key) +} + +// Store uses a helper and an input reader to save credentials. +// The reader must contain the JSON serialization of a Credentials struct. +func Store(helper Helper, reader io.Reader) error { + scanner := bufio.NewScanner(reader) + + buffer := new(bytes.Buffer) + for scanner.Scan() { + buffer.Write(scanner.Bytes()) + } + + if err := scanner.Err(); err != nil && err != io.EOF { + return err + } + + var creds Credentials + if err := json.NewDecoder(buffer).Decode(&creds); err != nil { + return err + } + + if ok, err := creds.isValid(); !ok { + return err + } + + return helper.Add(&creds) +} + +// Get retrieves the credentials for a given server url. +// The reader must contain the server URL to search. +// The writer is used to write the JSON serialization of the credentials. +func Get(helper Helper, reader io.Reader, writer io.Writer) error { + scanner := bufio.NewScanner(reader) + + buffer := new(bytes.Buffer) + for scanner.Scan() { + buffer.Write(scanner.Bytes()) + } + + if err := scanner.Err(); err != nil && err != io.EOF { + return err + } + + serverURL := strings.TrimSpace(buffer.String()) + if len(serverURL) == 0 { + return NewErrCredentialsMissingServerURL() + } + + username, secret, err := helper.Get(serverURL) + if err != nil { + return err + } + + resp := Credentials{ + ServerURL: serverURL, + Username: username, + Secret: secret, + } + + buffer.Reset() + if err := json.NewEncoder(buffer).Encode(resp); err != nil { + return err + } + + fmt.Fprint(writer, buffer.String()) + return nil +} + +// Erase removes credentials from the store. +// The reader must contain the server URL to remove. +func Erase(helper Helper, reader io.Reader) error { + scanner := bufio.NewScanner(reader) + + buffer := new(bytes.Buffer) + for scanner.Scan() { + buffer.Write(scanner.Bytes()) + } + + if err := scanner.Err(); err != nil && err != io.EOF { + return err + } + + serverURL := strings.TrimSpace(buffer.String()) + if len(serverURL) == 0 { + return NewErrCredentialsMissingServerURL() + } + + return helper.Delete(serverURL) +} + +//List returns all the serverURLs of keys in +//the OS store as a list of strings +func List(helper Helper, writer io.Writer) error { + accts, err := helper.List() + if err != nil { + return err + } + return json.NewEncoder(writer).Encode(accts) +} + +//PrintVersion outputs the current version. +func PrintVersion(writer io.Writer) error { + fmt.Fprintln(writer, Version) + return nil +} diff --git a/vendor/github.com/docker/docker-credential-helpers/credentials/error.go b/vendor/github.com/docker/docker-credential-helpers/credentials/error.go new file mode 100644 index 000000000..fe6a5aef4 --- /dev/null +++ b/vendor/github.com/docker/docker-credential-helpers/credentials/error.go @@ -0,0 +1,102 @@ +package credentials + +const ( + // ErrCredentialsNotFound standardizes the not found error, so every helper returns + // the same message and docker can handle it properly. + errCredentialsNotFoundMessage = "credentials not found in native keychain" + + // ErrCredentialsMissingServerURL and ErrCredentialsMissingUsername standardize + // invalid credentials or credentials management operations + errCredentialsMissingServerURLMessage = "no credentials server URL" + errCredentialsMissingUsernameMessage = "no credentials username" +) + +// errCredentialsNotFound represents an error +// raised when credentials are not in the store. +type errCredentialsNotFound struct{} + +// Error returns the standard error message +// for when the credentials are not in the store. +func (errCredentialsNotFound) Error() string { + return errCredentialsNotFoundMessage +} + +// NewErrCredentialsNotFound creates a new error +// for when the credentials are not in the store. +func NewErrCredentialsNotFound() error { + return errCredentialsNotFound{} +} + +// IsErrCredentialsNotFound returns true if the error +// was caused by not having a set of credentials in a store. +func IsErrCredentialsNotFound(err error) bool { + _, ok := err.(errCredentialsNotFound) + return ok +} + +// IsErrCredentialsNotFoundMessage returns true if the error +// was caused by not having a set of credentials in a store. +// +// This function helps to check messages returned by an +// external program via its standard output. +func IsErrCredentialsNotFoundMessage(err string) bool { + return err == errCredentialsNotFoundMessage +} + +// errCredentialsMissingServerURL represents an error raised +// when the credentials object has no server URL or when no +// server URL is provided to a credentials operation requiring +// one. +type errCredentialsMissingServerURL struct{} + +func (errCredentialsMissingServerURL) Error() string { + return errCredentialsMissingServerURLMessage +} + +// errCredentialsMissingUsername represents an error raised +// when the credentials object has no username or when no +// username is provided to a credentials operation requiring +// one. +type errCredentialsMissingUsername struct{} + +func (errCredentialsMissingUsername) Error() string { + return errCredentialsMissingUsernameMessage +} + +// NewErrCredentialsMissingServerURL creates a new error for +// errCredentialsMissingServerURL. +func NewErrCredentialsMissingServerURL() error { + return errCredentialsMissingServerURL{} +} + +// NewErrCredentialsMissingUsername creates a new error for +// errCredentialsMissingUsername. +func NewErrCredentialsMissingUsername() error { + return errCredentialsMissingUsername{} +} + +// IsCredentialsMissingServerURL returns true if the error +// was an errCredentialsMissingServerURL. +func IsCredentialsMissingServerURL(err error) bool { + _, ok := err.(errCredentialsMissingServerURL) + return ok +} + +// IsCredentialsMissingServerURLMessage checks for an +// errCredentialsMissingServerURL in the error message. +func IsCredentialsMissingServerURLMessage(err string) bool { + return err == errCredentialsMissingServerURLMessage +} + +// IsCredentialsMissingUsername returns true if the error +// was an errCredentialsMissingUsername. +func IsCredentialsMissingUsername(err error) bool { + _, ok := err.(errCredentialsMissingUsername) + return ok +} + +// IsCredentialsMissingUsernameMessage checks for an +// errCredentialsMissingUsername in the error message. +func IsCredentialsMissingUsernameMessage(err string) bool { + return err == errCredentialsMissingUsernameMessage +} diff --git a/vendor/github.com/docker/docker-credential-helpers/credentials/helper.go b/vendor/github.com/docker/docker-credential-helpers/credentials/helper.go new file mode 100644 index 000000000..135acd254 --- /dev/null +++ b/vendor/github.com/docker/docker-credential-helpers/credentials/helper.go @@ -0,0 +1,14 @@ +package credentials + +// Helper is the interface a credentials store helper must implement. +type Helper interface { + // Add appends credentials to the store. + Add(*Credentials) error + // Delete removes credentials from the store. + Delete(serverURL string) error + // Get retrieves credentials from the store. + // It returns username and secret as strings. + Get(serverURL string) (string, string, error) + // List returns the stored serverURLs and their associated usernames. + List() (map[string]string, error) +} diff --git a/vendor/github.com/docker/docker-credential-helpers/credentials/version.go b/vendor/github.com/docker/docker-credential-helpers/credentials/version.go new file mode 100644 index 000000000..033a5fee5 --- /dev/null +++ b/vendor/github.com/docker/docker-credential-helpers/credentials/version.go @@ -0,0 +1,4 @@ +package credentials + +// Version holds a string describing the current version +const Version = "0.6.0" diff --git a/vendor/github.com/docker/docker/api/types/auth.go b/vendor/github.com/docker/docker/api/types/auth.go index 056af6b84..ddf15bb18 100644 --- a/vendor/github.com/docker/docker/api/types/auth.go +++ b/vendor/github.com/docker/docker/api/types/auth.go @@ -1,4 +1,4 @@ -package types +package types // import "github.com/docker/docker/api/types" // AuthConfig contains authorization information for connecting to a Registry type AuthConfig struct { diff --git a/vendor/github.com/docker/docker/api/types/blkiodev/blkio.go b/vendor/github.com/docker/docker/api/types/blkiodev/blkio.go index 931ae10ab..bf3463b90 100644 --- a/vendor/github.com/docker/docker/api/types/blkiodev/blkio.go +++ b/vendor/github.com/docker/docker/api/types/blkiodev/blkio.go @@ -1,4 +1,4 @@ -package blkiodev +package blkiodev // import "github.com/docker/docker/api/types/blkiodev" import "fmt" diff --git a/vendor/github.com/docker/docker/api/types/client.go b/vendor/github.com/docker/docker/api/types/client.go index 93ca42854..3b698c2c2 100644 --- a/vendor/github.com/docker/docker/api/types/client.go +++ b/vendor/github.com/docker/docker/api/types/client.go @@ -1,4 +1,4 @@ -package types +package types // import "github.com/docker/docker/api/types" import ( "bufio" @@ -181,8 +181,24 @@ type ImageBuildOptions struct { Target string SessionID string Platform string + // Version specifies the version of the unerlying builder to use + Version BuilderVersion + // BuildID is an optional identifier that can be passed together with the + // build request. The same identifier can be used to gracefully cancel the + // build with the cancel request. + BuildID string } +// BuilderVersion sets the version of underlying builder to use +type BuilderVersion string + +const ( + // BuilderV1 is the first generation builder in docker daemon + BuilderV1 BuilderVersion = "1" + // BuilderBuildKit is builder based on moby/buildkit project + BuilderBuildKit = "2" +) + // ImageBuildResponse holds information // returned by a server after building // an image. diff --git a/vendor/github.com/docker/docker/api/types/configs.go b/vendor/github.com/docker/docker/api/types/configs.go index 20c19f213..f6537a27f 100644 --- a/vendor/github.com/docker/docker/api/types/configs.go +++ b/vendor/github.com/docker/docker/api/types/configs.go @@ -1,4 +1,4 @@ -package types +package types // import "github.com/docker/docker/api/types" import ( "github.com/docker/docker/api/types/container" @@ -25,19 +25,6 @@ type ContainerRmConfig struct { ForceRemove, RemoveVolume, RemoveLink bool } -// ContainerCommitConfig contains build configs for commit operation, -// and is used when making a commit with the current state of the container. -type ContainerCommitConfig struct { - Pause bool - Repo string - Tag string - Author string - Comment string - // merge container config into commit config before commit - MergeConfigs bool - Config *container.Config -} - // ExecConfig is a small subset of the Config struct that holds the configuration // for the exec feature of docker. type ExecConfig struct { @@ -50,6 +37,7 @@ type ExecConfig struct { Detach bool // Execute in detach mode DetachKeys string // Escape keys for detach Env []string // Environment variables + WorkingDir string // Working directory Cmd []string // Execution commands and args } diff --git a/vendor/github.com/docker/docker/api/types/container/config.go b/vendor/github.com/docker/docker/api/types/container/config.go index fc050e5db..89ad08c23 100644 --- a/vendor/github.com/docker/docker/api/types/container/config.go +++ b/vendor/github.com/docker/docker/api/types/container/config.go @@ -1,4 +1,4 @@ -package container +package container // import "github.com/docker/docker/api/types/container" import ( "time" @@ -7,6 +7,12 @@ import ( "github.com/docker/go-connections/nat" ) +// MinimumDuration puts a minimum on user configured duration. +// This is to prevent API error on time unit. For example, API may +// set 3 as healthcheck interval with intention of 3 seconds, but +// Docker interprets it as 3 nanoseconds. +const MinimumDuration = 1 * time.Millisecond + // HealthConfig holds configuration settings for the HEALTHCHECK feature. type HealthConfig struct { // Test is the test to perform to check that the container is healthy. @@ -19,8 +25,9 @@ type HealthConfig struct { Test []string `json:",omitempty"` // Zero means to inherit. Durations are expressed as integer nanoseconds. - Interval time.Duration `json:",omitempty"` // Interval is the time to wait between checks. - Timeout time.Duration `json:",omitempty"` // Timeout is the time to wait before considering the check to have hung. + Interval time.Duration `json:",omitempty"` // Interval is the time to wait between checks. + Timeout time.Duration `json:",omitempty"` // Timeout is the time to wait before considering the check to have hung. + StartPeriod time.Duration `json:",omitempty"` // The start period for the container to initialize before the retries starts to count down. // Retries is the number of consecutive failures needed to consider a container as unhealthy. // Zero means inherit. diff --git a/vendor/github.com/docker/docker/api/types/container/container_changes.go b/vendor/github.com/docker/docker/api/types/container/container_changes.go index 767945a53..c909d6ca3 100644 --- a/vendor/github.com/docker/docker/api/types/container/container_changes.go +++ b/vendor/github.com/docker/docker/api/types/container/container_changes.go @@ -7,7 +7,7 @@ package container // See hack/generate-swagger-api.sh // ---------------------------------------------------------------------------- -// ContainerChangeResponseItem container change response item +// ContainerChangeResponseItem change item in response to ContainerChanges operation // swagger:model ContainerChangeResponseItem type ContainerChangeResponseItem struct { diff --git a/vendor/github.com/docker/docker/api/types/container/container_create.go b/vendor/github.com/docker/docker/api/types/container/container_create.go index c95023b81..49efa0f2c 100644 --- a/vendor/github.com/docker/docker/api/types/container/container_create.go +++ b/vendor/github.com/docker/docker/api/types/container/container_create.go @@ -7,7 +7,7 @@ package container // See hack/generate-swagger-api.sh // ---------------------------------------------------------------------------- -// ContainerCreateCreatedBody container create created body +// ContainerCreateCreatedBody OK response to ContainerCreate operation // swagger:model ContainerCreateCreatedBody type ContainerCreateCreatedBody struct { diff --git a/vendor/github.com/docker/docker/api/types/container/container_top.go b/vendor/github.com/docker/docker/api/types/container/container_top.go index 78bc37ee5..ba41edcf3 100644 --- a/vendor/github.com/docker/docker/api/types/container/container_top.go +++ b/vendor/github.com/docker/docker/api/types/container/container_top.go @@ -7,7 +7,7 @@ package container // See hack/generate-swagger-api.sh // ---------------------------------------------------------------------------- -// ContainerTopOKBody container top o k body +// ContainerTopOKBody OK response to ContainerTop operation // swagger:model ContainerTopOKBody type ContainerTopOKBody struct { diff --git a/vendor/github.com/docker/docker/api/types/container/container_update.go b/vendor/github.com/docker/docker/api/types/container/container_update.go index 2339366fb..7630ae54c 100644 --- a/vendor/github.com/docker/docker/api/types/container/container_update.go +++ b/vendor/github.com/docker/docker/api/types/container/container_update.go @@ -7,7 +7,7 @@ package container // See hack/generate-swagger-api.sh // ---------------------------------------------------------------------------- -// ContainerUpdateOKBody container update o k body +// ContainerUpdateOKBody OK response to ContainerUpdate operation // swagger:model ContainerUpdateOKBody type ContainerUpdateOKBody struct { diff --git a/vendor/github.com/docker/docker/api/types/container/container_wait.go b/vendor/github.com/docker/docker/api/types/container/container_wait.go index 77ecdbaf7..9e3910a6b 100644 --- a/vendor/github.com/docker/docker/api/types/container/container_wait.go +++ b/vendor/github.com/docker/docker/api/types/container/container_wait.go @@ -7,10 +7,22 @@ package container // See hack/generate-swagger-api.sh // ---------------------------------------------------------------------------- -// ContainerWaitOKBody container wait o k body +// ContainerWaitOKBodyError container waiting error, if any +// swagger:model ContainerWaitOKBodyError +type ContainerWaitOKBodyError struct { + + // Details of an error + Message string `json:"Message,omitempty"` +} + +// ContainerWaitOKBody OK response to ContainerWait operation // swagger:model ContainerWaitOKBody type ContainerWaitOKBody struct { + // error + // Required: true + Error *ContainerWaitOKBodyError `json:"Error"` + // Exit code of the container // Required: true StatusCode int64 `json:"StatusCode"` diff --git a/vendor/github.com/docker/docker/api/types/container/host_config.go b/vendor/github.com/docker/docker/api/types/container/host_config.go index 15a84b82d..4ef26fa6c 100644 --- a/vendor/github.com/docker/docker/api/types/container/host_config.go +++ b/vendor/github.com/docker/docker/api/types/container/host_config.go @@ -1,4 +1,4 @@ -package container +package container // import "github.com/docker/docker/api/types/container" import ( "strings" @@ -10,9 +10,6 @@ import ( "github.com/docker/go-units" ) -// NetworkMode represents the container network stack. -type NetworkMode string - // Isolation represents the isolation technology of a container. The supported // values are platform specific type Isolation string @@ -23,42 +20,101 @@ func (i Isolation) IsDefault() bool { return strings.ToLower(string(i)) == "default" || string(i) == "" } +// IsHyperV indicates the use of a Hyper-V partition for isolation +func (i Isolation) IsHyperV() bool { + return strings.ToLower(string(i)) == "hyperv" +} + +// IsProcess indicates the use of process isolation +func (i Isolation) IsProcess() bool { + return strings.ToLower(string(i)) == "process" +} + +const ( + // IsolationEmpty is unspecified (same behavior as default) + IsolationEmpty = Isolation("") + // IsolationDefault is the default isolation mode on current daemon + IsolationDefault = Isolation("default") + // IsolationProcess is process isolation mode + IsolationProcess = Isolation("process") + // IsolationHyperV is HyperV isolation mode + IsolationHyperV = Isolation("hyperv") +) + // IpcMode represents the container ipc stack. type IpcMode string -// IsPrivate indicates whether the container uses its private ipc stack. +// IsPrivate indicates whether the container uses its own private ipc namespace which can not be shared. func (n IpcMode) IsPrivate() bool { - return !(n.IsHost() || n.IsContainer()) + return n == "private" } -// IsHost indicates whether the container uses the host's ipc stack. +// IsHost indicates whether the container shares the host's ipc namespace. func (n IpcMode) IsHost() bool { return n == "host" } -// IsContainer indicates whether the container uses a container's ipc stack. +// IsShareable indicates whether the container's ipc namespace can be shared with another container. +func (n IpcMode) IsShareable() bool { + return n == "shareable" +} + +// IsContainer indicates whether the container uses another container's ipc namespace. func (n IpcMode) IsContainer() bool { parts := strings.SplitN(string(n), ":", 2) return len(parts) > 1 && parts[0] == "container" } -// Valid indicates whether the ipc stack is valid. +// IsNone indicates whether container IpcMode is set to "none". +func (n IpcMode) IsNone() bool { + return n == "none" +} + +// IsEmpty indicates whether container IpcMode is empty +func (n IpcMode) IsEmpty() bool { + return n == "" +} + +// Valid indicates whether the ipc mode is valid. func (n IpcMode) Valid() bool { - parts := strings.Split(string(n), ":") - switch mode := parts[0]; mode { - case "", "host": - case "container": - if len(parts) != 2 || parts[1] == "" { - return false - } - default: - return false - } - return true + return n.IsEmpty() || n.IsNone() || n.IsPrivate() || n.IsHost() || n.IsShareable() || n.IsContainer() } // Container returns the name of the container ipc stack is going to be used. func (n IpcMode) Container() string { + parts := strings.SplitN(string(n), ":", 2) + if len(parts) > 1 && parts[0] == "container" { + return parts[1] + } + return "" +} + +// NetworkMode represents the container network stack. +type NetworkMode string + +// IsNone indicates whether container isn't using a network stack. +func (n NetworkMode) IsNone() bool { + return n == "none" +} + +// IsDefault indicates whether container uses the default network stack. +func (n NetworkMode) IsDefault() bool { + return n == "default" +} + +// IsPrivate indicates whether container uses its private network stack. +func (n NetworkMode) IsPrivate() bool { + return !(n.IsHost() || n.IsContainer()) +} + +// IsContainer indicates whether container uses a container network stack. +func (n NetworkMode) IsContainer() bool { + parts := strings.SplitN(string(n), ":", 2) + return len(parts) > 1 && parts[0] == "container" +} + +// ConnectedContainer is the id of the container which network this container is connected to. +func (n NetworkMode) ConnectedContainer() string { parts := strings.SplitN(string(n), ":", 2) if len(parts) > 1 { return parts[1] @@ -66,6 +122,14 @@ func (n IpcMode) Container() string { return "" } +//UserDefined indicates user-created network +func (n NetworkMode) UserDefined() string { + if n.IsUserDefined() { + return string(n) + } + return "" +} + // UsernsMode represents userns mode in the container. type UsernsMode string @@ -337,9 +401,12 @@ type HostConfig struct { // Mounts specs used by the container Mounts []mount.Mount `json:",omitempty"` + // MaskedPaths is the list of paths to be masked inside the container (this overrides the default set of paths) + MaskedPaths []string + + // ReadonlyPaths is the list of paths to be set as read-only inside the container (this overrides the default set of paths) + ReadonlyPaths []string + // Run a custom init inside the container, if null, use the daemon's configured settings Init *bool `json:",omitempty"` - - // Custom init path - InitPath string `json:",omitempty"` } diff --git a/vendor/github.com/docker/docker/api/types/container/hostconfig_unix.go b/vendor/github.com/docker/docker/api/types/container/hostconfig_unix.go index 9fb79bed6..cf6fdf440 100644 --- a/vendor/github.com/docker/docker/api/types/container/hostconfig_unix.go +++ b/vendor/github.com/docker/docker/api/types/container/hostconfig_unix.go @@ -1,24 +1,12 @@ // +build !windows -package container - -import "strings" +package container // import "github.com/docker/docker/api/types/container" // IsValid indicates if an isolation technology is valid func (i Isolation) IsValid() bool { return i.IsDefault() } -// IsPrivate indicates whether container uses its private network stack. -func (n NetworkMode) IsPrivate() bool { - return !(n.IsHost() || n.IsContainer()) -} - -// IsDefault indicates whether container uses the default network stack. -func (n NetworkMode) IsDefault() bool { - return n == "default" -} - // NetworkName returns the name of the network stack. func (n NetworkMode) NetworkName() string { if n.IsBridge() { @@ -47,35 +35,7 @@ func (n NetworkMode) IsHost() bool { return n == "host" } -// IsContainer indicates whether container uses a container network stack. -func (n NetworkMode) IsContainer() bool { - parts := strings.SplitN(string(n), ":", 2) - return len(parts) > 1 && parts[0] == "container" -} - -// IsNone indicates whether container isn't using a network stack. -func (n NetworkMode) IsNone() bool { - return n == "none" -} - -// ConnectedContainer is the id of the container which network this container is connected to. -func (n NetworkMode) ConnectedContainer() string { - parts := strings.SplitN(string(n), ":", 2) - if len(parts) > 1 { - return parts[1] - } - return "" -} - // IsUserDefined indicates user-created network func (n NetworkMode) IsUserDefined() bool { return !n.IsDefault() && !n.IsBridge() && !n.IsHost() && !n.IsNone() && !n.IsContainer() } - -//UserDefined indicates user-created network -func (n NetworkMode) UserDefined() string { - if n.IsUserDefined() { - return string(n) - } - return "" -} diff --git a/vendor/github.com/docker/docker/api/types/container/hostconfig_windows.go b/vendor/github.com/docker/docker/api/types/container/hostconfig_windows.go index 0ee332ba6..99f803a5b 100644 --- a/vendor/github.com/docker/docker/api/types/container/hostconfig_windows.go +++ b/vendor/github.com/docker/docker/api/types/container/hostconfig_windows.go @@ -1,24 +1,4 @@ -package container - -import ( - "strings" -) - -// IsDefault indicates whether container uses the default network stack. -func (n NetworkMode) IsDefault() bool { - return n == "default" -} - -// IsNone indicates whether container isn't using a network stack. -func (n NetworkMode) IsNone() bool { - return n == "none" -} - -// IsContainer indicates whether container uses a container network stack. -// Returns false as windows doesn't support this mode -func (n NetworkMode) IsContainer() bool { - return false -} +package container // import "github.com/docker/docker/api/types/container" // IsBridge indicates whether container uses the bridge network stack // in windows it is given the name NAT @@ -32,30 +12,9 @@ func (n NetworkMode) IsHost() bool { return false } -// IsPrivate indicates whether container uses its private network stack. -func (n NetworkMode) IsPrivate() bool { - return !(n.IsHost() || n.IsContainer()) -} - -// ConnectedContainer is the id of the container which network this container is connected to. -// Returns blank string on windows -func (n NetworkMode) ConnectedContainer() string { - return "" -} - // IsUserDefined indicates user-created network func (n NetworkMode) IsUserDefined() bool { - return !n.IsDefault() && !n.IsNone() && !n.IsBridge() -} - -// IsHyperV indicates the use of a Hyper-V partition for isolation -func (i Isolation) IsHyperV() bool { - return strings.ToLower(string(i)) == "hyperv" -} - -// IsProcess indicates the use of process isolation -func (i Isolation) IsProcess() bool { - return strings.ToLower(string(i)) == "process" + return !n.IsDefault() && !n.IsNone() && !n.IsBridge() && !n.IsContainer() } // IsValid indicates if an isolation technology is valid @@ -71,17 +30,11 @@ func (n NetworkMode) NetworkName() string { return "nat" } else if n.IsNone() { return "none" + } else if n.IsContainer() { + return "container" } else if n.IsUserDefined() { return n.UserDefined() } return "" } - -//UserDefined indicates user-created network -func (n NetworkMode) UserDefined() string { - if n.IsUserDefined() { - return string(n) - } - return "" -} diff --git a/vendor/github.com/docker/docker/api/types/container/waitcondition.go b/vendor/github.com/docker/docker/api/types/container/waitcondition.go new file mode 100644 index 000000000..cd8311f99 --- /dev/null +++ b/vendor/github.com/docker/docker/api/types/container/waitcondition.go @@ -0,0 +1,22 @@ +package container // import "github.com/docker/docker/api/types/container" + +// WaitCondition is a type used to specify a container state for which +// to wait. +type WaitCondition string + +// Possible WaitCondition Values. +// +// WaitConditionNotRunning (default) is used to wait for any of the non-running +// states: "created", "exited", "dead", "removing", or "removed". +// +// WaitConditionNextExit is used to wait for the next time the state changes +// to a non-running state. If the state is currently "created" or "exited", +// this would cause Wait() to block until either the container runs and exits +// or is removed. +// +// WaitConditionRemoved is used to wait for the container to be removed. +const ( + WaitConditionNotRunning WaitCondition = "not-running" + WaitConditionNextExit WaitCondition = "next-exit" + WaitConditionRemoved WaitCondition = "removed" +) diff --git a/vendor/github.com/docker/docker/api/types/filters/parse.go b/vendor/github.com/docker/docker/api/types/filters/parse.go index beec3d494..a41e3d8d9 100644 --- a/vendor/github.com/docker/docker/api/types/filters/parse.go +++ b/vendor/github.com/docker/docker/api/types/filters/parse.go @@ -1,38 +1,45 @@ -// Package filters provides helper function to parse and handle command line -// filter, used for example in docker ps or docker images commands. -package filters +/*Package filters provides tools for encoding a mapping of keys to a set of +multiple values. +*/ +package filters // import "github.com/docker/docker/api/types/filters" import ( "encoding/json" "errors" - "fmt" "regexp" "strings" "github.com/docker/docker/api/types/versions" ) -// Args stores filter arguments as map key:{map key: bool}. -// It contains an aggregation of the map of arguments (which are in the form -// of -f 'key=value') based on the key, and stores values for the same key -// in a map with string keys and boolean values. -// e.g given -f 'label=label1=1' -f 'label=label2=2' -f 'image.name=ubuntu' -// the args will be {"image.name":{"ubuntu":true},"label":{"label1=1":true,"label2=2":true}} +// Args stores a mapping of keys to a set of multiple values. type Args struct { fields map[string]map[string]bool } -// NewArgs initializes a new Args struct. -func NewArgs() Args { - return Args{fields: map[string]map[string]bool{}} +// KeyValuePair are used to initialize a new Args +type KeyValuePair struct { + Key string + Value string } -// ParseFlag parses the argument to the filter flag. Like +// Arg creates a new KeyValuePair for initializing Args +func Arg(key, value string) KeyValuePair { + return KeyValuePair{Key: key, Value: value} +} + +// NewArgs returns a new Args populated with the initial args +func NewArgs(initialArgs ...KeyValuePair) Args { + args := Args{fields: map[string]map[string]bool{}} + for _, arg := range initialArgs { + args.Add(arg.Key, arg.Value) + } + return args +} + +// ParseFlag parses a key=value string and adds it to an Args. // -// `docker ps -f 'created=today' -f 'image.name=ubuntu*'` -// -// If prev map is provided, then it is appended to, and returned. By default a new -// map is created. +// Deprecated: Use Args.Add() func ParseFlag(arg string, prev Args) (Args, error) { filters := prev if len(arg) == 0 { @@ -53,74 +60,95 @@ func ParseFlag(arg string, prev Args) (Args, error) { return filters, nil } -// ErrBadFormat is an error returned in case of bad format for a filter. +// ErrBadFormat is an error returned when a filter is not in the form key=value +// +// Deprecated: this error will be removed in a future version var ErrBadFormat = errors.New("bad format of filter (expected name=value)") -// ToParam packs the Args into a string for easy transport from client to server. +// ToParam encodes the Args as args JSON encoded string +// +// Deprecated: use ToJSON func ToParam(a Args) (string, error) { - // this way we don't URL encode {}, just empty space + return ToJSON(a) +} + +// MarshalJSON returns a JSON byte representation of the Args +func (args Args) MarshalJSON() ([]byte, error) { + if len(args.fields) == 0 { + return []byte{}, nil + } + return json.Marshal(args.fields) +} + +// ToJSON returns the Args as a JSON encoded string +func ToJSON(a Args) (string, error) { if a.Len() == 0 { return "", nil } - - buf, err := json.Marshal(a.fields) - if err != nil { - return "", err - } - return string(buf), nil + buf, err := json.Marshal(a) + return string(buf), err } -// ToParamWithVersion packs the Args into a string for easy transport from client to server. -// The generated string will depend on the specified version (corresponding to the API version). +// ToParamWithVersion encodes Args as a JSON string. If version is less than 1.22 +// then the encoded format will use an older legacy format where the values are a +// list of strings, instead of a set. +// +// Deprecated: Use ToJSON func ToParamWithVersion(version string, a Args) (string, error) { - // this way we don't URL encode {}, just empty space if a.Len() == 0 { return "", nil } - // for daemons older than v1.10, filter must be of the form map[string][]string - var buf []byte - var err error if version != "" && versions.LessThan(version, "1.22") { - buf, err = json.Marshal(convertArgsToSlice(a.fields)) - } else { - buf, err = json.Marshal(a.fields) + buf, err := json.Marshal(convertArgsToSlice(a.fields)) + return string(buf), err } - if err != nil { - return "", err - } - return string(buf), nil + + return ToJSON(a) } -// FromParam unpacks the filter Args. +// FromParam decodes a JSON encoded string into Args +// +// Deprecated: use FromJSON func FromParam(p string) (Args, error) { - if len(p) == 0 { - return NewArgs(), nil - } - - r := strings.NewReader(p) - d := json.NewDecoder(r) - - m := map[string]map[string]bool{} - if err := d.Decode(&m); err != nil { - r.Seek(0, 0) - - // Allow parsing old arguments in slice format. - // Because other libraries might be sending them in this format. - deprecated := map[string][]string{} - if deprecatedErr := d.Decode(&deprecated); deprecatedErr == nil { - m = deprecatedArgs(deprecated) - } else { - return NewArgs(), err - } - } - return Args{m}, nil + return FromJSON(p) } -// Get returns the list of values associates with a field. -// It returns a slice of strings to keep backwards compatibility with old code. -func (filters Args) Get(field string) []string { - values := filters.fields[field] +// FromJSON decodes a JSON encoded string into Args +func FromJSON(p string) (Args, error) { + args := NewArgs() + + if p == "" { + return args, nil + } + + raw := []byte(p) + err := json.Unmarshal(raw, &args) + if err == nil { + return args, nil + } + + // Fallback to parsing arguments in the legacy slice format + deprecated := map[string][]string{} + if legacyErr := json.Unmarshal(raw, &deprecated); legacyErr != nil { + return args, err + } + + args.fields = deprecatedArgs(deprecated) + return args, nil +} + +// UnmarshalJSON populates the Args from JSON encode bytes +func (args Args) UnmarshalJSON(raw []byte) error { + if len(raw) == 0 { + return nil + } + return json.Unmarshal(raw, &args.fields) +} + +// Get returns the list of values associated with the key +func (args Args) Get(key string) []string { + values := args.fields[key] if values == nil { return make([]string, 0) } @@ -131,37 +159,34 @@ func (filters Args) Get(field string) []string { return slice } -// Add adds a new value to a filter field. -func (filters Args) Add(name, value string) { - if _, ok := filters.fields[name]; ok { - filters.fields[name][value] = true +// Add a new value to the set of values +func (args Args) Add(key, value string) { + if _, ok := args.fields[key]; ok { + args.fields[key][value] = true } else { - filters.fields[name] = map[string]bool{value: true} + args.fields[key] = map[string]bool{value: true} } } -// Del removes a value from a filter field. -func (filters Args) Del(name, value string) { - if _, ok := filters.fields[name]; ok { - delete(filters.fields[name], value) - if len(filters.fields[name]) == 0 { - delete(filters.fields, name) +// Del removes a value from the set +func (args Args) Del(key, value string) { + if _, ok := args.fields[key]; ok { + delete(args.fields[key], value) + if len(args.fields[key]) == 0 { + delete(args.fields, key) } } } -// Len returns the number of fields in the arguments. -func (filters Args) Len() int { - return len(filters.fields) +// Len returns the number of keys in the mapping +func (args Args) Len() int { + return len(args.fields) } -// MatchKVList returns true if the values for the specified field matches the ones -// from the sources. -// e.g. given Args are {'label': {'label1=1','label2=1'}, 'image.name', {'ubuntu'}}, -// field is 'label' and sources are {'label1': '1', 'label2': '2'} -// it returns true. -func (filters Args) MatchKVList(field string, sources map[string]string) bool { - fieldValues := filters.fields[field] +// MatchKVList returns true if all the pairs in sources exist as key=value +// pairs in the mapping at key, or if there are no values at key. +func (args Args) MatchKVList(key string, sources map[string]string) bool { + fieldValues := args.fields[key] //do not filter if there is no filter set or cannot determine filter if len(fieldValues) == 0 { @@ -172,8 +197,8 @@ func (filters Args) MatchKVList(field string, sources map[string]string) bool { return false } - for name2match := range fieldValues { - testKV := strings.SplitN(name2match, "=", 2) + for value := range fieldValues { + testKV := strings.SplitN(value, "=", 2) v, ok := sources[testKV[0]] if !ok { @@ -187,16 +212,13 @@ func (filters Args) MatchKVList(field string, sources map[string]string) bool { return true } -// Match returns true if the values for the specified field matches the source string -// e.g. given Args are {'label': {'label1=1','label2=1'}, 'image.name', {'ubuntu'}}, -// field is 'image.name' and source is 'ubuntu' -// it returns true. -func (filters Args) Match(field, source string) bool { - if filters.ExactMatch(field, source) { +// Match returns true if any of the values at key match the source string +func (args Args) Match(field, source string) bool { + if args.ExactMatch(field, source) { return true } - fieldValues := filters.fields[field] + fieldValues := args.fields[field] for name2match := range fieldValues { match, err := regexp.MatchString(name2match, source) if err != nil { @@ -209,9 +231,9 @@ func (filters Args) Match(field, source string) bool { return false } -// ExactMatch returns true if the source matches exactly one of the filters. -func (filters Args) ExactMatch(field, source string) bool { - fieldValues, ok := filters.fields[field] +// ExactMatch returns true if the source matches exactly one of the values. +func (args Args) ExactMatch(key, source string) bool { + fieldValues, ok := args.fields[key] //do not filter if there is no filter set or cannot determine filter if !ok || len(fieldValues) == 0 { return true @@ -221,14 +243,15 @@ func (filters Args) ExactMatch(field, source string) bool { return fieldValues[source] } -// UniqueExactMatch returns true if there is only one filter and the source matches exactly this one. -func (filters Args) UniqueExactMatch(field, source string) bool { - fieldValues := filters.fields[field] +// UniqueExactMatch returns true if there is only one value and the source +// matches exactly the value. +func (args Args) UniqueExactMatch(key, source string) bool { + fieldValues := args.fields[key] //do not filter if there is no filter set or cannot determine filter if len(fieldValues) == 0 { return true } - if len(filters.fields[field]) != 1 { + if len(args.fields[key]) != 1 { return false } @@ -236,14 +259,14 @@ func (filters Args) UniqueExactMatch(field, source string) bool { return fieldValues[source] } -// FuzzyMatch returns true if the source matches exactly one of the filters, -// or the source has one of the filters as a prefix. -func (filters Args) FuzzyMatch(field, source string) bool { - if filters.ExactMatch(field, source) { +// FuzzyMatch returns true if the source matches exactly one value, or the +// source has one of the values as a prefix. +func (args Args) FuzzyMatch(key, source string) bool { + if args.ExactMatch(key, source) { return true } - fieldValues := filters.fields[field] + fieldValues := args.fields[key] for prefix := range fieldValues { if strings.HasPrefix(source, prefix) { return true @@ -252,30 +275,47 @@ func (filters Args) FuzzyMatch(field, source string) bool { return false } -// Include returns true if the name of the field to filter is in the filters. -func (filters Args) Include(field string) bool { - _, ok := filters.fields[field] +// Include returns true if the key exists in the mapping +// +// Deprecated: use Contains +func (args Args) Include(field string) bool { + _, ok := args.fields[field] return ok } -// Validate ensures that all the fields in the filter are valid. -// It returns an error as soon as it finds an invalid field. -func (filters Args) Validate(accepted map[string]bool) error { - for name := range filters.fields { +// Contains returns true if the key exists in the mapping +func (args Args) Contains(field string) bool { + _, ok := args.fields[field] + return ok +} + +type invalidFilter string + +func (e invalidFilter) Error() string { + return "Invalid filter '" + string(e) + "'" +} + +func (invalidFilter) InvalidParameter() {} + +// Validate compared the set of accepted keys against the keys in the mapping. +// An error is returned if any mapping keys are not in the accepted set. +func (args Args) Validate(accepted map[string]bool) error { + for name := range args.fields { if !accepted[name] { - return fmt.Errorf("Invalid filter '%s'", name) + return invalidFilter(name) } } return nil } -// WalkValues iterates over the list of filtered values for a field. -// It stops the iteration if it finds an error and it returns that error. -func (filters Args) WalkValues(field string, op func(value string) error) error { - if _, ok := filters.fields[field]; !ok { +// WalkValues iterates over the list of values for a key in the mapping and calls +// op() for each value. If op returns an error the iteration stops and the +// error is returned. +func (args Args) WalkValues(field string, op func(value string) error) error { + if _, ok := args.fields[field]; !ok { return nil } - for v := range filters.fields[field] { + for v := range args.fields[field] { if err := op(v); err != nil { return err } diff --git a/vendor/github.com/docker/docker/api/types/mount/mount.go b/vendor/github.com/docker/docker/api/types/mount/mount.go index 8ee16711e..3fef974df 100644 --- a/vendor/github.com/docker/docker/api/types/mount/mount.go +++ b/vendor/github.com/docker/docker/api/types/mount/mount.go @@ -1,4 +1,4 @@ -package mount +package mount // import "github.com/docker/docker/api/types/mount" import ( "os" @@ -15,6 +15,8 @@ const ( TypeVolume Type = "volume" // TypeTmpfs is the type for mounting tmpfs TypeTmpfs Type = "tmpfs" + // TypeNamedPipe is the type for mounting Windows named pipes + TypeNamedPipe Type = "npipe" ) // Mount represents a mount (volume). @@ -23,9 +25,10 @@ type Mount struct { // Source specifies the name of the mount. Depending on mount type, this // may be a volume name or a host path, or even ignored. // Source is not supported for tmpfs (must be an empty value) - Source string `json:",omitempty"` - Target string `json:",omitempty"` - ReadOnly bool `json:",omitempty"` + Source string `json:",omitempty"` + Target string `json:",omitempty"` + ReadOnly bool `json:",omitempty"` + Consistency Consistency `json:",omitempty"` BindOptions *BindOptions `json:",omitempty"` VolumeOptions *VolumeOptions `json:",omitempty"` @@ -60,6 +63,20 @@ var Propagations = []Propagation{ PropagationSlave, } +// Consistency represents the consistency requirements of a mount. +type Consistency string + +const ( + // ConsistencyFull guarantees bind mount-like consistency + ConsistencyFull Consistency = "consistent" + // ConsistencyCached mounts can cache read data and FS structure + ConsistencyCached Consistency = "cached" + // ConsistencyDelegated mounts can cache read and written data and structure + ConsistencyDelegated Consistency = "delegated" + // ConsistencyDefault provides "consistent" behavior unless overridden + ConsistencyDefault Consistency = "default" +) + // BindOptions defines options specific to mounts of type "bind". type BindOptions struct { Propagation Propagation `json:",omitempty"` diff --git a/vendor/github.com/docker/docker/api/types/network/network.go b/vendor/github.com/docker/docker/api/types/network/network.go index 7c7dbacc8..761d0b34f 100644 --- a/vendor/github.com/docker/docker/api/types/network/network.go +++ b/vendor/github.com/docker/docker/api/types/network/network.go @@ -1,4 +1,4 @@ -package network +package network // import "github.com/docker/docker/api/types/network" // Address represents an IP address type Address struct { diff --git a/vendor/github.com/docker/docker/api/types/plugin.go b/vendor/github.com/docker/docker/api/types/plugin.go index cab333e01..abae48b9a 100644 --- a/vendor/github.com/docker/docker/api/types/plugin.go +++ b/vendor/github.com/docker/docker/api/types/plugin.go @@ -121,6 +121,9 @@ type PluginConfigArgs struct { // swagger:model PluginConfigInterface type PluginConfigInterface struct { + // Protocol to use for clients connecting to the plugin. + ProtocolScheme string `json:"ProtocolScheme,omitempty"` + // socket // Required: true Socket string `json:"Socket"` diff --git a/vendor/github.com/docker/docker/api/types/plugin_responses.go b/vendor/github.com/docker/docker/api/types/plugin_responses.go index 18f743fcd..60d1fb5ad 100644 --- a/vendor/github.com/docker/docker/api/types/plugin_responses.go +++ b/vendor/github.com/docker/docker/api/types/plugin_responses.go @@ -1,4 +1,4 @@ -package types +package types // import "github.com/docker/docker/api/types" import ( "encoding/json" diff --git a/vendor/github.com/docker/docker/api/types/port.go b/vendor/github.com/docker/docker/api/types/port.go index ad52d46d5..d91234744 100644 --- a/vendor/github.com/docker/docker/api/types/port.go +++ b/vendor/github.com/docker/docker/api/types/port.go @@ -7,7 +7,7 @@ package types // swagger:model Port type Port struct { - // IP + // Host IP address that the container's port is mapped to IP string `json:"IP,omitempty"` // Port on the container diff --git a/vendor/github.com/docker/docker/api/types/registry/authenticate.go b/vendor/github.com/docker/docker/api/types/registry/authenticate.go index 42cac4430..f0a2113e4 100644 --- a/vendor/github.com/docker/docker/api/types/registry/authenticate.go +++ b/vendor/github.com/docker/docker/api/types/registry/authenticate.go @@ -1,4 +1,4 @@ -package registry +package registry // import "github.com/docker/docker/api/types/registry" // ---------------------------------------------------------------------------- // DO NOT EDIT THIS FILE diff --git a/vendor/github.com/docker/docker/api/types/registry/registry.go b/vendor/github.com/docker/docker/api/types/registry/registry.go index b98a943a1..8789ad3b3 100644 --- a/vendor/github.com/docker/docker/api/types/registry/registry.go +++ b/vendor/github.com/docker/docker/api/types/registry/registry.go @@ -1,4 +1,4 @@ -package registry +package registry // import "github.com/docker/docker/api/types/registry" import ( "encoding/json" diff --git a/vendor/github.com/docker/docker/api/types/seccomp.go b/vendor/github.com/docker/docker/api/types/seccomp.go index 7d62c9a43..67a41e1a8 100644 --- a/vendor/github.com/docker/docker/api/types/seccomp.go +++ b/vendor/github.com/docker/docker/api/types/seccomp.go @@ -1,4 +1,4 @@ -package types +package types // import "github.com/docker/docker/api/types" // Seccomp represents the config for a seccomp profile for syscall restriction. type Seccomp struct { diff --git a/vendor/github.com/docker/docker/api/types/stats.go b/vendor/github.com/docker/docker/api/types/stats.go index 7ca76a5b6..60175c061 100644 --- a/vendor/github.com/docker/docker/api/types/stats.go +++ b/vendor/github.com/docker/docker/api/types/stats.go @@ -1,6 +1,6 @@ // Package types is used for API stability in the types and response to the // consumers of the API stats endpoint. -package types +package types // import "github.com/docker/docker/api/types" import "time" diff --git a/vendor/github.com/docker/docker/api/types/strslice/strslice.go b/vendor/github.com/docker/docker/api/types/strslice/strslice.go index bad493fb8..82921cebc 100644 --- a/vendor/github.com/docker/docker/api/types/strslice/strslice.go +++ b/vendor/github.com/docker/docker/api/types/strslice/strslice.go @@ -1,4 +1,4 @@ -package strslice +package strslice // import "github.com/docker/docker/api/types/strslice" import "encoding/json" diff --git a/vendor/github.com/docker/docker/api/types/swarm/common.go b/vendor/github.com/docker/docker/api/types/swarm/common.go index 2834cf202..ef020f458 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/common.go +++ b/vendor/github.com/docker/docker/api/types/swarm/common.go @@ -1,4 +1,4 @@ -package swarm +package swarm // import "github.com/docker/docker/api/types/swarm" import "time" diff --git a/vendor/github.com/docker/docker/api/types/swarm/config.go b/vendor/github.com/docker/docker/api/types/swarm/config.go index 0fb021ce9..a1555cf43 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/config.go +++ b/vendor/github.com/docker/docker/api/types/swarm/config.go @@ -1,4 +1,4 @@ -package swarm +package swarm // import "github.com/docker/docker/api/types/swarm" import "os" @@ -13,6 +13,10 @@ type Config struct { type ConfigSpec struct { Annotations Data []byte `json:",omitempty"` + + // Templating controls whether and how to evaluate the config payload as + // a template. If it is not set, no templating is used. + Templating *Driver `json:",omitempty"` } // ConfigReferenceFileTarget is a file target in a config reference diff --git a/vendor/github.com/docker/docker/api/types/swarm/container.go b/vendor/github.com/docker/docker/api/types/swarm/container.go index 734236c4b..151211ff5 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/container.go +++ b/vendor/github.com/docker/docker/api/types/swarm/container.go @@ -1,4 +1,4 @@ -package swarm +package swarm // import "github.com/docker/docker/api/types/swarm" import ( "time" @@ -55,6 +55,7 @@ type ContainerSpec struct { User string `json:",omitempty"` Groups []string `json:",omitempty"` Privileges *Privileges `json:",omitempty"` + Init *bool `json:",omitempty"` StopSignal string `json:",omitempty"` TTY bool `json:",omitempty"` OpenStdin bool `json:",omitempty"` diff --git a/vendor/github.com/docker/docker/api/types/swarm/network.go b/vendor/github.com/docker/docker/api/types/swarm/network.go index 97c484e14..98ef3284d 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/network.go +++ b/vendor/github.com/docker/docker/api/types/swarm/network.go @@ -1,4 +1,4 @@ -package swarm +package swarm // import "github.com/docker/docker/api/types/swarm" import ( "github.com/docker/docker/api/types/network" @@ -62,6 +62,8 @@ const ( PortConfigProtocolTCP PortConfigProtocol = "tcp" // PortConfigProtocolUDP UDP PortConfigProtocolUDP PortConfigProtocol = "udp" + // PortConfigProtocolSCTP SCTP + PortConfigProtocolSCTP PortConfigProtocol = "sctp" ) // EndpointVirtualIP represents the virtual ip of a port. diff --git a/vendor/github.com/docker/docker/api/types/swarm/node.go b/vendor/github.com/docker/docker/api/types/swarm/node.go index 28c6851e9..1e30f5fa1 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/node.go +++ b/vendor/github.com/docker/docker/api/types/swarm/node.go @@ -1,4 +1,4 @@ -package swarm +package swarm // import "github.com/docker/docker/api/types/swarm" // Node represents a node. type Node struct { diff --git a/vendor/github.com/docker/docker/api/types/swarm/runtime.go b/vendor/github.com/docker/docker/api/types/swarm/runtime.go index c4c731dc8..0c77403cc 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/runtime.go +++ b/vendor/github.com/docker/docker/api/types/swarm/runtime.go @@ -1,4 +1,4 @@ -package swarm +package swarm // import "github.com/docker/docker/api/types/swarm" // RuntimeType is the type of runtime used for the TaskSpec type RuntimeType string @@ -11,9 +11,17 @@ const ( RuntimeContainer RuntimeType = "container" // RuntimePlugin is the plugin based runtime RuntimePlugin RuntimeType = "plugin" + // RuntimeNetworkAttachment is the network attachment runtime + RuntimeNetworkAttachment RuntimeType = "attachment" // RuntimeURLContainer is the proto url for the container type RuntimeURLContainer RuntimeURL = "types.docker.com/RuntimeContainer" // RuntimeURLPlugin is the proto url for the plugin type RuntimeURLPlugin RuntimeURL = "types.docker.com/RuntimePlugin" ) + +// NetworkAttachmentSpec represents the runtime spec type for network +// attachment tasks +type NetworkAttachmentSpec struct { + ContainerID string +} diff --git a/vendor/github.com/docker/docker/api/types/swarm/runtime/gen.go b/vendor/github.com/docker/docker/api/types/swarm/runtime/gen.go index 47ae234ef..98c2806c3 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/runtime/gen.go +++ b/vendor/github.com/docker/docker/api/types/swarm/runtime/gen.go @@ -1,3 +1,3 @@ //go:generate protoc -I . --gogofast_out=import_path=github.com/docker/docker/api/types/swarm/runtime:. plugin.proto -package runtime +package runtime // import "github.com/docker/docker/api/types/swarm/runtime" diff --git a/vendor/github.com/docker/docker/api/types/swarm/runtime/plugin.proto b/vendor/github.com/docker/docker/api/types/swarm/runtime/plugin.proto index 06eb7ba65..6d63b7783 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/runtime/plugin.proto +++ b/vendor/github.com/docker/docker/api/types/swarm/runtime/plugin.proto @@ -1,5 +1,7 @@ syntax = "proto3"; +option go_package = "github.com/docker/docker/api/types/swarm/runtime;runtime"; + // PluginSpec defines the base payload which clients can specify for creating // a service with the plugin runtime. message PluginSpec { diff --git a/vendor/github.com/docker/docker/api/types/swarm/secret.go b/vendor/github.com/docker/docker/api/types/swarm/secret.go index f9b1e9266..d5213ec98 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/secret.go +++ b/vendor/github.com/docker/docker/api/types/swarm/secret.go @@ -1,4 +1,4 @@ -package swarm +package swarm // import "github.com/docker/docker/api/types/swarm" import "os" @@ -14,6 +14,10 @@ type SecretSpec struct { Annotations Data []byte `json:",omitempty"` Driver *Driver `json:",omitempty"` // name of the secrets driver used to fetch the secret's value from an external secret store + + // Templating controls whether and how to evaluate the secret payload as + // a template. If it is not set, no templating is used. + Templating *Driver `json:",omitempty"` } // SecretReferenceFileTarget is a file target in a secret reference diff --git a/vendor/github.com/docker/docker/api/types/swarm/service.go b/vendor/github.com/docker/docker/api/types/swarm/service.go index fa31a7ec8..abf192e75 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/service.go +++ b/vendor/github.com/docker/docker/api/types/swarm/service.go @@ -1,4 +1,4 @@ -package swarm +package swarm // import "github.com/docker/docker/api/types/swarm" import "time" diff --git a/vendor/github.com/docker/docker/api/types/swarm/swarm.go b/vendor/github.com/docker/docker/api/types/swarm/swarm.go index b65fa86da..1b111d725 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/swarm.go +++ b/vendor/github.com/docker/docker/api/types/swarm/swarm.go @@ -1,4 +1,4 @@ -package swarm +package swarm // import "github.com/docker/docker/api/types/swarm" import "time" diff --git a/vendor/github.com/docker/docker/api/types/swarm/task.go b/vendor/github.com/docker/docker/api/types/swarm/task.go index ff11b07e7..b35605d12 100644 --- a/vendor/github.com/docker/docker/api/types/swarm/task.go +++ b/vendor/github.com/docker/docker/api/types/swarm/task.go @@ -1,4 +1,4 @@ -package swarm +package swarm // import "github.com/docker/docker/api/types/swarm" import ( "time" @@ -36,6 +36,10 @@ const ( TaskStateFailed TaskState = "failed" // TaskStateRejected REJECTED TaskStateRejected TaskState = "rejected" + // TaskStateRemove REMOVE + TaskStateRemove TaskState = "remove" + // TaskStateOrphaned ORPHANED + TaskStateOrphaned TaskState = "orphaned" ) // Task represents a task. @@ -56,10 +60,13 @@ type Task struct { // TaskSpec represents the spec of a task. type TaskSpec struct { - // ContainerSpec and PluginSpec are mutually exclusive. - // PluginSpec will only be used when the `Runtime` field is set to `plugin` - ContainerSpec *ContainerSpec `json:",omitempty"` - PluginSpec *runtime.PluginSpec `json:",omitempty"` + // ContainerSpec, NetworkAttachmentSpec, and PluginSpec are mutually exclusive. + // PluginSpec is only used when the `Runtime` field is set to `plugin` + // NetworkAttachmentSpec is used if the `Runtime` field is set to + // `attachment`. + ContainerSpec *ContainerSpec `json:",omitempty"` + PluginSpec *runtime.PluginSpec `json:",omitempty"` + NetworkAttachmentSpec *NetworkAttachmentSpec `json:",omitempty"` Resources *ResourceRequirements `json:",omitempty"` RestartPolicy *RestartPolicy `json:",omitempty"` @@ -162,19 +169,19 @@ const ( // TaskStatus represents the status of a task. type TaskStatus struct { - Timestamp time.Time `json:",omitempty"` - State TaskState `json:",omitempty"` - Message string `json:",omitempty"` - Err string `json:",omitempty"` - ContainerStatus ContainerStatus `json:",omitempty"` - PortStatus PortStatus `json:",omitempty"` + Timestamp time.Time `json:",omitempty"` + State TaskState `json:",omitempty"` + Message string `json:",omitempty"` + Err string `json:",omitempty"` + ContainerStatus *ContainerStatus `json:",omitempty"` + PortStatus PortStatus `json:",omitempty"` } // ContainerStatus represents the status of a container. type ContainerStatus struct { - ContainerID string `json:",omitempty"` - PID int `json:",omitempty"` - ExitCode int `json:",omitempty"` + ContainerID string + PID int + ExitCode int } // PortStatus represents the port status of a task's host ports whose diff --git a/vendor/github.com/docker/docker/api/types/types.go b/vendor/github.com/docker/docker/api/types/types.go index f7ac77297..06c0ca3a6 100644 --- a/vendor/github.com/docker/docker/api/types/types.go +++ b/vendor/github.com/docker/docker/api/types/types.go @@ -1,4 +1,4 @@ -package types +package types // import "github.com/docker/docker/api/types" import ( "errors" @@ -107,9 +107,21 @@ type Ping struct { Experimental bool } +// ComponentVersion describes the version information for a specific component. +type ComponentVersion struct { + Name string + Version string + Details map[string]string `json:",omitempty"` +} + // Version contains response of Engine API: // GET "/version" type Version struct { + Platform struct{ Name string } `json:",omitempty"` + Components []ComponentVersion `json:",omitempty"` + + // The following fields are deprecated, they relate to the Engine component and are kept for backwards compatibility + Version string APIVersion string `json:"ApiVersion"` MinAPIVersion string `json:"MinAPIVersion,omitempty"` @@ -500,7 +512,8 @@ type DiskUsage struct { Images []*ImageSummary Containers []*Container Volumes []*Volume - BuilderSize int64 + BuildCache []*BuildCache + BuilderSize int64 // deprecated } // ContainersPruneReport contains the response for Engine API: @@ -573,3 +586,17 @@ type PushResult struct { type BuildResult struct { ID string } + +// BuildCache contains information about a build cache record +type BuildCache struct { + ID string + Mutable bool + InUse bool + Size int64 + + CreatedAt time.Time + LastUsedAt *time.Time + UsageCount int + Parent string + Description string +} diff --git a/vendor/github.com/docker/docker/api/types/versions/compare.go b/vendor/github.com/docker/docker/api/types/versions/compare.go index 611d4fed6..8ccb0aa92 100644 --- a/vendor/github.com/docker/docker/api/types/versions/compare.go +++ b/vendor/github.com/docker/docker/api/types/versions/compare.go @@ -1,4 +1,4 @@ -package versions +package versions // import "github.com/docker/docker/api/types/versions" import ( "strconv" diff --git a/vendor/github.com/docker/docker/cli/config/configfile/file.go b/vendor/github.com/docker/docker/cli/config/configfile/file.go deleted file mode 100644 index 39097133a..000000000 --- a/vendor/github.com/docker/docker/cli/config/configfile/file.go +++ /dev/null @@ -1,183 +0,0 @@ -package configfile - -import ( - "encoding/base64" - "encoding/json" - "fmt" - "io" - "io/ioutil" - "os" - "path/filepath" - "strings" - - "github.com/docker/docker/api/types" -) - -const ( - // This constant is only used for really old config files when the - // URL wasn't saved as part of the config file and it was just - // assumed to be this value. - defaultIndexserver = "https://index.docker.io/v1/" -) - -// ConfigFile ~/.docker/config.json file info -type ConfigFile struct { - AuthConfigs map[string]types.AuthConfig `json:"auths"` - HTTPHeaders map[string]string `json:"HttpHeaders,omitempty"` - PsFormat string `json:"psFormat,omitempty"` - ImagesFormat string `json:"imagesFormat,omitempty"` - NetworksFormat string `json:"networksFormat,omitempty"` - VolumesFormat string `json:"volumesFormat,omitempty"` - StatsFormat string `json:"statsFormat,omitempty"` - DetachKeys string `json:"detachKeys,omitempty"` - CredentialsStore string `json:"credsStore,omitempty"` - CredentialHelpers map[string]string `json:"credHelpers,omitempty"` - Filename string `json:"-"` // Note: for internal use only - ServiceInspectFormat string `json:"serviceInspectFormat,omitempty"` -} - -// LegacyLoadFromReader reads the non-nested configuration data given and sets up the -// auth config information with given directory and populates the receiver object -func (configFile *ConfigFile) LegacyLoadFromReader(configData io.Reader) error { - b, err := ioutil.ReadAll(configData) - if err != nil { - return err - } - - if err := json.Unmarshal(b, &configFile.AuthConfigs); err != nil { - arr := strings.Split(string(b), "\n") - if len(arr) < 2 { - return fmt.Errorf("The Auth config file is empty") - } - authConfig := types.AuthConfig{} - origAuth := strings.Split(arr[0], " = ") - if len(origAuth) != 2 { - return fmt.Errorf("Invalid Auth config file") - } - authConfig.Username, authConfig.Password, err = decodeAuth(origAuth[1]) - if err != nil { - return err - } - authConfig.ServerAddress = defaultIndexserver - configFile.AuthConfigs[defaultIndexserver] = authConfig - } else { - for k, authConfig := range configFile.AuthConfigs { - authConfig.Username, authConfig.Password, err = decodeAuth(authConfig.Auth) - if err != nil { - return err - } - authConfig.Auth = "" - authConfig.ServerAddress = k - configFile.AuthConfigs[k] = authConfig - } - } - return nil -} - -// LoadFromReader reads the configuration data given and sets up the auth config -// information with given directory and populates the receiver object -func (configFile *ConfigFile) LoadFromReader(configData io.Reader) error { - if err := json.NewDecoder(configData).Decode(&configFile); err != nil { - return err - } - var err error - for addr, ac := range configFile.AuthConfigs { - ac.Username, ac.Password, err = decodeAuth(ac.Auth) - if err != nil { - return err - } - ac.Auth = "" - ac.ServerAddress = addr - configFile.AuthConfigs[addr] = ac - } - return nil -} - -// ContainsAuth returns whether there is authentication configured -// in this file or not. -func (configFile *ConfigFile) ContainsAuth() bool { - return configFile.CredentialsStore != "" || - len(configFile.CredentialHelpers) > 0 || - len(configFile.AuthConfigs) > 0 -} - -// SaveToWriter encodes and writes out all the authorization information to -// the given writer -func (configFile *ConfigFile) SaveToWriter(writer io.Writer) error { - // Encode sensitive data into a new/temp struct - tmpAuthConfigs := make(map[string]types.AuthConfig, len(configFile.AuthConfigs)) - for k, authConfig := range configFile.AuthConfigs { - authCopy := authConfig - // encode and save the authstring, while blanking out the original fields - authCopy.Auth = encodeAuth(&authCopy) - authCopy.Username = "" - authCopy.Password = "" - authCopy.ServerAddress = "" - tmpAuthConfigs[k] = authCopy - } - - saveAuthConfigs := configFile.AuthConfigs - configFile.AuthConfigs = tmpAuthConfigs - defer func() { configFile.AuthConfigs = saveAuthConfigs }() - - data, err := json.MarshalIndent(configFile, "", "\t") - if err != nil { - return err - } - _, err = writer.Write(data) - return err -} - -// Save encodes and writes out all the authorization information -func (configFile *ConfigFile) Save() error { - if configFile.Filename == "" { - return fmt.Errorf("Can't save config with empty filename") - } - - if err := os.MkdirAll(filepath.Dir(configFile.Filename), 0700); err != nil { - return err - } - f, err := os.OpenFile(configFile.Filename, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0600) - if err != nil { - return err - } - defer f.Close() - return configFile.SaveToWriter(f) -} - -// encodeAuth creates a base64 encoded string to containing authorization information -func encodeAuth(authConfig *types.AuthConfig) string { - if authConfig.Username == "" && authConfig.Password == "" { - return "" - } - - authStr := authConfig.Username + ":" + authConfig.Password - msg := []byte(authStr) - encoded := make([]byte, base64.StdEncoding.EncodedLen(len(msg))) - base64.StdEncoding.Encode(encoded, msg) - return string(encoded) -} - -// decodeAuth decodes a base64 encoded string and returns username and password -func decodeAuth(authStr string) (string, string, error) { - if authStr == "" { - return "", "", nil - } - - decLen := base64.StdEncoding.DecodedLen(len(authStr)) - decoded := make([]byte, decLen) - authByte := []byte(authStr) - n, err := base64.StdEncoding.Decode(decoded, authByte) - if err != nil { - return "", "", err - } - if n > decLen { - return "", "", fmt.Errorf("Something went wrong decoding auth config") - } - arr := strings.SplitN(string(decoded), ":", 2) - if len(arr) != 2 { - return "", "", fmt.Errorf("Invalid auth configuration file") - } - password := strings.Trim(arr[1], "\x00") - return arr[0], password, nil -} diff --git a/vendor/github.com/docker/docker/contrib/docker-engine-selinux/LICENSE b/vendor/github.com/docker/docker/contrib/docker-engine-selinux/LICENSE deleted file mode 100644 index 5b6e7c66c..000000000 --- a/vendor/github.com/docker/docker/contrib/docker-engine-selinux/LICENSE +++ /dev/null @@ -1,340 +0,0 @@ - GNU GENERAL PUBLIC LICENSE - Version 2, June 1991 - - Copyright (C) 1989, 1991 Free Software Foundation, Inc. - 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA - Everyone is permitted to copy and distribute verbatim copies - of this license document, but changing it is not allowed. - - Preamble - - The licenses for most software are designed to take away your -freedom to share and change it. By contrast, the GNU General Public -License is intended to guarantee your freedom to share and change free -software--to make sure the software is free for all its users. This -General Public License applies to most of the Free Software -Foundation's software and to any other program whose authors commit to -using it. (Some other Free Software Foundation software is covered by -the GNU Library General Public License instead.) You can apply it to -your programs, too. - - When we speak of free software, we are referring to freedom, not -price. Our General Public Licenses are designed to make sure that you -have the freedom to distribute copies of free software (and charge for -this service if you wish), that you receive source code or can get it -if you want it, that you can change the software or use pieces of it -in new free programs; and that you know you can do these things. - - To protect your rights, we need to make restrictions that forbid -anyone to deny you these rights or to ask you to surrender the rights. -These restrictions translate to certain responsibilities for you if you -distribute copies of the software, or if you modify it. - - For example, if you distribute copies of such a program, whether -gratis or for a fee, you must give the recipients all the rights that -you have. You must make sure that they, too, receive or can get the -source code. And you must show them these terms so they know their -rights. - - We protect your rights with two steps: (1) copyright the software, and -(2) offer you this license which gives you legal permission to copy, -distribute and/or modify the software. - - Also, for each author's protection and ours, we want to make certain -that everyone understands that there is no warranty for this free -software. If the software is modified by someone else and passed on, we -want its recipients to know that what they have is not the original, so -that any problems introduced by others will not reflect on the original -authors' reputations. - - Finally, any free program is threatened constantly by software -patents. We wish to avoid the danger that redistributors of a free -program will individually obtain patent licenses, in effect making the -program proprietary. To prevent this, we have made it clear that any -patent must be licensed for everyone's free use or not licensed at all. - - The precise terms and conditions for copying, distribution and -modification follow. - - GNU GENERAL PUBLIC LICENSE - TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION - - 0. This License applies to any program or other work which contains -a notice placed by the copyright holder saying it may be distributed -under the terms of this General Public License. The "Program", below, -refers to any such program or work, and a "work based on the Program" -means either the Program or any derivative work under copyright law: -that is to say, a work containing the Program or a portion of it, -either verbatim or with modifications and/or translated into another -language. (Hereinafter, translation is included without limitation in -the term "modification".) Each licensee is addressed as "you". - -Activities other than copying, distribution and modification are not -covered by this License; they are outside its scope. The act of -running the Program is not restricted, and the output from the Program -is covered only if its contents constitute a work based on the -Program (independent of having been made by running the Program). -Whether that is true depends on what the Program does. - - 1. You may copy and distribute verbatim copies of the Program's -source code as you receive it, in any medium, provided that you -conspicuously and appropriately publish on each copy an appropriate -copyright notice and disclaimer of warranty; keep intact all the -notices that refer to this License and to the absence of any warranty; -and give any other recipients of the Program a copy of this License -along with the Program. - -You may charge a fee for the physical act of transferring a copy, and -you may at your option offer warranty protection in exchange for a fee. - - 2. You may modify your copy or copies of the Program or any portion -of it, thus forming a work based on the Program, and copy and -distribute such modifications or work under the terms of Section 1 -above, provided that you also meet all of these conditions: - - a) You must cause the modified files to carry prominent notices - stating that you changed the files and the date of any change. - - b) You must cause any work that you distribute or publish, that in - whole or in part contains or is derived from the Program or any - part thereof, to be licensed as a whole at no charge to all third - parties under the terms of this License. - - c) If the modified program normally reads commands interactively - when run, you must cause it, when started running for such - interactive use in the most ordinary way, to print or display an - announcement including an appropriate copyright notice and a - notice that there is no warranty (or else, saying that you provide - a warranty) and that users may redistribute the program under - these conditions, and telling the user how to view a copy of this - License. (Exception: if the Program itself is interactive but - does not normally print such an announcement, your work based on - the Program is not required to print an announcement.) - -These requirements apply to the modified work as a whole. If -identifiable sections of that work are not derived from the Program, -and can be reasonably considered independent and separate works in -themselves, then this License, and its terms, do not apply to those -sections when you distribute them as separate works. But when you -distribute the same sections as part of a whole which is a work based -on the Program, the distribution of the whole must be on the terms of -this License, whose permissions for other licensees extend to the -entire whole, and thus to each and every part regardless of who wrote it. - -Thus, it is not the intent of this section to claim rights or contest -your rights to work written entirely by you; rather, the intent is to -exercise the right to control the distribution of derivative or -collective works based on the Program. - -In addition, mere aggregation of another work not based on the Program -with the Program (or with a work based on the Program) on a volume of -a storage or distribution medium does not bring the other work under -the scope of this License. - - 3. You may copy and distribute the Program (or a work based on it, -under Section 2) in object code or executable form under the terms of -Sections 1 and 2 above provided that you also do one of the following: - - a) Accompany it with the complete corresponding machine-readable - source code, which must be distributed under the terms of Sections - 1 and 2 above on a medium customarily used for software interchange; or, - - b) Accompany it with a written offer, valid for at least three - years, to give any third party, for a charge no more than your - cost of physically performing source distribution, a complete - machine-readable copy of the corresponding source code, to be - distributed under the terms of Sections 1 and 2 above on a medium - customarily used for software interchange; or, - - c) Accompany it with the information you received as to the offer - to distribute corresponding source code. (This alternative is - allowed only for noncommercial distribution and only if you - received the program in object code or executable form with such - an offer, in accord with Subsection b above.) - -The source code for a work means the preferred form of the work for -making modifications to it. For an executable work, complete source -code means all the source code for all modules it contains, plus any -associated interface definition files, plus the scripts used to -control compilation and installation of the executable. However, as a -special exception, the source code distributed need not include -anything that is normally distributed (in either source or binary -form) with the major components (compiler, kernel, and so on) of the -operating system on which the executable runs, unless that component -itself accompanies the executable. - -If distribution of executable or object code is made by offering -access to copy from a designated place, then offering equivalent -access to copy the source code from the same place counts as -distribution of the source code, even though third parties are not -compelled to copy the source along with the object code. - - 4. You may not copy, modify, sublicense, or distribute the Program -except as expressly provided under this License. Any attempt -otherwise to copy, modify, sublicense or distribute the Program is -void, and will automatically terminate your rights under this License. -However, parties who have received copies, or rights, from you under -this License will not have their licenses terminated so long as such -parties remain in full compliance. - - 5. You are not required to accept this License, since you have not -signed it. However, nothing else grants you permission to modify or -distribute the Program or its derivative works. These actions are -prohibited by law if you do not accept this License. Therefore, by -modifying or distributing the Program (or any work based on the -Program), you indicate your acceptance of this License to do so, and -all its terms and conditions for copying, distributing or modifying -the Program or works based on it. - - 6. Each time you redistribute the Program (or any work based on the -Program), the recipient automatically receives a license from the -original licensor to copy, distribute or modify the Program subject to -these terms and conditions. You may not impose any further -restrictions on the recipients' exercise of the rights granted herein. -You are not responsible for enforcing compliance by third parties to -this License. - - 7. If, as a consequence of a court judgment or allegation of patent -infringement or for any other reason (not limited to patent issues), -conditions are imposed on you (whether by court order, agreement or -otherwise) that contradict the conditions of this License, they do not -excuse you from the conditions of this License. If you cannot -distribute so as to satisfy simultaneously your obligations under this -License and any other pertinent obligations, then as a consequence you -may not distribute the Program at all. For example, if a patent -license would not permit royalty-free redistribution of the Program by -all those who receive copies directly or indirectly through you, then -the only way you could satisfy both it and this License would be to -refrain entirely from distribution of the Program. - -If any portion of this section is held invalid or unenforceable under -any particular circumstance, the balance of the section is intended to -apply and the section as a whole is intended to apply in other -circumstances. - -It is not the purpose of this section to induce you to infringe any -patents or other property right claims or to contest validity of any -such claims; this section has the sole purpose of protecting the -integrity of the free software distribution system, which is -implemented by public license practices. Many people have made -generous contributions to the wide range of software distributed -through that system in reliance on consistent application of that -system; it is up to the author/donor to decide if he or she is willing -to distribute software through any other system and a licensee cannot -impose that choice. - -This section is intended to make thoroughly clear what is believed to -be a consequence of the rest of this License. - - 8. If the distribution and/or use of the Program is restricted in -certain countries either by patents or by copyrighted interfaces, the -original copyright holder who places the Program under this License -may add an explicit geographical distribution limitation excluding -those countries, so that distribution is permitted only in or among -countries not thus excluded. In such case, this License incorporates -the limitation as if written in the body of this License. - - 9. The Free Software Foundation may publish revised and/or new versions -of the General Public License from time to time. Such new versions will -be similar in spirit to the present version, but may differ in detail to -address new problems or concerns. - -Each version is given a distinguishing version number. If the Program -specifies a version number of this License which applies to it and "any -later version", you have the option of following the terms and conditions -either of that version or of any later version published by the Free -Software Foundation. If the Program does not specify a version number of -this License, you may choose any version ever published by the Free Software -Foundation. - - 10. If you wish to incorporate parts of the Program into other free -programs whose distribution conditions are different, write to the author -to ask for permission. For software which is copyrighted by the Free -Software Foundation, write to the Free Software Foundation; we sometimes -make exceptions for this. Our decision will be guided by the two goals -of preserving the free status of all derivatives of our free software and -of promoting the sharing and reuse of software generally. - - NO WARRANTY - - 11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY -FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN -OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES -PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED -OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF -MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS -TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE -PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, -REPAIR OR CORRECTION. - - 12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING -WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR -REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, -INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING -OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED -TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY -YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER -PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE -POSSIBILITY OF SUCH DAMAGES. - - END OF TERMS AND CONDITIONS - - How to Apply These Terms to Your New Programs - - If you develop a new program, and you want it to be of the greatest -possible use to the public, the best way to achieve this is to make it -free software which everyone can redistribute and change under these terms. - - To do so, attach the following notices to the program. It is safest -to attach them to the start of each source file to most effectively -convey the exclusion of warranty; and each file should have at least -the "copyright" line and a pointer to where the full notice is found. - - - Copyright (C) - - This program is free software; you can redistribute it and/or modify - it under the terms of the GNU General Public License as published by - the Free Software Foundation; either version 2 of the License, or - (at your option) any later version. - - This program is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - GNU General Public License for more details. - - You should have received a copy of the GNU General Public License - along with this program; if not, write to the Free Software - Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA - - -Also add information on how to contact you by electronic and paper mail. - -If the program is interactive, make it output a short notice like this -when it starts in an interactive mode: - - Gnomovision version 69, Copyright (C) year name of author - Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'. - This is free software, and you are welcome to redistribute it - under certain conditions; type `show c' for details. - -The hypothetical commands `show w' and `show c' should show the appropriate -parts of the General Public License. Of course, the commands you use may -be called something other than `show w' and `show c'; they could even be -mouse-clicks or menu items--whatever suits your program. - -You should also get your employer (if you work as a programmer) or your -school, if any, to sign a "copyright disclaimer" for the program, if -necessary. Here is a sample; alter the names: - - Yoyodyne, Inc., hereby disclaims all copyright interest in the program - `Gnomovision' (which makes passes at compilers) written by James Hacker. - - , 1 April 1989 - Ty Coon, President of Vice - -This General Public License does not permit incorporating your program into -proprietary programs. If your program is a subroutine library, you may -consider it more useful to permit linking proprietary applications with the -library. If this is what you want to do, use the GNU Library General -Public License instead of this License. diff --git a/vendor/github.com/docker/docker/contrib/syntax/vim/LICENSE b/vendor/github.com/docker/docker/contrib/syntax/vim/LICENSE deleted file mode 100644 index e67cdabd2..000000000 --- a/vendor/github.com/docker/docker/contrib/syntax/vim/LICENSE +++ /dev/null @@ -1,22 +0,0 @@ -Copyright (c) 2013 Honza Pokorny -All rights reserved. - -Redistribution and use in source and binary forms, with or without -modification, are permitted provided that the following conditions are met: - -1. Redistributions of source code must retain the above copyright - notice, this list of conditions and the following disclaimer. -2. Redistributions in binary form must reproduce the above copyright - notice, this list of conditions and the following disclaimer in the - documentation and/or other materials provided with the distribution. - -THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND -ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED -WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE -DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR -ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES -(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; -LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND -ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS -SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. diff --git a/vendor/github.com/docker/docker/errdefs/defs.go b/vendor/github.com/docker/docker/errdefs/defs.go new file mode 100644 index 000000000..e6a2275b2 --- /dev/null +++ b/vendor/github.com/docker/docker/errdefs/defs.go @@ -0,0 +1,74 @@ +package errdefs // import "github.com/docker/docker/errdefs" + +// ErrNotFound signals that the requested object doesn't exist +type ErrNotFound interface { + NotFound() +} + +// ErrInvalidParameter signals that the user input is invalid +type ErrInvalidParameter interface { + InvalidParameter() +} + +// ErrConflict signals that some internal state conflicts with the requested action and can't be performed. +// A change in state should be able to clear this error. +type ErrConflict interface { + Conflict() +} + +// ErrUnauthorized is used to signify that the user is not authorized to perform a specific action +type ErrUnauthorized interface { + Unauthorized() +} + +// ErrUnavailable signals that the requested action/subsystem is not available. +type ErrUnavailable interface { + Unavailable() +} + +// ErrForbidden signals that the requested action cannot be performed under any circumstances. +// When a ErrForbidden is returned, the caller should never retry the action. +type ErrForbidden interface { + Forbidden() +} + +// ErrSystem signals that some internal error occurred. +// An example of this would be a failed mount request. +type ErrSystem interface { + System() +} + +// ErrNotModified signals that an action can't be performed because it's already in the desired state +type ErrNotModified interface { + NotModified() +} + +// ErrAlreadyExists is a special case of ErrConflict which signals that the desired object already exists +type ErrAlreadyExists interface { + AlreadyExists() +} + +// ErrNotImplemented signals that the requested action/feature is not implemented on the system as configured. +type ErrNotImplemented interface { + NotImplemented() +} + +// ErrUnknown signals that the kind of error that occurred is not known. +type ErrUnknown interface { + Unknown() +} + +// ErrCancelled signals that the action was cancelled. +type ErrCancelled interface { + Cancelled() +} + +// ErrDeadline signals that the deadline was reached before the action completed. +type ErrDeadline interface { + DeadlineExceeded() +} + +// ErrDataLoss indicates that data was lost or there is data corruption. +type ErrDataLoss interface { + DataLoss() +} diff --git a/vendor/github.com/docker/docker/errdefs/doc.go b/vendor/github.com/docker/docker/errdefs/doc.go new file mode 100644 index 000000000..c211f174f --- /dev/null +++ b/vendor/github.com/docker/docker/errdefs/doc.go @@ -0,0 +1,8 @@ +// Package errdefs defines a set of error interfaces that packages should use for communicating classes of errors. +// Errors that cross the package boundary should implement one (and only one) of these interfaces. +// +// Packages should not reference these interfaces directly, only implement them. +// To check if a particular error implements one of these interfaces, there are helper +// functions provided (e.g. `Is`) which can be used rather than asserting the interfaces directly. +// If you must assert on these interfaces, be sure to check the causal chain (`err.Cause()`). +package errdefs // import "github.com/docker/docker/errdefs" diff --git a/vendor/github.com/docker/docker/errdefs/helpers.go b/vendor/github.com/docker/docker/errdefs/helpers.go new file mode 100644 index 000000000..6169c2bc6 --- /dev/null +++ b/vendor/github.com/docker/docker/errdefs/helpers.go @@ -0,0 +1,240 @@ +package errdefs // import "github.com/docker/docker/errdefs" + +import "context" + +type errNotFound struct{ error } + +func (errNotFound) NotFound() {} + +func (e errNotFound) Cause() error { + return e.error +} + +// NotFound is a helper to create an error of the class with the same name from any error type +func NotFound(err error) error { + if err == nil { + return nil + } + return errNotFound{err} +} + +type errInvalidParameter struct{ error } + +func (errInvalidParameter) InvalidParameter() {} + +func (e errInvalidParameter) Cause() error { + return e.error +} + +// InvalidParameter is a helper to create an error of the class with the same name from any error type +func InvalidParameter(err error) error { + if err == nil { + return nil + } + return errInvalidParameter{err} +} + +type errConflict struct{ error } + +func (errConflict) Conflict() {} + +func (e errConflict) Cause() error { + return e.error +} + +// Conflict is a helper to create an error of the class with the same name from any error type +func Conflict(err error) error { + if err == nil { + return nil + } + return errConflict{err} +} + +type errUnauthorized struct{ error } + +func (errUnauthorized) Unauthorized() {} + +func (e errUnauthorized) Cause() error { + return e.error +} + +// Unauthorized is a helper to create an error of the class with the same name from any error type +func Unauthorized(err error) error { + if err == nil { + return nil + } + return errUnauthorized{err} +} + +type errUnavailable struct{ error } + +func (errUnavailable) Unavailable() {} + +func (e errUnavailable) Cause() error { + return e.error +} + +// Unavailable is a helper to create an error of the class with the same name from any error type +func Unavailable(err error) error { + return errUnavailable{err} +} + +type errForbidden struct{ error } + +func (errForbidden) Forbidden() {} + +func (e errForbidden) Cause() error { + return e.error +} + +// Forbidden is a helper to create an error of the class with the same name from any error type +func Forbidden(err error) error { + if err == nil { + return nil + } + return errForbidden{err} +} + +type errSystem struct{ error } + +func (errSystem) System() {} + +func (e errSystem) Cause() error { + return e.error +} + +// System is a helper to create an error of the class with the same name from any error type +func System(err error) error { + if err == nil { + return nil + } + return errSystem{err} +} + +type errNotModified struct{ error } + +func (errNotModified) NotModified() {} + +func (e errNotModified) Cause() error { + return e.error +} + +// NotModified is a helper to create an error of the class with the same name from any error type +func NotModified(err error) error { + if err == nil { + return nil + } + return errNotModified{err} +} + +type errAlreadyExists struct{ error } + +func (errAlreadyExists) AlreadyExists() {} + +func (e errAlreadyExists) Cause() error { + return e.error +} + +// AlreadyExists is a helper to create an error of the class with the same name from any error type +func AlreadyExists(err error) error { + if err == nil { + return nil + } + return errAlreadyExists{err} +} + +type errNotImplemented struct{ error } + +func (errNotImplemented) NotImplemented() {} + +func (e errNotImplemented) Cause() error { + return e.error +} + +// NotImplemented is a helper to create an error of the class with the same name from any error type +func NotImplemented(err error) error { + if err == nil { + return nil + } + return errNotImplemented{err} +} + +type errUnknown struct{ error } + +func (errUnknown) Unknown() {} + +func (e errUnknown) Cause() error { + return e.error +} + +// Unknown is a helper to create an error of the class with the same name from any error type +func Unknown(err error) error { + if err == nil { + return nil + } + return errUnknown{err} +} + +type errCancelled struct{ error } + +func (errCancelled) Cancelled() {} + +func (e errCancelled) Cause() error { + return e.error +} + +// Cancelled is a helper to create an error of the class with the same name from any error type +func Cancelled(err error) error { + if err == nil { + return nil + } + return errCancelled{err} +} + +type errDeadline struct{ error } + +func (errDeadline) DeadlineExceeded() {} + +func (e errDeadline) Cause() error { + return e.error +} + +// Deadline is a helper to create an error of the class with the same name from any error type +func Deadline(err error) error { + if err == nil { + return nil + } + return errDeadline{err} +} + +type errDataLoss struct{ error } + +func (errDataLoss) DataLoss() {} + +func (e errDataLoss) Cause() error { + return e.error +} + +// DataLoss is a helper to create an error of the class with the same name from any error type +func DataLoss(err error) error { + if err == nil { + return nil + } + return errDataLoss{err} +} + +// FromContext returns the error class from the passed in context +func FromContext(ctx context.Context) error { + e := ctx.Err() + if e == nil { + return nil + } + + if e == context.Canceled { + return Cancelled(e) + } + if e == context.DeadlineExceeded { + return Deadline(e) + } + return Unknown(e) +} diff --git a/vendor/github.com/docker/docker/errdefs/is.go b/vendor/github.com/docker/docker/errdefs/is.go new file mode 100644 index 000000000..e0513331b --- /dev/null +++ b/vendor/github.com/docker/docker/errdefs/is.go @@ -0,0 +1,114 @@ +package errdefs // import "github.com/docker/docker/errdefs" + +type causer interface { + Cause() error +} + +func getImplementer(err error) error { + switch e := err.(type) { + case + ErrNotFound, + ErrInvalidParameter, + ErrConflict, + ErrUnauthorized, + ErrUnavailable, + ErrForbidden, + ErrSystem, + ErrNotModified, + ErrAlreadyExists, + ErrNotImplemented, + ErrCancelled, + ErrDeadline, + ErrDataLoss, + ErrUnknown: + return err + case causer: + return getImplementer(e.Cause()) + default: + return err + } +} + +// IsNotFound returns if the passed in error is an ErrNotFound +func IsNotFound(err error) bool { + _, ok := getImplementer(err).(ErrNotFound) + return ok +} + +// IsInvalidParameter returns if the passed in error is an ErrInvalidParameter +func IsInvalidParameter(err error) bool { + _, ok := getImplementer(err).(ErrInvalidParameter) + return ok +} + +// IsConflict returns if the passed in error is an ErrConflict +func IsConflict(err error) bool { + _, ok := getImplementer(err).(ErrConflict) + return ok +} + +// IsUnauthorized returns if the passed in error is an ErrUnauthorized +func IsUnauthorized(err error) bool { + _, ok := getImplementer(err).(ErrUnauthorized) + return ok +} + +// IsUnavailable returns if the passed in error is an ErrUnavailable +func IsUnavailable(err error) bool { + _, ok := getImplementer(err).(ErrUnavailable) + return ok +} + +// IsForbidden returns if the passed in error is an ErrForbidden +func IsForbidden(err error) bool { + _, ok := getImplementer(err).(ErrForbidden) + return ok +} + +// IsSystem returns if the passed in error is an ErrSystem +func IsSystem(err error) bool { + _, ok := getImplementer(err).(ErrSystem) + return ok +} + +// IsNotModified returns if the passed in error is a NotModified error +func IsNotModified(err error) bool { + _, ok := getImplementer(err).(ErrNotModified) + return ok +} + +// IsAlreadyExists returns if the passed in error is a AlreadyExists error +func IsAlreadyExists(err error) bool { + _, ok := getImplementer(err).(ErrAlreadyExists) + return ok +} + +// IsNotImplemented returns if the passed in error is an ErrNotImplemented +func IsNotImplemented(err error) bool { + _, ok := getImplementer(err).(ErrNotImplemented) + return ok +} + +// IsUnknown returns if the passed in error is an ErrUnknown +func IsUnknown(err error) bool { + _, ok := getImplementer(err).(ErrUnknown) + return ok +} + +// IsCancelled returns if the passed in error is an ErrCancelled +func IsCancelled(err error) bool { + _, ok := getImplementer(err).(ErrCancelled) + return ok +} + +// IsDeadline returns if the passed in error is an ErrDeadline +func IsDeadline(err error) bool { + _, ok := getImplementer(err).(ErrDeadline) + return ok +} + +// IsDataLoss returns if the passed in error is an ErrDataLoss +func IsDataLoss(err error) bool { + _, ok := getImplementer(err).(ErrDataLoss) + return ok +} diff --git a/vendor/github.com/docker/docker/opts/address_pools.go b/vendor/github.com/docker/docker/opts/address_pools.go new file mode 100644 index 000000000..9b27a6285 --- /dev/null +++ b/vendor/github.com/docker/docker/opts/address_pools.go @@ -0,0 +1,84 @@ +package opts + +import ( + "encoding/csv" + "encoding/json" + "fmt" + "strconv" + "strings" + + types "github.com/docker/libnetwork/ipamutils" +) + +// PoolsOpt is a Value type for parsing the default address pools definitions +type PoolsOpt struct { + values []*types.NetworkToSplit +} + +// UnmarshalJSON fills values structure info from JSON input +func (p *PoolsOpt) UnmarshalJSON(raw []byte) error { + return json.Unmarshal(raw, &(p.values)) +} + +// Set predefined pools +func (p *PoolsOpt) Set(value string) error { + csvReader := csv.NewReader(strings.NewReader(value)) + fields, err := csvReader.Read() + if err != nil { + return err + } + + poolsDef := types.NetworkToSplit{} + + for _, field := range fields { + parts := strings.SplitN(field, "=", 2) + if len(parts) != 2 { + return fmt.Errorf("invalid field '%s' must be a key=value pair", field) + } + + key := strings.ToLower(parts[0]) + value := strings.ToLower(parts[1]) + + switch key { + case "base": + poolsDef.Base = value + case "size": + size, err := strconv.Atoi(value) + if err != nil { + return fmt.Errorf("invalid size value: %q (must be integer): %v", value, err) + } + poolsDef.Size = size + default: + return fmt.Errorf("unexpected key '%s' in '%s'", key, field) + } + } + + p.values = append(p.values, &poolsDef) + + return nil +} + +// Type returns the type of this option +func (p *PoolsOpt) Type() string { + return "pool-options" +} + +// String returns a string repr of this option +func (p *PoolsOpt) String() string { + var pools []string + for _, pool := range p.values { + repr := fmt.Sprintf("%s %d", pool.Base, pool.Size) + pools = append(pools, repr) + } + return strings.Join(pools, ", ") +} + +// Value returns the mounts +func (p *PoolsOpt) Value() []*types.NetworkToSplit { + return p.values +} + +// Name returns the flag name of this option +func (p *PoolsOpt) Name() string { + return "default-address-pools" +} diff --git a/vendor/github.com/docker/docker/opts/env.go b/vendor/github.com/docker/docker/opts/env.go new file mode 100644 index 000000000..f6e5e9074 --- /dev/null +++ b/vendor/github.com/docker/docker/opts/env.go @@ -0,0 +1,48 @@ +package opts // import "github.com/docker/docker/opts" + +import ( + "fmt" + "os" + "runtime" + "strings" + + "github.com/pkg/errors" +) + +// ValidateEnv validates an environment variable and returns it. +// If no value is specified, it returns the current value using os.Getenv. +// +// As on ParseEnvFile and related to #16585, environment variable names +// are not validate what so ever, it's up to application inside docker +// to validate them or not. +// +// The only validation here is to check if name is empty, per #25099 +func ValidateEnv(val string) (string, error) { + arr := strings.Split(val, "=") + if arr[0] == "" { + return "", errors.Errorf("invalid environment variable: %s", val) + } + if len(arr) > 1 { + return val, nil + } + if !doesEnvExist(val) { + return val, nil + } + return fmt.Sprintf("%s=%s", val, os.Getenv(val)), nil +} + +func doesEnvExist(name string) bool { + for _, entry := range os.Environ() { + parts := strings.SplitN(entry, "=", 2) + if runtime.GOOS == "windows" { + // Environment variable are case-insensitive on Windows. PaTh, path and PATH are equivalent. + if strings.EqualFold(parts[0], name) { + return true + } + } + if parts[0] == name { + return true + } + } + return false +} diff --git a/vendor/github.com/docker/docker/opts/hosts.go b/vendor/github.com/docker/docker/opts/hosts.go index ad1675923..2adf4211d 100644 --- a/vendor/github.com/docker/docker/opts/hosts.go +++ b/vendor/github.com/docker/docker/opts/hosts.go @@ -1,4 +1,4 @@ -package opts +package opts // import "github.com/docker/docker/opts" import ( "fmt" @@ -9,7 +9,7 @@ import ( ) var ( - // DefaultHTTPPort Default HTTP Port used if only the protocol is provided to -H flag e.g. docker daemon -H tcp:// + // DefaultHTTPPort Default HTTP Port used if only the protocol is provided to -H flag e.g. dockerd -H tcp:// // These are the IANA registered port numbers for use with Docker // see http://www.iana.org/assignments/service-names-port-numbers/service-names-port-numbers.xhtml?search=docker DefaultHTTPPort = 2375 // Default HTTP Port @@ -29,15 +29,15 @@ var ( // ValidateHost validates that the specified string is a valid host and returns it. func ValidateHost(val string) (string, error) { host := strings.TrimSpace(val) - // The empty string means default and is not handled by parseDockerDaemonHost + // The empty string means default and is not handled by parseDaemonHost if host != "" { - _, err := parseDockerDaemonHost(host) + _, err := parseDaemonHost(host) if err != nil { return val, err } } // Note: unlike most flag validators, we don't return the mutated value here - // we need to know what the user entered later (using ParseHost) to adjust for tls + // we need to know what the user entered later (using ParseHost) to adjust for TLS return val, nil } @@ -52,7 +52,7 @@ func ParseHost(defaultToTLS bool, val string) (string, error) { } } else { var err error - host, err = parseDockerDaemonHost(host) + host, err = parseDaemonHost(host) if err != nil { return val, err } @@ -60,17 +60,17 @@ func ParseHost(defaultToTLS bool, val string) (string, error) { return host, nil } -// parseDockerDaemonHost parses the specified address and returns an address that will be used as the host. +// parseDaemonHost parses the specified address and returns an address that will be used as the host. // Depending of the address specified, this may return one of the global Default* strings defined in hosts.go. -func parseDockerDaemonHost(addr string) (string, error) { - addrParts := strings.Split(addr, "://") +func parseDaemonHost(addr string) (string, error) { + addrParts := strings.SplitN(addr, "://", 2) if len(addrParts) == 1 && addrParts[0] != "" { addrParts = []string{"tcp", addrParts[0]} } switch addrParts[0] { case "tcp": - return parseTCPAddr(addrParts[1], DefaultTCPHost) + return ParseTCPAddr(addrParts[1], DefaultTCPHost) case "unix": return parseSimpleProtoAddr("unix", addrParts[1], DefaultUnixSocket) case "npipe": @@ -97,12 +97,12 @@ func parseSimpleProtoAddr(proto, addr, defaultAddr string) (string, error) { return fmt.Sprintf("%s://%s", proto, addr), nil } -// parseTCPAddr parses and validates that the specified address is a valid TCP +// ParseTCPAddr parses and validates that the specified address is a valid TCP // address. It returns a formatted TCP address, either using the address parsed // from tryAddr, or the contents of defaultAddr if tryAddr is a blank string. // tryAddr is expected to have already been Trim()'d // defaultAddr must be in the full `tcp://host:port` form -func parseTCPAddr(tryAddr string, defaultAddr string) (string, error) { +func ParseTCPAddr(tryAddr string, defaultAddr string) (string, error) { if tryAddr == "" || tryAddr == "tcp://" { return defaultAddr, nil } @@ -127,8 +127,11 @@ func parseTCPAddr(tryAddr string, defaultAddr string) (string, error) { if err != nil { return "", err } - host, port, err := net.SplitHostPort(u.Host) + if err != nil { + // try port addition once + host, port, err = net.SplitHostPort(net.JoinHostPort(u.Host, defaultPort)) + } if err != nil { return "", fmt.Errorf("Invalid bind address format: %s", tryAddr) } @@ -146,3 +149,17 @@ func parseTCPAddr(tryAddr string, defaultAddr string) (string, error) { return fmt.Sprintf("tcp://%s%s", net.JoinHostPort(host, port), u.Path), nil } + +// ValidateExtraHost validates that the specified string is a valid extrahost and returns it. +// ExtraHost is in the form of name:ip where the ip has to be a valid ip (IPv4 or IPv6). +func ValidateExtraHost(val string) (string, error) { + // allow for IPv6 addresses in extra hosts by only splitting on first ":" + arr := strings.SplitN(val, ":", 2) + if len(arr) != 2 || len(arr[0]) == 0 { + return "", fmt.Errorf("bad format for add-host: %q", val) + } + if _, err := ValidateIPAddress(arr[1]); err != nil { + return "", fmt.Errorf("invalid IP address in add-host: %q", arr[1]) + } + return val, nil +} diff --git a/vendor/github.com/docker/docker/opts/hosts_unix.go b/vendor/github.com/docker/docker/opts/hosts_unix.go index 611407a9d..9d5bb6456 100644 --- a/vendor/github.com/docker/docker/opts/hosts_unix.go +++ b/vendor/github.com/docker/docker/opts/hosts_unix.go @@ -1,6 +1,6 @@ // +build !windows -package opts +package opts // import "github.com/docker/docker/opts" import "fmt" diff --git a/vendor/github.com/docker/docker/opts/hosts_windows.go b/vendor/github.com/docker/docker/opts/hosts_windows.go index 7c239e00f..906eba53e 100644 --- a/vendor/github.com/docker/docker/opts/hosts_windows.go +++ b/vendor/github.com/docker/docker/opts/hosts_windows.go @@ -1,6 +1,4 @@ -// +build windows - -package opts +package opts // import "github.com/docker/docker/opts" // DefaultHost constant defines the default host string used by docker on Windows var DefaultHost = "npipe://" + DefaultNamedPipe diff --git a/vendor/github.com/docker/docker/opts/ip.go b/vendor/github.com/docker/docker/opts/ip.go index c7b0dc994..cfbff3a9f 100644 --- a/vendor/github.com/docker/docker/opts/ip.go +++ b/vendor/github.com/docker/docker/opts/ip.go @@ -1,4 +1,4 @@ -package opts +package opts // import "github.com/docker/docker/opts" import ( "fmt" @@ -22,7 +22,7 @@ func NewIPOpt(ref *net.IP, defaultVal string) *IPOpt { } // Set sets an IPv4 or IPv6 address from a given string. If the given -// string is not parseable as an IP address it returns an error. +// string is not parsable as an IP address it returns an error. func (o *IPOpt) Set(val string) error { ip := net.ParseIP(val) if ip == nil { @@ -40,3 +40,8 @@ func (o *IPOpt) String() string { } return o.IP.String() } + +// Type returns the type of the option +func (o *IPOpt) Type() string { + return "ip" +} diff --git a/vendor/github.com/docker/docker/opts/opts.go b/vendor/github.com/docker/docker/opts/opts.go index 0b0998177..de8aacb80 100644 --- a/vendor/github.com/docker/docker/opts/opts.go +++ b/vendor/github.com/docker/docker/opts/opts.go @@ -1,10 +1,13 @@ -package opts +package opts // import "github.com/docker/docker/opts" import ( "fmt" "net" + "path" "regexp" "strings" + + "github.com/docker/go-units" ) var ( @@ -33,7 +36,10 @@ func NewListOptsRef(values *[]string, validator ValidatorFctType) *ListOpts { } func (opts *ListOpts) String() string { - return fmt.Sprintf("%v", []string((*opts.values))) + if len(*opts.values) == 0 { + return "" + } + return fmt.Sprintf("%v", *opts.values) } // Set validates if needed the input value and adds it to the @@ -46,7 +52,7 @@ func (opts *ListOpts) Set(value string) error { } value = v } - (*opts.values) = append((*opts.values), value) + *opts.values = append(*opts.values, value) return nil } @@ -54,7 +60,7 @@ func (opts *ListOpts) Set(value string) error { func (opts *ListOpts) Delete(key string) { for i, k := range *opts.values { if k == key { - (*opts.values) = append((*opts.values)[:i], (*opts.values)[i+1:]...) + *opts.values = append((*opts.values)[:i], (*opts.values)[i+1:]...) return } } @@ -72,7 +78,7 @@ func (opts *ListOpts) GetMap() map[string]struct{} { // GetAll returns the values of slice. func (opts *ListOpts) GetAll() []string { - return (*opts.values) + return *opts.values } // GetAllOrEmpty returns the values of the slice @@ -97,7 +103,18 @@ func (opts *ListOpts) Get(key string) bool { // Len returns the amount of element in the slice. func (opts *ListOpts) Len() int { - return len((*opts.values)) + return len(*opts.values) +} + +// Type returns a string name for this Option type +func (opts *ListOpts) Type() string { + return "list" +} + +// WithValidator returns the ListOpts with validator set. +func (opts *ListOpts) WithValidator(validator ValidatorFctType) *ListOpts { + opts.validator = validator + return opts } // NamedOption is an interface that list and map options @@ -129,7 +146,7 @@ func (o *NamedListOpts) Name() string { return o.name } -//MapOpts holds a map of values and a validation function. +// MapOpts holds a map of values and a validation function. type MapOpts struct { values map[string]string validator ValidatorFctType @@ -160,7 +177,12 @@ func (opts *MapOpts) GetAll() map[string]string { } func (opts *MapOpts) String() string { - return fmt.Sprintf("%v", map[string]string((opts.values))) + return fmt.Sprintf("%v", opts.values) +} + +// Type returns a string name for this Option type +func (opts *MapOpts) Type() string { + return "map" } // NewMapOpts creates a new MapOpts with the specified map of values and a validator. @@ -241,34 +263,75 @@ func ValidateLabel(val string) (string, error) { return val, nil } -// ValidateSysctl validates an sysctl and returns it. -func ValidateSysctl(val string) (string, error) { - validSysctlMap := map[string]bool{ - "kernel.msgmax": true, - "kernel.msgmnb": true, - "kernel.msgmni": true, - "kernel.sem": true, - "kernel.shmall": true, - "kernel.shmmax": true, - "kernel.shmmni": true, - "kernel.shm_rmid_forced": true, +// ValidateSingleGenericResource validates that a single entry in the +// generic resource list is valid. +// i.e 'GPU=UID1' is valid however 'GPU:UID1' or 'UID1' isn't +func ValidateSingleGenericResource(val string) (string, error) { + if strings.Count(val, "=") < 1 { + return "", fmt.Errorf("invalid node-generic-resource format `%s` expected `name=value`", val) } - validSysctlPrefixes := []string{ - "net.", - "fs.mqueue.", - } - arr := strings.Split(val, "=") - if len(arr) < 2 { - return "", fmt.Errorf("sysctl '%s' is not whitelisted", val) - } - if validSysctlMap[arr[0]] { - return val, nil - } - - for _, vp := range validSysctlPrefixes { - if strings.HasPrefix(arr[0], vp) { - return val, nil - } - } - return "", fmt.Errorf("sysctl '%s' is not whitelisted", val) + return val, nil +} + +// ParseLink parses and validates the specified string as a link format (name:alias) +func ParseLink(val string) (string, string, error) { + if val == "" { + return "", "", fmt.Errorf("empty string specified for links") + } + arr := strings.Split(val, ":") + if len(arr) > 2 { + return "", "", fmt.Errorf("bad format for links: %s", val) + } + if len(arr) == 1 { + return val, val, nil + } + // This is kept because we can actually get a HostConfig with links + // from an already created container and the format is not `foo:bar` + // but `/foo:/c1/bar` + if strings.HasPrefix(arr[0], "/") { + _, alias := path.Split(arr[1]) + return arr[0][1:], alias, nil + } + return arr[0], arr[1], nil +} + +// MemBytes is a type for human readable memory bytes (like 128M, 2g, etc) +type MemBytes int64 + +// String returns the string format of the human readable memory bytes +func (m *MemBytes) String() string { + // NOTE: In spf13/pflag/flag.go, "0" is considered as "zero value" while "0 B" is not. + // We return "0" in case value is 0 here so that the default value is hidden. + // (Sometimes "default 0 B" is actually misleading) + if m.Value() != 0 { + return units.BytesSize(float64(m.Value())) + } + return "0" +} + +// Set sets the value of the MemBytes by passing a string +func (m *MemBytes) Set(value string) error { + val, err := units.RAMInBytes(value) + *m = MemBytes(val) + return err +} + +// Type returns the type +func (m *MemBytes) Type() string { + return "bytes" +} + +// Value returns the value in int64 +func (m *MemBytes) Value() int64 { + return int64(*m) +} + +// UnmarshalJSON is the customized unmarshaler for MemBytes +func (m *MemBytes) UnmarshalJSON(s []byte) error { + if len(s) <= 2 || s[0] != '"' || s[len(s)-1] != '"' { + return fmt.Errorf("invalid size: %q", s) + } + val, err := units.RAMInBytes(string(s[1 : len(s)-1])) + *m = MemBytes(val) + return err } diff --git a/vendor/github.com/docker/docker/opts/opts_unix.go b/vendor/github.com/docker/docker/opts/opts_unix.go index f1ce844a8..0c32367cb 100644 --- a/vendor/github.com/docker/docker/opts/opts_unix.go +++ b/vendor/github.com/docker/docker/opts/opts_unix.go @@ -1,6 +1,6 @@ // +build !windows -package opts +package opts // import "github.com/docker/docker/opts" -// DefaultHTTPHost Default HTTP Host used if only port is provided to -H flag e.g. docker daemon -H tcp://:8080 +// DefaultHTTPHost Default HTTP Host used if only port is provided to -H flag e.g. dockerd -H tcp://:8080 const DefaultHTTPHost = "localhost" diff --git a/vendor/github.com/docker/docker/opts/opts_windows.go b/vendor/github.com/docker/docker/opts/opts_windows.go index ebe40c969..0e1b6c6d1 100644 --- a/vendor/github.com/docker/docker/opts/opts_windows.go +++ b/vendor/github.com/docker/docker/opts/opts_windows.go @@ -1,4 +1,4 @@ -package opts +package opts // import "github.com/docker/docker/opts" // TODO Windows. Identify bug in GOLang 1.5.1+ and/or Windows Server 2016 TP5. // @jhowardmsft, @swernli. @@ -52,5 +52,5 @@ package opts // to the delay if a user were to do 'docker run -H=tcp://localhost:2375...' // explicitly. -// DefaultHTTPHost Default HTTP Host used if only port is provided to -H flag e.g. docker daemon -H tcp://:8080 +// DefaultHTTPHost Default HTTP Host used if only port is provided to -H flag e.g. dockerd -H tcp://:8080 const DefaultHTTPHost = "127.0.0.1" diff --git a/vendor/github.com/docker/docker/opts/quotedstring.go b/vendor/github.com/docker/docker/opts/quotedstring.go new file mode 100644 index 000000000..6c889070e --- /dev/null +++ b/vendor/github.com/docker/docker/opts/quotedstring.go @@ -0,0 +1,37 @@ +package opts // import "github.com/docker/docker/opts" + +// QuotedString is a string that may have extra quotes around the value. The +// quotes are stripped from the value. +type QuotedString struct { + value *string +} + +// Set sets a new value +func (s *QuotedString) Set(val string) error { + *s.value = trimQuotes(val) + return nil +} + +// Type returns the type of the value +func (s *QuotedString) Type() string { + return "string" +} + +func (s *QuotedString) String() string { + return *s.value +} + +func trimQuotes(value string) string { + lastIndex := len(value) - 1 + for _, char := range []byte{'\'', '"'} { + if value[0] == char && value[lastIndex] == char { + return value[1:lastIndex] + } + } + return value +} + +// NewQuotedString returns a new quoted string option +func NewQuotedString(value *string) *QuotedString { + return &QuotedString{value: value} +} diff --git a/vendor/github.com/docker/docker/opts/runtime.go b/vendor/github.com/docker/docker/opts/runtime.go new file mode 100644 index 000000000..4b9babf0a --- /dev/null +++ b/vendor/github.com/docker/docker/opts/runtime.go @@ -0,0 +1,79 @@ +package opts // import "github.com/docker/docker/opts" + +import ( + "fmt" + "strings" + + "github.com/docker/docker/api/types" +) + +// RuntimeOpt defines a map of Runtimes +type RuntimeOpt struct { + name string + stockRuntimeName string + values *map[string]types.Runtime +} + +// NewNamedRuntimeOpt creates a new RuntimeOpt +func NewNamedRuntimeOpt(name string, ref *map[string]types.Runtime, stockRuntime string) *RuntimeOpt { + if ref == nil { + ref = &map[string]types.Runtime{} + } + return &RuntimeOpt{name: name, values: ref, stockRuntimeName: stockRuntime} +} + +// Name returns the name of the NamedListOpts in the configuration. +func (o *RuntimeOpt) Name() string { + return o.name +} + +// Set validates and updates the list of Runtimes +func (o *RuntimeOpt) Set(val string) error { + parts := strings.SplitN(val, "=", 2) + if len(parts) != 2 { + return fmt.Errorf("invalid runtime argument: %s", val) + } + + parts[0] = strings.TrimSpace(parts[0]) + parts[1] = strings.TrimSpace(parts[1]) + if parts[0] == "" || parts[1] == "" { + return fmt.Errorf("invalid runtime argument: %s", val) + } + + parts[0] = strings.ToLower(parts[0]) + if parts[0] == o.stockRuntimeName { + return fmt.Errorf("runtime name '%s' is reserved", o.stockRuntimeName) + } + + if _, ok := (*o.values)[parts[0]]; ok { + return fmt.Errorf("runtime '%s' was already defined", parts[0]) + } + + (*o.values)[parts[0]] = types.Runtime{Path: parts[1]} + + return nil +} + +// String returns Runtime values as a string. +func (o *RuntimeOpt) String() string { + var out []string + for k := range *o.values { + out = append(out, k) + } + + return fmt.Sprintf("%v", out) +} + +// GetMap returns a map of Runtimes (name: path) +func (o *RuntimeOpt) GetMap() map[string]types.Runtime { + if o.values != nil { + return *o.values + } + + return map[string]types.Runtime{} +} + +// Type returns the type of the option +func (o *RuntimeOpt) Type() string { + return "runtime" +} diff --git a/vendor/github.com/docker/docker/opts/ulimit.go b/vendor/github.com/docker/docker/opts/ulimit.go new file mode 100644 index 000000000..0e2a36236 --- /dev/null +++ b/vendor/github.com/docker/docker/opts/ulimit.go @@ -0,0 +1,81 @@ +package opts // import "github.com/docker/docker/opts" + +import ( + "fmt" + + "github.com/docker/go-units" +) + +// UlimitOpt defines a map of Ulimits +type UlimitOpt struct { + values *map[string]*units.Ulimit +} + +// NewUlimitOpt creates a new UlimitOpt +func NewUlimitOpt(ref *map[string]*units.Ulimit) *UlimitOpt { + if ref == nil { + ref = &map[string]*units.Ulimit{} + } + return &UlimitOpt{ref} +} + +// Set validates a Ulimit and sets its name as a key in UlimitOpt +func (o *UlimitOpt) Set(val string) error { + l, err := units.ParseUlimit(val) + if err != nil { + return err + } + + (*o.values)[l.Name] = l + + return nil +} + +// String returns Ulimit values as a string. +func (o *UlimitOpt) String() string { + var out []string + for _, v := range *o.values { + out = append(out, v.String()) + } + + return fmt.Sprintf("%v", out) +} + +// GetList returns a slice of pointers to Ulimits. +func (o *UlimitOpt) GetList() []*units.Ulimit { + var ulimits []*units.Ulimit + for _, v := range *o.values { + ulimits = append(ulimits, v) + } + + return ulimits +} + +// Type returns the option type +func (o *UlimitOpt) Type() string { + return "ulimit" +} + +// NamedUlimitOpt defines a named map of Ulimits +type NamedUlimitOpt struct { + name string + UlimitOpt +} + +var _ NamedOption = &NamedUlimitOpt{} + +// NewNamedUlimitOpt creates a new NamedUlimitOpt +func NewNamedUlimitOpt(name string, ref *map[string]*units.Ulimit) *NamedUlimitOpt { + if ref == nil { + ref = &map[string]*units.Ulimit{} + } + return &NamedUlimitOpt{ + name: name, + UlimitOpt: *NewUlimitOpt(ref), + } +} + +// Name returns the option name +func (o *NamedUlimitOpt) Name() string { + return o.name +} diff --git a/vendor/github.com/docker/docker/pkg/archive/archive.go b/vendor/github.com/docker/docker/pkg/archive/archive.go index 1603a2302..daddebded 100644 --- a/vendor/github.com/docker/docker/pkg/archive/archive.go +++ b/vendor/github.com/docker/docker/pkg/archive/archive.go @@ -1,4 +1,4 @@ -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "archive/tar" @@ -6,7 +6,7 @@ import ( "bytes" "compress/bzip2" "compress/gzip" - "errors" + "context" "fmt" "io" "io/ioutil" @@ -14,29 +14,36 @@ import ( "os/exec" "path/filepath" "runtime" + "strconv" "strings" "syscall" + "time" - "github.com/Sirupsen/logrus" "github.com/docker/docker/pkg/fileutils" "github.com/docker/docker/pkg/idtools" "github.com/docker/docker/pkg/ioutils" "github.com/docker/docker/pkg/pools" - "github.com/docker/docker/pkg/promise" "github.com/docker/docker/pkg/system" + "github.com/sirupsen/logrus" ) +var unpigzPath string + +func init() { + if path, err := exec.LookPath("unpigz"); err != nil { + logrus.Debug("unpigz binary not found in PATH, falling back to go gzip library") + } else { + logrus.Debugf("Using unpigz binary found at path %s", path) + unpigzPath = path + } +} + type ( - // Archive is a type of io.ReadCloser which has two interfaces Read and Closer. - Archive io.ReadCloser - // Reader is a type of io.Reader. - Reader io.Reader // Compression is the state represents if compressed or not. Compression int - // TarChownOptions wraps the chown options UID and GID. - TarChownOptions struct { - UID, GID int - } + // WhiteoutFormat is the format of whiteouts unpacked + WhiteoutFormat int + // TarOptions wraps the tar options. TarOptions struct { IncludeFiles []string @@ -45,42 +52,39 @@ type ( NoLchown bool UIDMaps []idtools.IDMap GIDMaps []idtools.IDMap - ChownOpts *TarChownOptions + ChownOpts *idtools.IDPair IncludeSourceDir bool + // WhiteoutFormat is the expected on disk format for whiteout files. + // This format will be converted to the standard format on pack + // and from the standard format on unpack. + WhiteoutFormat WhiteoutFormat // When unpacking, specifies whether overwriting a directory with a // non-directory is allowed and vice versa. NoOverwriteDirNonDir bool // For each include when creating an archive, the included name will be // replaced with the matching name from this map. RebaseNames map[string]string + InUserNS bool } - - // Archiver allows the reuse of most utility functions of this package - // with a pluggable Untar function. Also, to facilitate the passing of - // specific id mappings for untar, an archiver can be created with maps - // which will then be passed to Untar operations - Archiver struct { - Untar func(io.Reader, string, *TarOptions) error - UIDMaps []idtools.IDMap - GIDMaps []idtools.IDMap - } - - // breakoutError is used to differentiate errors related to breaking out - // When testing archive breakout in the unit tests, this error is expected - // in order for the test to pass. - breakoutError error ) -var ( - // ErrNotImplemented is the error message of function not implemented. - ErrNotImplemented = errors.New("Function not implemented") - defaultArchiver = &Archiver{Untar: Untar, UIDMaps: nil, GIDMaps: nil} -) +// Archiver implements the Archiver interface and allows the reuse of most utility functions of +// this package with a pluggable Untar function. Also, to facilitate the passing of specific id +// mappings for untar, an Archiver can be created with maps which will then be passed to Untar operations. +type Archiver struct { + Untar func(io.Reader, string, *TarOptions) error + IDMappingsVar *idtools.IDMappings +} -const ( - // HeaderSize is the size in bytes of a tar header - HeaderSize = 512 -) +// NewDefaultArchiver returns a new Archiver without any IDMappings +func NewDefaultArchiver() *Archiver { + return &Archiver{Untar: Untar, IDMappingsVar: &idtools.IDMappings{}} +} + +// breakoutError is used to differentiate errors related to breaking out +// When testing archive breakout in the unit tests, this error is expected +// in order for the test to pass. +type breakoutError error const ( // Uncompressed represents the uncompressed. @@ -93,17 +97,23 @@ const ( Xz ) -// IsArchive checks for the magic bytes of a tar or any supported compression -// algorithm. -func IsArchive(header []byte) bool { - compression := DetectCompression(header) - if compression != Uncompressed { - return true - } - r := tar.NewReader(bytes.NewBuffer(header)) - _, err := r.Next() - return err == nil -} +const ( + // AUFSWhiteoutFormat is the default format for whiteouts + AUFSWhiteoutFormat WhiteoutFormat = iota + // OverlayWhiteoutFormat formats whiteout according to the overlay + // standard. + OverlayWhiteoutFormat +) + +const ( + modeISDIR = 040000 // Directory + modeISFIFO = 010000 // FIFO + modeISREG = 0100000 // Regular file + modeISLNK = 0120000 // Symbolic link + modeISBLK = 060000 // Block special file + modeISCHR = 020000 // Character special file + modeISSOCK = 0140000 // Socket +) // IsArchivePath checks if the (possibly compressed) file at the given path // starts with a tar file header. @@ -117,6 +127,7 @@ func IsArchivePath(path string) bool { if err != nil { return false } + defer rdr.Close() r := tar.NewReader(rdr) _, err = r.Next() return err == nil @@ -130,20 +141,44 @@ func DetectCompression(source []byte) Compression { Xz: {0xFD, 0x37, 0x7A, 0x58, 0x5A, 0x00}, } { if len(source) < len(m) { - logrus.Debugf("Len too short") + logrus.Debug("Len too short") continue } - if bytes.Compare(m, source[:len(m)]) == 0 { + if bytes.Equal(m, source[:len(m)]) { return compression } } return Uncompressed } -func xzDecompress(archive io.Reader) (io.ReadCloser, <-chan struct{}, error) { +func xzDecompress(ctx context.Context, archive io.Reader) (io.ReadCloser, error) { args := []string{"xz", "-d", "-c", "-q"} - return cmdStream(exec.Command(args[0], args[1:]...), archive) + return cmdStream(exec.CommandContext(ctx, args[0], args[1:]...), archive) +} + +func gzDecompress(ctx context.Context, buf io.Reader) (io.ReadCloser, error) { + if unpigzPath == "" { + return gzip.NewReader(buf) + } + + disablePigzEnv := os.Getenv("MOBY_DISABLE_PIGZ") + if disablePigzEnv != "" { + if disablePigz, err := strconv.ParseBool(disablePigzEnv); err != nil { + return nil, err + } else if disablePigz { + return gzip.NewReader(buf) + } + } + + return cmdStream(exec.CommandContext(ctx, unpigzPath, "-d", "-c"), buf) +} + +func wrapReadCloser(readBuf io.ReadCloser, cancel context.CancelFunc) io.ReadCloser { + return ioutils.NewReadCloserWrapper(readBuf, func() error { + cancel() + return readBuf.Close() + }) } // DecompressStream decompresses the archive and returns a ReaderCloser with the decompressed archive. @@ -167,32 +202,35 @@ func DecompressStream(archive io.Reader) (io.ReadCloser, error) { readBufWrapper := p.NewReadCloserWrapper(buf, buf) return readBufWrapper, nil case Gzip: - gzReader, err := gzip.NewReader(buf) + ctx, cancel := context.WithCancel(context.Background()) + + gzReader, err := gzDecompress(ctx, buf) if err != nil { + cancel() return nil, err } readBufWrapper := p.NewReadCloserWrapper(buf, gzReader) - return readBufWrapper, nil + return wrapReadCloser(readBufWrapper, cancel), nil case Bzip2: bz2Reader := bzip2.NewReader(buf) readBufWrapper := p.NewReadCloserWrapper(buf, bz2Reader) return readBufWrapper, nil case Xz: - xzReader, chdone, err := xzDecompress(buf) + ctx, cancel := context.WithCancel(context.Background()) + + xzReader, err := xzDecompress(ctx, buf) if err != nil { + cancel() return nil, err } readBufWrapper := p.NewReadCloserWrapper(buf, xzReader) - return ioutils.NewReadCloserWrapper(readBufWrapper, func() error { - <-chdone - return readBufWrapper.Close() - }), nil + return wrapReadCloser(readBufWrapper, cancel), nil default: return nil, fmt.Errorf("Unsupported compression format %s", (&compression).Extension()) } } -// CompressStream compresseses the dest with specified compression algorithm. +// CompressStream compresses the dest with specified compression algorithm. func CompressStream(dest io.Writer, compression Compression) (io.WriteCloser, error) { p := pools.BufioWriter32KPool buf := p.Get(dest) @@ -213,6 +251,93 @@ func CompressStream(dest io.Writer, compression Compression) (io.WriteCloser, er } } +// TarModifierFunc is a function that can be passed to ReplaceFileTarWrapper to +// modify the contents or header of an entry in the archive. If the file already +// exists in the archive the TarModifierFunc will be called with the Header and +// a reader which will return the files content. If the file does not exist both +// header and content will be nil. +type TarModifierFunc func(path string, header *tar.Header, content io.Reader) (*tar.Header, []byte, error) + +// ReplaceFileTarWrapper converts inputTarStream to a new tar stream. Files in the +// tar stream are modified if they match any of the keys in mods. +func ReplaceFileTarWrapper(inputTarStream io.ReadCloser, mods map[string]TarModifierFunc) io.ReadCloser { + pipeReader, pipeWriter := io.Pipe() + + go func() { + tarReader := tar.NewReader(inputTarStream) + tarWriter := tar.NewWriter(pipeWriter) + defer inputTarStream.Close() + defer tarWriter.Close() + + modify := func(name string, original *tar.Header, modifier TarModifierFunc, tarReader io.Reader) error { + header, data, err := modifier(name, original, tarReader) + switch { + case err != nil: + return err + case header == nil: + return nil + } + + header.Name = name + header.Size = int64(len(data)) + if err := tarWriter.WriteHeader(header); err != nil { + return err + } + if len(data) != 0 { + if _, err := tarWriter.Write(data); err != nil { + return err + } + } + return nil + } + + var err error + var originalHeader *tar.Header + for { + originalHeader, err = tarReader.Next() + if err == io.EOF { + break + } + if err != nil { + pipeWriter.CloseWithError(err) + return + } + + modifier, ok := mods[originalHeader.Name] + if !ok { + // No modifiers for this file, copy the header and data + if err := tarWriter.WriteHeader(originalHeader); err != nil { + pipeWriter.CloseWithError(err) + return + } + if _, err := pools.Copy(tarWriter, tarReader); err != nil { + pipeWriter.CloseWithError(err) + return + } + continue + } + delete(mods, originalHeader.Name) + + if err := modify(originalHeader.Name, originalHeader, modifier, tarReader); err != nil { + pipeWriter.CloseWithError(err) + return + } + } + + // Apply the modifiers that haven't matched any files in the archive + for name, modifier := range mods { + if err := modify(name, nil, modifier, nil); err != nil { + pipeWriter.CloseWithError(err) + return + } + } + + pipeWriter.Close() + + }() + return pipeReader +} + // Extension returns the extension of a file that uses the specified compression algorithm. func (compression *Compression) Extension() string { switch *compression { @@ -228,14 +353,96 @@ func (compression *Compression) Extension() string { return "" } +// FileInfoHeader creates a populated Header from fi. +// Compared to archive pkg this function fills in more information. +// Also, regardless of Go version, this function fills file type bits (e.g. hdr.Mode |= modeISDIR), +// which have been deleted since Go 1.9 archive/tar. +func FileInfoHeader(name string, fi os.FileInfo, link string) (*tar.Header, error) { + hdr, err := tar.FileInfoHeader(fi, link) + if err != nil { + return nil, err + } + hdr.Format = tar.FormatPAX + hdr.ModTime = hdr.ModTime.Truncate(time.Second) + hdr.AccessTime = time.Time{} + hdr.ChangeTime = time.Time{} + hdr.Mode = fillGo18FileTypeBits(int64(chmodTarEntry(os.FileMode(hdr.Mode))), fi) + name, err = canonicalTarName(name, fi.IsDir()) + if err != nil { + return nil, fmt.Errorf("tar: cannot canonicalize path: %v", err) + } + hdr.Name = name + if err := setHeaderForSpecialDevice(hdr, name, fi.Sys()); err != nil { + return nil, err + } + return hdr, nil +} + +// fillGo18FileTypeBits fills type bits which have been removed on Go 1.9 archive/tar +// https://github.com/golang/go/commit/66b5a2f +func fillGo18FileTypeBits(mode int64, fi os.FileInfo) int64 { + fm := fi.Mode() + switch { + case fm.IsRegular(): + mode |= modeISREG + case fi.IsDir(): + mode |= modeISDIR + case fm&os.ModeSymlink != 0: + mode |= modeISLNK + case fm&os.ModeDevice != 0: + if fm&os.ModeCharDevice != 0 { + mode |= modeISCHR + } else { + mode |= modeISBLK + } + case fm&os.ModeNamedPipe != 0: + mode |= modeISFIFO + case fm&os.ModeSocket != 0: + mode |= modeISSOCK + } + return mode +} + +// ReadSecurityXattrToTarHeader reads security.capability xattr from filesystem +// to a tar header +func ReadSecurityXattrToTarHeader(path string, hdr *tar.Header) error { + capability, _ := system.Lgetxattr(path, "security.capability") + if capability != nil { + hdr.Xattrs = make(map[string]string) + hdr.Xattrs["security.capability"] = string(capability) + } + return nil +} + +type tarWhiteoutConverter interface { + ConvertWrite(*tar.Header, string, os.FileInfo) (*tar.Header, error) + ConvertRead(*tar.Header, string) (bool, error) +} + type tarAppender struct { TarWriter *tar.Writer Buffer *bufio.Writer // for hardlink mapping - SeenFiles map[uint64]string - UIDMaps []idtools.IDMap - GIDMaps []idtools.IDMap + SeenFiles map[uint64]string + IDMappings *idtools.IDMappings + ChownOpts *idtools.IDPair + + // For packing and unpacking whiteout files in the + // non standard format. The whiteout files defined + // by the AUFS standard are used as the tar whiteout + // standard. + WhiteoutConverter tarWhiteoutConverter +} + +func newTarAppender(idMapping *idtools.IDMappings, writer io.Writer, chownOpts *idtools.IDPair) *tarAppender { + return &tarAppender{ + SeenFiles: make(map[uint64]string), + TarWriter: tar.NewWriter(writer), + Buffer: pools.BufioWriter32KPool.Get(nil), + IDMappings: idMapping, + ChownOpts: chownOpts, + } } // canonicalTarName provides a platform-independent and consistent posix-style @@ -253,39 +460,37 @@ func canonicalTarName(name string, isDir bool) (string, error) { return name, nil } +// addTarFile adds to the tar archive a file from `path` as `name` func (ta *tarAppender) addTarFile(path, name string) error { fi, err := os.Lstat(path) if err != nil { return err } - link := "" + var link string if fi.Mode()&os.ModeSymlink != 0 { - if link, err = os.Readlink(path); err != nil { + var err error + link, err = os.Readlink(path) + if err != nil { return err } } - hdr, err := tar.FileInfoHeader(fi, link) + hdr, err := FileInfoHeader(name, fi, link) if err != nil { return err } - hdr.Mode = int64(chmodTarEntry(os.FileMode(hdr.Mode))) - - name, err = canonicalTarName(name, fi.IsDir()) - if err != nil { - return fmt.Errorf("tar: cannot canonicalize path: %v", err) - } - hdr.Name = name - - inode, err := setHeaderForSpecialDevice(hdr, ta, name, fi.Sys()) - if err != nil { + if err := ReadSecurityXattrToTarHeader(path, hdr); err != nil { return err } // if it's not a directory and has more than 1 link, - // it's hardlinked, so set the type flag accordingly + // it's hard linked, so set the type flag accordingly if !fi.IsDir() && hasHardlinks(fi) { + inode, err := getInodeFromStat(fi.Sys()) + if err != nil { + return err + } // a link should have a name that it links too // and that linked name should be first in the tar archive if oldpath, ok := ta.SeenFiles[inode]; ok { @@ -297,38 +502,63 @@ func (ta *tarAppender) addTarFile(path, name string) error { } } - capability, _ := system.Lgetxattr(path, "security.capability") - if capability != nil { - hdr.Xattrs = make(map[string]string) - hdr.Xattrs["security.capability"] = string(capability) - } + //check whether the file is overlayfs whiteout + //if yes, skip re-mapping container ID mappings. + isOverlayWhiteout := fi.Mode()&os.ModeCharDevice != 0 && hdr.Devmajor == 0 && hdr.Devminor == 0 //handle re-mapping container ID mappings back to host ID mappings before //writing tar headers/files. We skip whiteout files because they were written //by the kernel and already have proper ownership relative to the host - if !strings.HasPrefix(filepath.Base(hdr.Name), WhiteoutPrefix) && (ta.UIDMaps != nil || ta.GIDMaps != nil) { - uid, gid, err := getFileUIDGID(fi.Sys()) + if !isOverlayWhiteout && + !strings.HasPrefix(filepath.Base(hdr.Name), WhiteoutPrefix) && + !ta.IDMappings.Empty() { + fileIDPair, err := getFileUIDGID(fi.Sys()) if err != nil { return err } - xUID, err := idtools.ToContainer(uid, ta.UIDMaps) + hdr.Uid, hdr.Gid, err = ta.IDMappings.ToContainer(fileIDPair) if err != nil { return err } - xGID, err := idtools.ToContainer(gid, ta.GIDMaps) + } + + // explicitly override with ChownOpts + if ta.ChownOpts != nil { + hdr.Uid = ta.ChownOpts.UID + hdr.Gid = ta.ChownOpts.GID + } + + if ta.WhiteoutConverter != nil { + wo, err := ta.WhiteoutConverter.ConvertWrite(hdr, path, fi) if err != nil { return err } - hdr.Uid = xUID - hdr.Gid = xGID + + // If a new whiteout file exists, write original hdr, then + // replace hdr with wo to be written after. Whiteouts should + // always be written after the original. Note the original + // hdr may have been updated to be a whiteout with returning + // a whiteout header + if wo != nil { + if err := ta.TarWriter.WriteHeader(hdr); err != nil { + return err + } + if hdr.Typeflag == tar.TypeReg && hdr.Size > 0 { + return fmt.Errorf("tar: cannot use whiteout for non-empty file") + } + hdr = wo + } } if err := ta.TarWriter.WriteHeader(hdr); err != nil { return err } - if hdr.Typeflag == tar.TypeReg { - file, err := os.Open(path) + if hdr.Typeflag == tar.TypeReg && hdr.Size > 0 { + // We use system.OpenSequential to ensure we use sequential file + // access on Windows to avoid depleting the standby list. + // On Linux, this equates to a regular os.Open. + file, err := system.OpenSequential(path) if err != nil { return err } @@ -349,7 +579,7 @@ func (ta *tarAppender) addTarFile(path, name string) error { return nil } -func createTarFile(path, extractDir string, hdr *tar.Header, reader io.Reader, Lchown bool, chownOpts *TarChownOptions) error { +func createTarFile(path, extractDir string, hdr *tar.Header, reader io.Reader, Lchown bool, chownOpts *idtools.IDPair, inUserns bool) error { // hdr.Mode is in linux format, which we can use for sycalls, // but for os.Foo() calls we need the mode converted to os.FileMode, // so use hdrInfo.Mode() (they differ for e.g. setuid bits) @@ -366,8 +596,10 @@ func createTarFile(path, extractDir string, hdr *tar.Header, reader io.Reader, L } case tar.TypeReg, tar.TypeRegA: - // Source is regular file - file, err := os.OpenFile(path, os.O_CREATE|os.O_WRONLY, hdrInfo.Mode()) + // Source is regular file. We use system.OpenFileSequential to use sequential + // file access to avoid depleting the standby list on Windows. + // On Linux, this equates to a regular os.OpenFile + file, err := system.OpenFileSequential(path, os.O_CREATE|os.O_WRONLY, hdrInfo.Mode()) if err != nil { return err } @@ -377,7 +609,16 @@ func createTarFile(path, extractDir string, hdr *tar.Header, reader io.Reader, L } file.Close() - case tar.TypeBlock, tar.TypeChar, tar.TypeFifo: + case tar.TypeBlock, tar.TypeChar: + if inUserns { // cannot create devices in a userns + return nil + } + // Handle this is an OS-specific way + if err := handleTarTypeBlockCharFifo(hdr, path); err != nil { + return err + } + + case tar.TypeFifo: // Handle this is an OS-specific way if err := handleTarTypeBlockCharFifo(hdr, path); err != nil { return err @@ -408,17 +649,17 @@ func createTarFile(path, extractDir string, hdr *tar.Header, reader io.Reader, L } case tar.TypeXGlobalHeader: - logrus.Debugf("PAX Global Extended Headers found and ignored") + logrus.Debug("PAX Global Extended Headers found and ignored") return nil default: - return fmt.Errorf("Unhandled tar header type %d\n", hdr.Typeflag) + return fmt.Errorf("unhandled tar header type %d", hdr.Typeflag) } // Lchown is not supported on Windows. if Lchown && runtime.GOOS != "windows" { if chownOpts == nil { - chownOpts = &TarChownOptions{UID: hdr.Uid, GID: hdr.Gid} + chownOpts = &idtools.IDPair{UID: hdr.Uid, GID: hdr.Gid} } if err := os.Lchown(path, chownOpts.UID, chownOpts.GID); err != nil { return err @@ -428,8 +669,15 @@ func createTarFile(path, extractDir string, hdr *tar.Header, reader io.Reader, L var errors []string for key, value := range hdr.Xattrs { if err := system.Lsetxattr(path, key, []byte(value), 0); err != nil { - // We ignore errors here because not all graphdrivers support xattrs. - errors = append(errors, err.Error()) + if err == syscall.ENOTSUP { + // We ignore errors here because not all graphdrivers support + // xattrs *cough* old versions of AUFS *cough*. However only + // ENOTSUP should be emitted in that case, otherwise we still + // bail. + errors = append(errors, err.Error()) + continue + } + return err } } @@ -486,8 +734,7 @@ func TarWithOptions(srcPath string, options *TarOptions) (io.ReadCloser, error) // on platforms other than Windows. srcPath = fixVolumePathPrefix(srcPath) - patterns, patDirs, exceptions, err := fileutils.CleanPatterns(options.ExcludePatterns) - + pm, err := fileutils.NewPatternMatcher(options.ExcludePatterns) if err != nil { return nil, err } @@ -500,13 +747,12 @@ func TarWithOptions(srcPath string, options *TarOptions) (io.ReadCloser, error) } go func() { - ta := &tarAppender{ - TarWriter: tar.NewWriter(compressWriter), - Buffer: pools.BufioWriter32KPool.Get(nil), - SeenFiles: make(map[uint64]string), - UIDMaps: options.UIDMaps, - GIDMaps: options.GIDMaps, - } + ta := newTarAppender( + idtools.NewIDMappingsFromMaps(options.UIDMaps, options.GIDMaps), + compressWriter, + options.ChownOpts, + ) + ta.WhiteoutConverter = getWhiteoutConverter(options.WhiteoutFormat) defer func() { // Make sure to check the error on Close. @@ -583,7 +829,7 @@ func TarWithOptions(srcPath string, options *TarOptions) (io.ReadCloser, error) // is asking for that file no matter what - which is true // for some files, like .dockerignore and Dockerfile (sometimes) if include != relFilePath { - skip, err = fileutils.OptimizedMatches(relFilePath, patterns, patDirs) + skip, err = pm.Matches(relFilePath) if err != nil { logrus.Errorf("Error matching %s: %v", relFilePath, err) return err @@ -593,7 +839,7 @@ func TarWithOptions(srcPath string, options *TarOptions) (io.ReadCloser, error) if skip { // If we want to skip this file and its a directory // then we should first check to see if there's an - // excludes pattern (eg !dir/file) that starts with this + // excludes pattern (e.g. !dir/file) that starts with this // dir. If so then we can't skip this dir. // Its not a dir then so we can just return/skip. @@ -602,18 +848,17 @@ func TarWithOptions(srcPath string, options *TarOptions) (io.ReadCloser, error) } // No exceptions (!...) in patterns so just skip dir - if !exceptions { + if !pm.Exclusions() { return filepath.SkipDir } dirSlash := relFilePath + string(filepath.Separator) - for _, pat := range patterns { - if pat[0] != '!' { + for _, pat := range pm.Patterns() { + if !pat.Exclusion() { continue } - pat = pat[1:] + string(filepath.Separator) - if strings.HasPrefix(pat, dirSlash) { + if strings.HasPrefix(pat.String()+string(filepath.Separator), dirSlash) { // found a match - so can't skip this dir return nil } @@ -663,10 +908,9 @@ func Unpack(decompressedArchive io.Reader, dest string, options *TarOptions) err defer pools.BufioReader32KPool.Put(trBuf) var dirs []*tar.Header - remappedRootUID, remappedRootGID, err := idtools.GetRootUIDGID(options.UIDMaps, options.GIDMaps) - if err != nil { - return err - } + idMappings := idtools.NewIDMappingsFromMaps(options.UIDMaps, options.GIDMaps) + rootIDs := idMappings.RootPair() + whiteoutConverter := getWhiteoutConverter(options.WhiteoutFormat) // Iterate through the files in the archive. loop: @@ -699,7 +943,7 @@ loop: parent := filepath.Dir(hdr.Name) parentPath := filepath.Join(dest, parent) if _, err := os.Lstat(parentPath); err != nil && os.IsNotExist(err) { - err = idtools.MkdirAllNewAs(parentPath, 0777, remappedRootUID, remappedRootGID) + err = idtools.MkdirAllAndChownNew(parentPath, 0777, rootIDs) if err != nil { return err } @@ -744,29 +988,21 @@ loop: } trBuf.Reset(tr) - // if the options contain a uid & gid maps, convert header uid/gid - // entries using the maps such that lchown sets the proper mapped - // uid/gid after writing the file. We only perform this mapping if - // the file isn't already owned by the remapped root UID or GID, as - // that specific uid/gid has no mapping from container -> host, and - // those files already have the proper ownership for inside the - // container. - if hdr.Uid != remappedRootUID { - xUID, err := idtools.ToHost(hdr.Uid, options.UIDMaps) - if err != nil { - return err - } - hdr.Uid = xUID - } - if hdr.Gid != remappedRootGID { - xGID, err := idtools.ToHost(hdr.Gid, options.GIDMaps) - if err != nil { - return err - } - hdr.Gid = xGID + if err := remapIDs(idMappings, hdr); err != nil { + return err } - if err := createTarFile(path, dest, hdr, trBuf, !options.NoLchown, options.ChownOpts); err != nil { + if whiteoutConverter != nil { + writeFile, err := whiteoutConverter.ConvertRead(hdr, path) + if err != nil { + return err + } + if !writeFile { + continue + } + } + + if err := createTarFile(path, dest, hdr, trBuf, !options.NoLchown, options.ChownOpts, options.InUserNS); err != nil { return err } @@ -838,23 +1074,13 @@ func (archiver *Archiver) TarUntar(src, dst string) error { return err } defer archive.Close() - - var options *TarOptions - if archiver.UIDMaps != nil || archiver.GIDMaps != nil { - options = &TarOptions{ - UIDMaps: archiver.UIDMaps, - GIDMaps: archiver.GIDMaps, - } + options := &TarOptions{ + UIDMaps: archiver.IDMappingsVar.UIDs(), + GIDMaps: archiver.IDMappingsVar.GIDs(), } return archiver.Untar(archive, dst, options) } -// TarUntar is a convenience function which calls Tar and Untar, with the output of one piped into the other. -// If either Tar or Untar fails, TarUntar aborts and returns the error. -func TarUntar(src, dst string) error { - return defaultArchiver.TarUntar(src, dst) -} - // UntarPath untar a file from path to a destination, src is the source tar file path. func (archiver *Archiver) UntarPath(src, dst string) error { archive, err := os.Open(src) @@ -862,22 +1088,13 @@ func (archiver *Archiver) UntarPath(src, dst string) error { return err } defer archive.Close() - var options *TarOptions - if archiver.UIDMaps != nil || archiver.GIDMaps != nil { - options = &TarOptions{ - UIDMaps: archiver.UIDMaps, - GIDMaps: archiver.GIDMaps, - } + options := &TarOptions{ + UIDMaps: archiver.IDMappingsVar.UIDs(), + GIDMaps: archiver.IDMappingsVar.GIDs(), } return archiver.Untar(archive, dst, options) } -// UntarPath is a convenience function which looks for an archive -// at filesystem path `src`, and unpacks it at `dst`. -func UntarPath(src, dst string) error { - return defaultArchiver.UntarPath(src, dst) -} - // CopyWithTar creates a tar archive of filesystem path `src`, and // unpacks it at filesystem path `dst`. // The archive is streamed directly with fixed buffering and no @@ -891,30 +1108,19 @@ func (archiver *Archiver) CopyWithTar(src, dst string) error { return archiver.CopyFileWithTar(src, dst) } - // if this archiver is set up with ID mapping we need to create + // if this Archiver is set up with ID mapping we need to create // the new destination directory with the remapped root UID/GID pair // as owner - rootUID, rootGID, err := idtools.GetRootUIDGID(archiver.UIDMaps, archiver.GIDMaps) - if err != nil { - return err - } + rootIDs := archiver.IDMappingsVar.RootPair() // Create dst, copy src's content into it logrus.Debugf("Creating dest directory: %s", dst) - if err := idtools.MkdirAllNewAs(dst, 0755, rootUID, rootGID); err != nil { + if err := idtools.MkdirAllAndChownNew(dst, 0755, rootIDs); err != nil { return err } logrus.Debugf("Calling TarUntar(%s, %s)", src, dst) return archiver.TarUntar(src, dst) } -// CopyWithTar creates a tar archive of filesystem path `src`, and -// unpacks it at filesystem path `dst`. -// The archive is streamed directly with fixed buffering and no -// intermediary disk IO. -func CopyWithTar(src, dst string) error { - return defaultArchiver.CopyWithTar(src, dst) -} - // CopyFileWithTar emulates the behavior of the 'cp' command-line // for a single file. It copies a regular file from path `src` to // path `dst`, and preserves all its metadata. @@ -935,61 +1141,53 @@ func (archiver *Archiver) CopyFileWithTar(src, dst string) (err error) { dst = filepath.Join(dst, filepath.Base(src)) } // Create the holding directory if necessary - if err := system.MkdirAll(filepath.Dir(dst), 0700); err != nil { + if err := system.MkdirAll(filepath.Dir(dst), 0700, ""); err != nil { return err } r, w := io.Pipe() - errC := promise.Go(func() error { - defer w.Close() + errC := make(chan error, 1) - srcF, err := os.Open(src) - if err != nil { - return err - } - defer srcF.Close() + go func() { + defer close(errC) - hdr, err := tar.FileInfoHeader(srcSt, "") - if err != nil { - return err - } - hdr.Name = filepath.Base(dst) - hdr.Mode = int64(chmodTarEntry(os.FileMode(hdr.Mode))) + errC <- func() error { + defer w.Close() - remappedRootUID, remappedRootGID, err := idtools.GetRootUIDGID(archiver.UIDMaps, archiver.GIDMaps) - if err != nil { - return err - } - - // only perform mapping if the file being copied isn't already owned by the - // uid or gid of the remapped root in the container - if remappedRootUID != hdr.Uid { - xUID, err := idtools.ToHost(hdr.Uid, archiver.UIDMaps) + srcF, err := os.Open(src) if err != nil { return err } - hdr.Uid = xUID - } - if remappedRootGID != hdr.Gid { - xGID, err := idtools.ToHost(hdr.Gid, archiver.GIDMaps) + defer srcF.Close() + + hdr, err := tar.FileInfoHeader(srcSt, "") if err != nil { return err } - hdr.Gid = xGID - } + hdr.Format = tar.FormatPAX + hdr.ModTime = hdr.ModTime.Truncate(time.Second) + hdr.AccessTime = time.Time{} + hdr.ChangeTime = time.Time{} + hdr.Name = filepath.Base(dst) + hdr.Mode = int64(chmodTarEntry(os.FileMode(hdr.Mode))) - tw := tar.NewWriter(w) - defer tw.Close() - if err := tw.WriteHeader(hdr); err != nil { - return err - } - if _, err := io.Copy(tw, srcF); err != nil { - return err - } - return nil - }) + if err := remapIDs(archiver.IDMappingsVar, hdr); err != nil { + return err + } + + tw := tar.NewWriter(w) + defer tw.Close() + if err := tw.WriteHeader(hdr); err != nil { + return err + } + if _, err := io.Copy(tw, srcF); err != nil { + return err + } + return nil + }() + }() defer func() { - if er := <-errC; err != nil { + if er := <-errC; err == nil && er != nil { err = er } }() @@ -1001,23 +1199,21 @@ func (archiver *Archiver) CopyFileWithTar(src, dst string) (err error) { return err } -// CopyFileWithTar emulates the behavior of the 'cp' command-line -// for a single file. It copies a regular file from path `src` to -// path `dst`, and preserves all its metadata. -// -// Destination handling is in an operating specific manner depending -// where the daemon is running. If `dst` ends with a trailing slash -// the final destination path will be `dst/base(src)` (Linux) or -// `dst\base(src)` (Windows). -func CopyFileWithTar(src, dst string) (err error) { - return defaultArchiver.CopyFileWithTar(src, dst) +// IDMappings returns the IDMappings of the archiver. +func (archiver *Archiver) IDMappings() *idtools.IDMappings { + return archiver.IDMappingsVar +} + +func remapIDs(idMappings *idtools.IDMappings, hdr *tar.Header) error { + ids, err := idMappings.ToHost(idtools.IDPair{UID: hdr.Uid, GID: hdr.Gid}) + hdr.Uid, hdr.Gid = ids.UID, ids.GID + return err } // cmdStream executes a command, and returns its stdout as a stream. // If the command fails to run or doesn't complete successfully, an error // will be returned, including anything written on stderr. -func cmdStream(cmd *exec.Cmd, input io.Reader) (io.ReadCloser, <-chan struct{}, error) { - chdone := make(chan struct{}) +func cmdStream(cmd *exec.Cmd, input io.Reader) (io.ReadCloser, error) { cmd.Stdin = input pipeR, pipeW := io.Pipe() cmd.Stdout = pipeW @@ -1026,7 +1222,7 @@ func cmdStream(cmd *exec.Cmd, input io.Reader) (io.ReadCloser, <-chan struct{}, // Run the command and return the pipe if err := cmd.Start(); err != nil { - return nil, nil, err + return nil, err } // Copy stdout to the returned pipe @@ -1036,16 +1232,15 @@ func cmdStream(cmd *exec.Cmd, input io.Reader) (io.ReadCloser, <-chan struct{}, } else { pipeW.Close() } - close(chdone) }() - return pipeR, chdone, nil + return pipeR, nil } // NewTempArchive reads the content of src into a temporary file, and returns the contents // of that file as an archive. The archive can only be read once - as soon as reading completes, // the file will be deleted. -func NewTempArchive(src Archive, dir string) (*TempArchive, error) { +func NewTempArchive(src io.Reader, dir string) (*TempArchive, error) { f, err := ioutil.TempFile(dir, "") if err != nil { return nil, err diff --git a/vendor/github.com/docker/docker/pkg/archive/archive_linux.go b/vendor/github.com/docker/docker/pkg/archive/archive_linux.go new file mode 100644 index 000000000..970d4d068 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/archive/archive_linux.go @@ -0,0 +1,92 @@ +package archive // import "github.com/docker/docker/pkg/archive" + +import ( + "archive/tar" + "os" + "path/filepath" + "strings" + + "github.com/docker/docker/pkg/system" + "golang.org/x/sys/unix" +) + +func getWhiteoutConverter(format WhiteoutFormat) tarWhiteoutConverter { + if format == OverlayWhiteoutFormat { + return overlayWhiteoutConverter{} + } + return nil +} + +type overlayWhiteoutConverter struct{} + +func (overlayWhiteoutConverter) ConvertWrite(hdr *tar.Header, path string, fi os.FileInfo) (wo *tar.Header, err error) { + // convert whiteouts to AUFS format + if fi.Mode()&os.ModeCharDevice != 0 && hdr.Devmajor == 0 && hdr.Devminor == 0 { + // we just rename the file and make it normal + dir, filename := filepath.Split(hdr.Name) + hdr.Name = filepath.Join(dir, WhiteoutPrefix+filename) + hdr.Mode = 0600 + hdr.Typeflag = tar.TypeReg + hdr.Size = 0 + } + + if fi.Mode()&os.ModeDir != 0 { + // convert opaque dirs to AUFS format by writing an empty file with the prefix + opaque, err := system.Lgetxattr(path, "trusted.overlay.opaque") + if err != nil { + return nil, err + } + if len(opaque) == 1 && opaque[0] == 'y' { + if hdr.Xattrs != nil { + delete(hdr.Xattrs, "trusted.overlay.opaque") + } + + // create a header for the whiteout file + // it should inherit some properties from the parent, but be a regular file + wo = &tar.Header{ + Typeflag: tar.TypeReg, + Mode: hdr.Mode & int64(os.ModePerm), + Name: filepath.Join(hdr.Name, WhiteoutOpaqueDir), + Size: 0, + Uid: hdr.Uid, + Uname: hdr.Uname, + Gid: hdr.Gid, + Gname: hdr.Gname, + AccessTime: hdr.AccessTime, + ChangeTime: hdr.ChangeTime, + } + } + } + + return +} + +func (overlayWhiteoutConverter) ConvertRead(hdr *tar.Header, path string) (bool, error) { + base := filepath.Base(path) + dir := filepath.Dir(path) + + // if a directory is marked as opaque by the AUFS special file, we need to translate that to overlay + if base == WhiteoutOpaqueDir { + err := unix.Setxattr(dir, "trusted.overlay.opaque", []byte{'y'}, 0) + // don't write the file itself + return false, err + } + + // if a file was deleted and we are using overlay, we need to create a character device + if strings.HasPrefix(base, WhiteoutPrefix) { + originalBase := base[len(WhiteoutPrefix):] + originalPath := filepath.Join(dir, originalBase) + + if err := unix.Mknod(originalPath, unix.S_IFCHR, 0); err != nil { + return false, err + } + if err := os.Chown(originalPath, hdr.Uid, hdr.Gid); err != nil { + return false, err + } + + // don't write the file itself + return false, nil + } + + return true, nil +} diff --git a/vendor/github.com/docker/docker/pkg/archive/archive_other.go b/vendor/github.com/docker/docker/pkg/archive/archive_other.go new file mode 100644 index 000000000..462dfc632 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/archive/archive_other.go @@ -0,0 +1,7 @@ +// +build !linux + +package archive // import "github.com/docker/docker/pkg/archive" + +func getWhiteoutConverter(format WhiteoutFormat) tarWhiteoutConverter { + return nil +} diff --git a/vendor/github.com/docker/docker/pkg/archive/archive_unix.go b/vendor/github.com/docker/docker/pkg/archive/archive_unix.go index fbc3bb8c4..e81076c17 100644 --- a/vendor/github.com/docker/docker/pkg/archive/archive_unix.go +++ b/vendor/github.com/docker/docker/pkg/archive/archive_unix.go @@ -1,6 +1,6 @@ // +build !windows -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "archive/tar" @@ -9,7 +9,10 @@ import ( "path/filepath" "syscall" + "github.com/docker/docker/pkg/idtools" "github.com/docker/docker/pkg/system" + rsystem "github.com/opencontainers/runc/libcontainer/system" + "golang.org/x/sys/unix" ) // fixVolumePathPrefix does platform specific processing to ensure that if @@ -40,60 +43,59 @@ func chmodTarEntry(perm os.FileMode) os.FileMode { return perm // noop for unix as golang APIs provide perm bits correctly } -func setHeaderForSpecialDevice(hdr *tar.Header, ta *tarAppender, name string, stat interface{}) (inode uint64, err error) { +func setHeaderForSpecialDevice(hdr *tar.Header, name string, stat interface{}) (err error) { s, ok := stat.(*syscall.Stat_t) - if !ok { - err = errors.New("cannot convert stat value to syscall.Stat_t") - return - } - - inode = uint64(s.Ino) - - // Currently go does not fill in the major/minors - if s.Mode&syscall.S_IFBLK != 0 || - s.Mode&syscall.S_IFCHR != 0 { - hdr.Devmajor = int64(major(uint64(s.Rdev))) - hdr.Devminor = int64(minor(uint64(s.Rdev))) + if ok { + // Currently go does not fill in the major/minors + if s.Mode&unix.S_IFBLK != 0 || + s.Mode&unix.S_IFCHR != 0 { + hdr.Devmajor = int64(unix.Major(uint64(s.Rdev))) // nolint: unconvert + hdr.Devminor = int64(unix.Minor(uint64(s.Rdev))) // nolint: unconvert + } } return } -func getFileUIDGID(stat interface{}) (int, int, error) { +func getInodeFromStat(stat interface{}) (inode uint64, err error) { + s, ok := stat.(*syscall.Stat_t) + + if ok { + inode = s.Ino + } + + return +} + +func getFileUIDGID(stat interface{}) (idtools.IDPair, error) { s, ok := stat.(*syscall.Stat_t) if !ok { - return -1, -1, errors.New("cannot convert stat value to syscall.Stat_t") + return idtools.IDPair{}, errors.New("cannot convert stat value to syscall.Stat_t") } - return int(s.Uid), int(s.Gid), nil -} - -func major(device uint64) uint64 { - return (device >> 8) & 0xfff -} - -func minor(device uint64) uint64 { - return (device & 0xff) | ((device >> 12) & 0xfff00) + return idtools.IDPair{UID: int(s.Uid), GID: int(s.Gid)}, nil } // handleTarTypeBlockCharFifo is an OS-specific helper function used by // createTarFile to handle the following types of header: Block; Char; Fifo func handleTarTypeBlockCharFifo(hdr *tar.Header, path string) error { + if rsystem.RunningInUserNS() { + // cannot create a device if running in user namespace + return nil + } + mode := uint32(hdr.Mode & 07777) switch hdr.Typeflag { case tar.TypeBlock: - mode |= syscall.S_IFBLK + mode |= unix.S_IFBLK case tar.TypeChar: - mode |= syscall.S_IFCHR + mode |= unix.S_IFCHR case tar.TypeFifo: - mode |= syscall.S_IFIFO + mode |= unix.S_IFIFO } - if err := system.Mknod(path, mode, int(system.Mkdev(hdr.Devmajor, hdr.Devminor))); err != nil { - return err - } - return nil + return system.Mknod(path, mode, int(system.Mkdev(hdr.Devmajor, hdr.Devminor))) } func handleLChmod(hdr *tar.Header, path string, hdrInfo os.FileInfo) error { diff --git a/vendor/github.com/docker/docker/pkg/archive/archive_windows.go b/vendor/github.com/docker/docker/pkg/archive/archive_windows.go index 5c3a1be34..69aadd823 100644 --- a/vendor/github.com/docker/docker/pkg/archive/archive_windows.go +++ b/vendor/github.com/docker/docker/pkg/archive/archive_windows.go @@ -1,6 +1,4 @@ -// +build windows - -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "archive/tar" @@ -9,6 +7,7 @@ import ( "path/filepath" "strings" + "github.com/docker/docker/pkg/idtools" "github.com/docker/docker/pkg/longpath" ) @@ -42,15 +41,23 @@ func CanonicalTarNameForPath(p string) (string, error) { // chmodTarEntry is used to adjust the file permissions used in tar header based // on the platform the archival is done. func chmodTarEntry(perm os.FileMode) os.FileMode { - perm &= 0755 + //perm &= 0755 // this 0-ed out tar flags (like link, regular file, directory marker etc.) + permPart := perm & os.ModePerm + noPermPart := perm &^ os.ModePerm // Add the x bit: make everything +x from windows - perm |= 0111 + permPart |= 0111 + permPart &= 0755 - return perm + return noPermPart | permPart } -func setHeaderForSpecialDevice(hdr *tar.Header, ta *tarAppender, name string, stat interface{}) (inode uint64, err error) { - // do nothing. no notion of Rdev, Inode, Nlink in stat on Windows +func setHeaderForSpecialDevice(hdr *tar.Header, name string, stat interface{}) (err error) { + // do nothing. no notion of Rdev, Nlink in stat on Windows + return +} + +func getInodeFromStat(stat interface{}) (inode uint64, err error) { + // do nothing. no notion of Inode in stat on Windows return } @@ -64,7 +71,7 @@ func handleLChmod(hdr *tar.Header, path string, hdrInfo os.FileInfo) error { return nil } -func getFileUIDGID(stat interface{}) (int, int, error) { +func getFileUIDGID(stat interface{}) (idtools.IDPair, error) { // no notion of file ownership mapping yet on Windows - return 0, 0, nil + return idtools.IDPair{UID: 0, GID: 0}, nil } diff --git a/vendor/github.com/docker/docker/pkg/archive/changes.go b/vendor/github.com/docker/docker/pkg/archive/changes.go index 81651c61d..43734db5b 100644 --- a/vendor/github.com/docker/docker/pkg/archive/changes.go +++ b/vendor/github.com/docker/docker/pkg/archive/changes.go @@ -1,4 +1,4 @@ -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "archive/tar" @@ -13,10 +13,10 @@ import ( "syscall" "time" - "github.com/Sirupsen/logrus" "github.com/docker/docker/pkg/idtools" "github.com/docker/docker/pkg/pools" "github.com/docker/docker/pkg/system" + "github.com/sirupsen/logrus" ) // ChangeType represents the change type. @@ -81,6 +81,33 @@ func sameFsTimeSpec(a, b syscall.Timespec) bool { // Changes walks the path rw and determines changes for the files in the path, // with respect to the parent layers func Changes(layers []string, rw string) ([]Change, error) { + return changes(layers, rw, aufsDeletedFile, aufsMetadataSkip) +} + +func aufsMetadataSkip(path string) (skip bool, err error) { + skip, err = filepath.Match(string(os.PathSeparator)+WhiteoutMetaPrefix+"*", path) + if err != nil { + skip = true + } + return +} + +func aufsDeletedFile(root, path string, fi os.FileInfo) (string, error) { + f := filepath.Base(path) + + // If there is a whiteout, then the file was removed + if strings.HasPrefix(f, WhiteoutPrefix) { + originalFile := f[len(WhiteoutPrefix):] + return filepath.Join(filepath.Dir(path), originalFile), nil + } + + return "", nil +} + +type skipChange func(string) (bool, error) +type deleteChange func(string, string, os.FileInfo) (string, error) + +func changes(layers []string, rw string, dc deleteChange, sc skipChange) ([]Change, error) { var ( changes []Change changedDirs = make(map[string]struct{}) @@ -105,21 +132,24 @@ func Changes(layers []string, rw string) ([]Change, error) { return nil } - // Skip AUFS metadata - if matched, err := filepath.Match(string(os.PathSeparator)+WhiteoutMetaPrefix+"*", path); err != nil || matched { - return err + if sc != nil { + if skip, err := sc(path); skip { + return err + } } change := Change{ Path: path, } + deletedFile, err := dc(rw, path, f) + if err != nil { + return err + } + // Find out what kind of modification happened - file := filepath.Base(path) - // If there is a whiteout, then the file was removed - if strings.HasPrefix(file, WhiteoutPrefix) { - originalFile := file[len(WhiteoutPrefix):] - change.Path = filepath.Join(filepath.Dir(path), originalFile) + if deletedFile != "" { + change.Path = deletedFile change.Kind = ChangeDelete } else { // Otherwise, the file was added @@ -237,7 +267,7 @@ func (info *FileInfo) addChanges(oldInfo *FileInfo, changes *[]Change) { } for name, newChild := range info.children { - oldChild, _ := oldChildren[name] + oldChild := oldChildren[name] if oldChild != nil { // change? oldStat := oldChild.stat @@ -249,7 +279,7 @@ func (info *FileInfo) addChanges(oldInfo *FileInfo, changes *[]Change) { // breaks down is if some code intentionally hides a change by setting // back mtime if statDifferent(oldStat, newStat) || - bytes.Compare(oldChild.capability, newChild.capability) != 0 { + !bytes.Equal(oldChild.capability, newChild.capability) { change := Change{ Path: newChild.path(), Kind: ChangeModify, @@ -361,16 +391,11 @@ func ChangesSize(newDir string, changes []Change) int64 { } // ExportChanges produces an Archive from the provided changes, relative to dir. -func ExportChanges(dir string, changes []Change, uidMaps, gidMaps []idtools.IDMap) (Archive, error) { +func ExportChanges(dir string, changes []Change, uidMaps, gidMaps []idtools.IDMap) (io.ReadCloser, error) { reader, writer := io.Pipe() go func() { - ta := &tarAppender{ - TarWriter: tar.NewWriter(writer), - Buffer: pools.BufioWriter32KPool.Get(nil), - SeenFiles: make(map[uint64]string), - UIDMaps: uidMaps, - GIDMaps: gidMaps, - } + ta := newTarAppender(idtools.NewIDMappingsFromMaps(uidMaps, gidMaps), writer, nil) + // this buffer is needed for the duration of this piped stream defer pools.BufioWriter32KPool.Put(ta.Buffer) diff --git a/vendor/github.com/docker/docker/pkg/archive/changes_linux.go b/vendor/github.com/docker/docker/pkg/archive/changes_linux.go index dee8b7c60..f8792b3d4 100644 --- a/vendor/github.com/docker/docker/pkg/archive/changes_linux.go +++ b/vendor/github.com/docker/docker/pkg/archive/changes_linux.go @@ -1,4 +1,4 @@ -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "bytes" @@ -10,6 +10,7 @@ import ( "unsafe" "github.com/docker/docker/pkg/system" + "golang.org/x/sys/unix" ) // walker is used to implement collectFileInfoForChanges on linux. Where this @@ -65,7 +66,7 @@ func walkchunk(path string, fi os.FileInfo, dir string, root *FileInfo) error { } parent := root.LookUp(filepath.Dir(path)) if parent == nil { - return fmt.Errorf("collectFileInfoForChanges: Unexpectedly no parent for %s", path) + return fmt.Errorf("walkchunk: Unexpectedly no parent for %s", path) } info := &FileInfo{ name: filepath.Base(path), @@ -233,7 +234,7 @@ func readdirnames(dirname string) (names []nameIno, err error) { // Refill the buffer if necessary if bufp >= nbuf { bufp = 0 - nbuf, err = syscall.ReadDirent(int(f.Fd()), buf) // getdents on linux + nbuf, err = unix.ReadDirent(int(f.Fd()), buf) // getdents on linux if nbuf < 0 { nbuf = 0 } @@ -255,12 +256,12 @@ func readdirnames(dirname string) (names []nameIno, err error) { return sl, nil } -// parseDirent is a minor modification of syscall.ParseDirent (linux version) +// parseDirent is a minor modification of unix.ParseDirent (linux version) // which returns {name,inode} pairs instead of just names. func parseDirent(buf []byte, names []nameIno) (consumed int, newnames []nameIno) { origlen := len(buf) for len(buf) > 0 { - dirent := (*syscall.Dirent)(unsafe.Pointer(&buf[0])) + dirent := (*unix.Dirent)(unsafe.Pointer(&buf[0])) buf = buf[dirent.Reclen:] if dirent.Ino == 0 { // File absent in directory. continue diff --git a/vendor/github.com/docker/docker/pkg/archive/changes_other.go b/vendor/github.com/docker/docker/pkg/archive/changes_other.go index da70ed37c..ba744741c 100644 --- a/vendor/github.com/docker/docker/pkg/archive/changes_other.go +++ b/vendor/github.com/docker/docker/pkg/archive/changes_other.go @@ -1,6 +1,6 @@ // +build !linux -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "fmt" diff --git a/vendor/github.com/docker/docker/pkg/archive/changes_unix.go b/vendor/github.com/docker/docker/pkg/archive/changes_unix.go index 3778b732c..c06a209d8 100644 --- a/vendor/github.com/docker/docker/pkg/archive/changes_unix.go +++ b/vendor/github.com/docker/docker/pkg/archive/changes_unix.go @@ -1,12 +1,13 @@ // +build !windows -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "os" "syscall" "github.com/docker/docker/pkg/system" + "golang.org/x/sys/unix" ) func statDifferent(oldStat *system.StatT, newStat *system.StatT) bool { @@ -16,7 +17,7 @@ func statDifferent(oldStat *system.StatT, newStat *system.StatT) bool { oldStat.GID() != newStat.GID() || oldStat.Rdev() != newStat.Rdev() || // Don't look at size for dirs, its not a good measure of change - (oldStat.Mode()&syscall.S_IFDIR != syscall.S_IFDIR && + (oldStat.Mode()&unix.S_IFDIR != unix.S_IFDIR && (!sameFsTimeSpec(oldStat.Mtim(), newStat.Mtim()) || (oldStat.Size() != newStat.Size()))) { return true } @@ -24,11 +25,11 @@ func statDifferent(oldStat *system.StatT, newStat *system.StatT) bool { } func (info *FileInfo) isDir() bool { - return info.parent == nil || info.stat.Mode()&syscall.S_IFDIR != 0 + return info.parent == nil || info.stat.Mode()&unix.S_IFDIR != 0 } func getIno(fi os.FileInfo) uint64 { - return uint64(fi.Sys().(*syscall.Stat_t).Ino) + return fi.Sys().(*syscall.Stat_t).Ino } func hasHardlinks(fi os.FileInfo) bool { diff --git a/vendor/github.com/docker/docker/pkg/archive/changes_windows.go b/vendor/github.com/docker/docker/pkg/archive/changes_windows.go index af94243fc..6555c0136 100644 --- a/vendor/github.com/docker/docker/pkg/archive/changes_windows.go +++ b/vendor/github.com/docker/docker/pkg/archive/changes_windows.go @@ -1,4 +1,4 @@ -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "os" @@ -9,16 +9,16 @@ import ( func statDifferent(oldStat *system.StatT, newStat *system.StatT) bool { // Don't look at size for dirs, its not a good measure of change - if oldStat.ModTime() != newStat.ModTime() || + if oldStat.Mtim() != newStat.Mtim() || oldStat.Mode() != newStat.Mode() || - oldStat.Size() != newStat.Size() && !oldStat.IsDir() { + oldStat.Size() != newStat.Size() && !oldStat.Mode().IsDir() { return true } return false } func (info *FileInfo) isDir() bool { - return info.parent == nil || info.stat.IsDir() + return info.parent == nil || info.stat.Mode().IsDir() } func getIno(fi os.FileInfo) (inode uint64) { diff --git a/vendor/github.com/docker/docker/pkg/archive/copy.go b/vendor/github.com/docker/docker/pkg/archive/copy.go index e1fa73f37..d0f13ca79 100644 --- a/vendor/github.com/docker/docker/pkg/archive/copy.go +++ b/vendor/github.com/docker/docker/pkg/archive/copy.go @@ -1,4 +1,4 @@ -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "archive/tar" @@ -9,8 +9,8 @@ import ( "path/filepath" "strings" - "github.com/Sirupsen/logrus" "github.com/docker/docker/pkg/system" + "github.com/sirupsen/logrus" ) // Errors used or returned by this file. @@ -27,23 +27,23 @@ var ( // path (from before being processed by utility functions from the path or // filepath stdlib packages) ends with a trailing `/.` or `/`. If the cleaned // path already ends in a `.` path segment, then another is not added. If the -// clean path already ends in a path separator, then another is not added. -func PreserveTrailingDotOrSeparator(cleanedPath, originalPath string) string { +// clean path already ends in the separator, then another is not added. +func PreserveTrailingDotOrSeparator(cleanedPath string, originalPath string, sep byte) string { // Ensure paths are in platform semantics - cleanedPath = normalizePath(cleanedPath) - originalPath = normalizePath(originalPath) + cleanedPath = strings.Replace(cleanedPath, "/", string(sep), -1) + originalPath = strings.Replace(originalPath, "/", string(sep), -1) if !specifiesCurrentDir(cleanedPath) && specifiesCurrentDir(originalPath) { - if !hasTrailingPathSeparator(cleanedPath) { + if !hasTrailingPathSeparator(cleanedPath, sep) { // Add a separator if it doesn't already end with one (a cleaned // path would only end in a separator if it is the root). - cleanedPath += string(filepath.Separator) + cleanedPath += string(sep) } cleanedPath += "." } - if !hasTrailingPathSeparator(cleanedPath) && hasTrailingPathSeparator(originalPath) { - cleanedPath += string(filepath.Separator) + if !hasTrailingPathSeparator(cleanedPath, sep) && hasTrailingPathSeparator(originalPath, sep) { + cleanedPath += string(sep) } return cleanedPath @@ -52,14 +52,14 @@ func PreserveTrailingDotOrSeparator(cleanedPath, originalPath string) string { // assertsDirectory returns whether the given path is // asserted to be a directory, i.e., the path ends with // a trailing '/' or `/.`, assuming a path separator of `/`. -func assertsDirectory(path string) bool { - return hasTrailingPathSeparator(path) || specifiesCurrentDir(path) +func assertsDirectory(path string, sep byte) bool { + return hasTrailingPathSeparator(path, sep) || specifiesCurrentDir(path) } // hasTrailingPathSeparator returns whether the given // path ends with the system's path separator character. -func hasTrailingPathSeparator(path string) bool { - return len(path) > 0 && os.IsPathSeparator(path[len(path)-1]) +func hasTrailingPathSeparator(path string, sep byte) bool { + return len(path) > 0 && path[len(path)-1] == sep } // specifiesCurrentDir returns whether the given path specifies @@ -72,10 +72,10 @@ func specifiesCurrentDir(path string) bool { // basename by first cleaning the path but preserves a trailing "." if the // original path specified the current directory. func SplitPathDirEntry(path string) (dir, base string) { - cleanedPath := filepath.Clean(normalizePath(path)) + cleanedPath := filepath.Clean(filepath.FromSlash(path)) if specifiesCurrentDir(path) { - cleanedPath += string(filepath.Separator) + "." + cleanedPath += string(os.PathSeparator) + "." } return filepath.Dir(cleanedPath), filepath.Base(cleanedPath) @@ -88,13 +88,13 @@ func SplitPathDirEntry(path string) (dir, base string) { // This function acts as a convenient wrapper around TarWithOptions, which // requires a directory as the source path. TarResource accepts either a // directory or a file path and correctly sets the Tar options. -func TarResource(sourceInfo CopyInfo) (content Archive, err error) { +func TarResource(sourceInfo CopyInfo) (content io.ReadCloser, err error) { return TarResourceRebase(sourceInfo.Path, sourceInfo.RebaseName) } // TarResourceRebase is like TarResource but renames the first path element of // items in the resulting tar archive to match the given rebaseName if not "". -func TarResourceRebase(sourcePath, rebaseName string) (content Archive, err error) { +func TarResourceRebase(sourcePath, rebaseName string) (content io.ReadCloser, err error) { sourcePath = normalizePath(sourcePath) if _, err = os.Lstat(sourcePath); err != nil { // Catches the case where the source does not exist or is not a @@ -103,22 +103,27 @@ func TarResourceRebase(sourcePath, rebaseName string) (content Archive, err erro return } - // Separate the source path between it's directory and + // Separate the source path between its directory and // the entry in that directory which we are archiving. sourceDir, sourceBase := SplitPathDirEntry(sourcePath) - - filter := []string{sourceBase} + opts := TarResourceRebaseOpts(sourceBase, rebaseName) logrus.Debugf("copying %q from %q", sourceBase, sourceDir) + return TarWithOptions(sourceDir, opts) +} - return TarWithOptions(sourceDir, &TarOptions{ +// TarResourceRebaseOpts does not preform the Tar, but instead just creates the rebase +// parameters to be sent to TarWithOptions (the TarOptions struct) +func TarResourceRebaseOpts(sourceBase string, rebaseName string) *TarOptions { + filter := []string{sourceBase} + return &TarOptions{ Compression: Uncompressed, IncludeFiles: filter, IncludeSourceDir: true, RebaseNames: map[string]string{ sourceBase: rebaseName, }, - }) + } } // CopyInfo holds basic info about the source @@ -218,7 +223,7 @@ func CopyInfoDestinationPath(path string) (info CopyInfo, err error) { // Ensure destination parent dir exists. dstParent, _ := SplitPathDirEntry(path) - parentDirStat, err := os.Lstat(dstParent) + parentDirStat, err := os.Stat(dstParent) if err != nil { return CopyInfo{}, err } @@ -241,7 +246,7 @@ func CopyInfoDestinationPath(path string) (info CopyInfo, err error) { // contain the archived resource described by srcInfo, to the destination // described by dstInfo. Returns the possibly modified content archive along // with the path to the destination directory which it should be extracted to. -func PrepareArchiveCopy(srcContent Reader, srcInfo, dstInfo CopyInfo) (dstDir string, content Archive, err error) { +func PrepareArchiveCopy(srcContent io.Reader, srcInfo, dstInfo CopyInfo) (dstDir string, content io.ReadCloser, err error) { // Ensure in platform semantics srcInfo.Path = normalizePath(srcInfo.Path) dstInfo.Path = normalizePath(dstInfo.Path) @@ -281,7 +286,7 @@ func PrepareArchiveCopy(srcContent Reader, srcInfo, dstInfo CopyInfo) (dstDir st srcBase = srcInfo.RebaseName } return dstDir, RebaseArchiveEntries(srcContent, srcBase, dstBase), nil - case assertsDirectory(dstInfo.Path): + case assertsDirectory(dstInfo.Path, os.PathSeparator): // The destination does not exist and is asserted to be created as a // directory, but the source content is not a directory. This is an // error condition since you cannot create a directory from a file @@ -304,7 +309,7 @@ func PrepareArchiveCopy(srcContent Reader, srcInfo, dstInfo CopyInfo) (dstDir st // RebaseArchiveEntries rewrites the given srcContent archive replacing // an occurrence of oldBase with newBase at the beginning of entry names. -func RebaseArchiveEntries(srcContent Reader, oldBase, newBase string) Archive { +func RebaseArchiveEntries(srcContent io.Reader, oldBase, newBase string) io.ReadCloser { if oldBase == string(os.PathSeparator) { // If oldBase specifies the root directory, use an empty string as // oldBase instead so that newBase doesn't replace the path separator @@ -332,6 +337,9 @@ func RebaseArchiveEntries(srcContent Reader, oldBase, newBase string) Archive { } hdr.Name = strings.Replace(hdr.Name, oldBase, newBase, 1) + if hdr.Typeflag == tar.TypeLink { + hdr.Linkname = strings.Replace(hdr.Linkname, oldBase, newBase, 1) + } if err = rebasedTar.WriteHeader(hdr); err != nil { w.CloseWithError(err) @@ -348,6 +356,9 @@ func RebaseArchiveEntries(srcContent Reader, oldBase, newBase string) Archive { return rebased } +// TODO @gupta-ak. These might have to be changed in the future to be +// continuity driver aware as well to support LCOW. + // CopyResource performs an archive copy from the given source path to the // given destination path. The source path MUST exist and the destination // path's parent directory must exist. @@ -362,8 +373,8 @@ func CopyResource(srcPath, dstPath string, followLink bool) error { dstPath = normalizePath(dstPath) // Clean the source and destination paths. - srcPath = PreserveTrailingDotOrSeparator(filepath.Clean(srcPath), srcPath) - dstPath = PreserveTrailingDotOrSeparator(filepath.Clean(dstPath), dstPath) + srcPath = PreserveTrailingDotOrSeparator(filepath.Clean(srcPath), srcPath, os.PathSeparator) + dstPath = PreserveTrailingDotOrSeparator(filepath.Clean(dstPath), dstPath, os.PathSeparator) if srcInfo, err = CopyInfoSourcePath(srcPath, followLink); err != nil { return err @@ -380,7 +391,7 @@ func CopyResource(srcPath, dstPath string, followLink bool) error { // CopyTo handles extracting the given content whose // entries should be sourced from srcInfo to dstPath. -func CopyTo(content Reader, srcInfo CopyInfo, dstPath string) error { +func CopyTo(content io.Reader, srcInfo CopyInfo, dstPath string) error { // The destination path need not exist, but CopyInfoDestinationPath will // ensure that at least the parent directory exists. dstInfo, err := CopyInfoDestinationPath(normalizePath(dstPath)) @@ -426,7 +437,8 @@ func ResolveHostSourcePath(path string, followLink bool) (resolvedPath, rebaseNa // resolvedDirPath will have been cleaned (no trailing path separators) so // we can manually join it with the base path element. resolvedPath = resolvedDirPath + string(filepath.Separator) + basePath - if hasTrailingPathSeparator(path) && filepath.Base(path) != filepath.Base(resolvedPath) { + if hasTrailingPathSeparator(path, os.PathSeparator) && + filepath.Base(path) != filepath.Base(resolvedPath) { rebaseName = filepath.Base(path) } } @@ -439,11 +451,13 @@ func GetRebaseName(path, resolvedPath string) (string, string) { // linkTarget will have been cleaned (no trailing path separators and dot) so // we can manually join it with them var rebaseName string - if specifiesCurrentDir(path) && !specifiesCurrentDir(resolvedPath) { + if specifiesCurrentDir(path) && + !specifiesCurrentDir(resolvedPath) { resolvedPath += string(filepath.Separator) + "." } - if hasTrailingPathSeparator(path) && !hasTrailingPathSeparator(resolvedPath) { + if hasTrailingPathSeparator(path, os.PathSeparator) && + !hasTrailingPathSeparator(resolvedPath, os.PathSeparator) { resolvedPath += string(filepath.Separator) } diff --git a/vendor/github.com/docker/docker/pkg/archive/copy_unix.go b/vendor/github.com/docker/docker/pkg/archive/copy_unix.go index e305b5e4a..3958364f5 100644 --- a/vendor/github.com/docker/docker/pkg/archive/copy_unix.go +++ b/vendor/github.com/docker/docker/pkg/archive/copy_unix.go @@ -1,6 +1,6 @@ // +build !windows -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "path/filepath" diff --git a/vendor/github.com/docker/docker/pkg/archive/copy_windows.go b/vendor/github.com/docker/docker/pkg/archive/copy_windows.go index 2b775b45c..a878d1bac 100644 --- a/vendor/github.com/docker/docker/pkg/archive/copy_windows.go +++ b/vendor/github.com/docker/docker/pkg/archive/copy_windows.go @@ -1,4 +1,4 @@ -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "path/filepath" diff --git a/vendor/github.com/docker/docker/pkg/archive/diff.go b/vendor/github.com/docker/docker/pkg/archive/diff.go index 1b08ad33a..fae4b9de0 100644 --- a/vendor/github.com/docker/docker/pkg/archive/diff.go +++ b/vendor/github.com/docker/docker/pkg/archive/diff.go @@ -1,4 +1,4 @@ -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "archive/tar" @@ -10,16 +10,16 @@ import ( "runtime" "strings" - "github.com/Sirupsen/logrus" "github.com/docker/docker/pkg/idtools" "github.com/docker/docker/pkg/pools" "github.com/docker/docker/pkg/system" + "github.com/sirupsen/logrus" ) // UnpackLayer unpack `layer` to a `dest`. The stream `layer` can be // compressed or uncompressed. // Returns the size in bytes of the contents of the layer. -func UnpackLayer(dest string, layer Reader, options *TarOptions) (size int64, err error) { +func UnpackLayer(dest string, layer io.Reader, options *TarOptions) (size int64, err error) { tr := tar.NewReader(layer) trBuf := pools.BufioReader32KPool.Get(tr) defer pools.BufioReader32KPool.Put(trBuf) @@ -33,17 +33,11 @@ func UnpackLayer(dest string, layer Reader, options *TarOptions) (size int64, er if options.ExcludePatterns == nil { options.ExcludePatterns = []string{} } - remappedRootUID, remappedRootGID, err := idtools.GetRootUIDGID(options.UIDMaps, options.GIDMaps) - if err != nil { - return 0, err - } + idMappings := idtools.NewIDMappingsFromMaps(options.UIDMaps, options.GIDMaps) aufsTempdir := "" aufsHardlinks := make(map[string]*tar.Header) - if options == nil { - options = &TarOptions{} - } // Iterate through the files in the archive. for { hdr, err := tr.Next() @@ -90,7 +84,7 @@ func UnpackLayer(dest string, layer Reader, options *TarOptions) (size int64, er parentPath := filepath.Join(dest, parent) if _, err := os.Lstat(parentPath); err != nil && os.IsNotExist(err) { - err = system.MkdirAll(parentPath, 0600) + err = system.MkdirAll(parentPath, 0600, "") if err != nil { return 0, err } @@ -111,7 +105,7 @@ func UnpackLayer(dest string, layer Reader, options *TarOptions) (size int64, er } defer os.RemoveAll(aufsTempdir) } - if err := createTarFile(filepath.Join(aufsTempdir, basename), dest, hdr, tr, true, nil); err != nil { + if err := createTarFile(filepath.Join(aufsTempdir, basename), dest, hdr, tr, true, nil, options.InUserNS); err != nil { return 0, err } } @@ -198,28 +192,11 @@ func UnpackLayer(dest string, layer Reader, options *TarOptions) (size int64, er srcData = tmpFile } - // if the options contain a uid & gid maps, convert header uid/gid - // entries using the maps such that lchown sets the proper mapped - // uid/gid after writing the file. We only perform this mapping if - // the file isn't already owned by the remapped root UID or GID, as - // that specific uid/gid has no mapping from container -> host, and - // those files already have the proper ownership for inside the - // container. - if srcHdr.Uid != remappedRootUID { - xUID, err := idtools.ToHost(srcHdr.Uid, options.UIDMaps) - if err != nil { - return 0, err - } - srcHdr.Uid = xUID + if err := remapIDs(idMappings, srcHdr); err != nil { + return 0, err } - if srcHdr.Gid != remappedRootGID { - xGID, err := idtools.ToHost(srcHdr.Gid, options.GIDMaps) - if err != nil { - return 0, err - } - srcHdr.Gid = xGID - } - if err := createTarFile(path, dest, srcHdr, srcData, true, nil); err != nil { + + if err := createTarFile(path, dest, srcHdr, srcData, true, nil, options.InUserNS); err != nil { return 0, err } @@ -246,7 +223,7 @@ func UnpackLayer(dest string, layer Reader, options *TarOptions) (size int64, er // and applies it to the directory `dest`. The stream `layer` can be // compressed or uncompressed. // Returns the size in bytes of the contents of the layer. -func ApplyLayer(dest string, layer Reader) (int64, error) { +func ApplyLayer(dest string, layer io.Reader) (int64, error) { return applyLayerHandler(dest, layer, &TarOptions{}, true) } @@ -254,12 +231,12 @@ func ApplyLayer(dest string, layer Reader) (int64, error) { // `layer`, and applies it to the directory `dest`. The stream `layer` // can only be uncompressed. // Returns the size in bytes of the contents of the layer. -func ApplyUncompressedLayer(dest string, layer Reader, options *TarOptions) (int64, error) { +func ApplyUncompressedLayer(dest string, layer io.Reader, options *TarOptions) (int64, error) { return applyLayerHandler(dest, layer, options, false) } // do the bulk load of ApplyLayer, but allow for not calling DecompressStream -func applyLayerHandler(dest string, layer Reader, options *TarOptions, decompress bool) (int64, error) { +func applyLayerHandler(dest string, layer io.Reader, options *TarOptions, decompress bool) (int64, error) { dest = filepath.Clean(dest) // We need to be able to set any perms @@ -270,10 +247,12 @@ func applyLayerHandler(dest string, layer Reader, options *TarOptions, decompres defer system.Umask(oldmask) // ignore err, ErrNotSupportedPlatform if decompress { - layer, err = DecompressStream(layer) + decompLayer, err := DecompressStream(layer) if err != nil { return 0, err } + defer decompLayer.Close() + layer = decompLayer } return UnpackLayer(dest, layer, options) } diff --git a/vendor/github.com/docker/docker/pkg/archive/example_changes.go b/vendor/github.com/docker/docker/pkg/archive/example_changes.go index cedd46a40..495db809e 100644 --- a/vendor/github.com/docker/docker/pkg/archive/example_changes.go +++ b/vendor/github.com/docker/docker/pkg/archive/example_changes.go @@ -13,8 +13,8 @@ import ( "os" "path" - "github.com/Sirupsen/logrus" "github.com/docker/docker/pkg/archive" + "github.com/sirupsen/logrus" ) var ( diff --git a/vendor/github.com/docker/docker/pkg/archive/time_linux.go b/vendor/github.com/docker/docker/pkg/archive/time_linux.go index 3448569b1..797143ee8 100644 --- a/vendor/github.com/docker/docker/pkg/archive/time_linux.go +++ b/vendor/github.com/docker/docker/pkg/archive/time_linux.go @@ -1,4 +1,4 @@ -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "syscall" @@ -9,7 +9,7 @@ func timeToTimespec(time time.Time) (ts syscall.Timespec) { if time.IsZero() { // Return UTIME_OMIT special value ts.Sec = 0 - ts.Nsec = ((1 << 30) - 2) + ts.Nsec = (1 << 30) - 2 return } return syscall.NsecToTimespec(time.UnixNano()) diff --git a/vendor/github.com/docker/docker/pkg/archive/time_unsupported.go b/vendor/github.com/docker/docker/pkg/archive/time_unsupported.go index e85aac054..f58bf227f 100644 --- a/vendor/github.com/docker/docker/pkg/archive/time_unsupported.go +++ b/vendor/github.com/docker/docker/pkg/archive/time_unsupported.go @@ -1,6 +1,6 @@ // +build !linux -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "syscall" diff --git a/vendor/github.com/docker/docker/pkg/archive/whiteouts.go b/vendor/github.com/docker/docker/pkg/archive/whiteouts.go index d20478a10..4c072a87e 100644 --- a/vendor/github.com/docker/docker/pkg/archive/whiteouts.go +++ b/vendor/github.com/docker/docker/pkg/archive/whiteouts.go @@ -1,4 +1,4 @@ -package archive +package archive // import "github.com/docker/docker/pkg/archive" // Whiteouts are files with a special meaning for the layered filesystem. // Docker uses AUFS whiteout files inside exported archives. In other diff --git a/vendor/github.com/docker/docker/pkg/archive/wrap.go b/vendor/github.com/docker/docker/pkg/archive/wrap.go index dfb335c0b..85435694c 100644 --- a/vendor/github.com/docker/docker/pkg/archive/wrap.go +++ b/vendor/github.com/docker/docker/pkg/archive/wrap.go @@ -1,9 +1,9 @@ -package archive +package archive // import "github.com/docker/docker/pkg/archive" import ( "archive/tar" "bytes" - "io/ioutil" + "io" ) // Generate generates a new archive from the content provided @@ -22,7 +22,7 @@ import ( // // FIXME: stream content instead of buffering // FIXME: specify permissions and other archive metadata -func Generate(input ...string) (Archive, error) { +func Generate(input ...string) (io.Reader, error) { files := parseStringPairs(input...) buf := new(bytes.Buffer) tw := tar.NewWriter(buf) @@ -42,7 +42,7 @@ func Generate(input ...string) (Archive, error) { if err := tw.Close(); err != nil { return nil, err } - return ioutil.NopCloser(buf), nil + return buf, nil } func parseStringPairs(input ...string) (output [][2]string) { diff --git a/vendor/github.com/docker/docker/pkg/fileutils/fileutils.go b/vendor/github.com/docker/docker/pkg/fileutils/fileutils.go index 763d8d279..28cad499a 100644 --- a/vendor/github.com/docker/docker/pkg/fileutils/fileutils.go +++ b/vendor/github.com/docker/docker/pkg/fileutils/fileutils.go @@ -1,4 +1,4 @@ -package fileutils +package fileutils // import "github.com/docker/docker/pkg/fileutils" import ( "errors" @@ -10,101 +10,77 @@ import ( "strings" "text/scanner" - "github.com/Sirupsen/logrus" + "github.com/sirupsen/logrus" ) -// exclusion returns true if the specified pattern is an exclusion -func exclusion(pattern string) bool { - return pattern[0] == '!' +// PatternMatcher allows checking paths against a list of patterns +type PatternMatcher struct { + patterns []*Pattern + exclusions bool } -// empty returns true if the specified pattern is empty -func empty(pattern string) bool { - return pattern == "" -} - -// CleanPatterns takes a slice of patterns returns a new -// slice of patterns cleaned with filepath.Clean, stripped -// of any empty patterns and lets the caller know whether the -// slice contains any exception patterns (prefixed with !). -func CleanPatterns(patterns []string) ([]string, [][]string, bool, error) { - // Loop over exclusion patterns and: - // 1. Clean them up. - // 2. Indicate whether we are dealing with any exception rules. - // 3. Error if we see a single exclusion marker on it's own (!). - cleanedPatterns := []string{} - patternDirs := [][]string{} - exceptions := false - for _, pattern := range patterns { +// NewPatternMatcher creates a new matcher object for specific patterns that can +// be used later to match against patterns against paths +func NewPatternMatcher(patterns []string) (*PatternMatcher, error) { + pm := &PatternMatcher{ + patterns: make([]*Pattern, 0, len(patterns)), + } + for _, p := range patterns { // Eliminate leading and trailing whitespace. - pattern = strings.TrimSpace(pattern) - if empty(pattern) { + p = strings.TrimSpace(p) + if p == "" { continue } - if exclusion(pattern) { - if len(pattern) == 1 { - return nil, nil, false, errors.New("Illegal exclusion pattern: !") + p = filepath.Clean(p) + newp := &Pattern{} + if p[0] == '!' { + if len(p) == 1 { + return nil, errors.New("illegal exclusion pattern: \"!\"") } - exceptions = true + newp.exclusion = true + p = p[1:] + pm.exclusions = true } - pattern = filepath.Clean(pattern) - cleanedPatterns = append(cleanedPatterns, pattern) - if exclusion(pattern) { - pattern = pattern[1:] + // Do some syntax checking on the pattern. + // filepath's Match() has some really weird rules that are inconsistent + // so instead of trying to dup their logic, just call Match() for its + // error state and if there is an error in the pattern return it. + // If this becomes an issue we can remove this since its really only + // needed in the error (syntax) case - which isn't really critical. + if _, err := filepath.Match(p, "."); err != nil { + return nil, err } - patternDirs = append(patternDirs, strings.Split(pattern, string(os.PathSeparator))) + newp.cleanedPattern = p + newp.dirs = strings.Split(p, string(os.PathSeparator)) + pm.patterns = append(pm.patterns, newp) } - - return cleanedPatterns, patternDirs, exceptions, nil + return pm, nil } -// Matches returns true if file matches any of the patterns -// and isn't excluded by any of the subsequent patterns. -func Matches(file string, patterns []string) (bool, error) { - file = filepath.Clean(file) - - if file == "." { - // Don't let them exclude everything, kind of silly. - return false, nil - } - - patterns, patDirs, _, err := CleanPatterns(patterns) - if err != nil { - return false, err - } - - return OptimizedMatches(file, patterns, patDirs) -} - -// OptimizedMatches is basically the same as fileutils.Matches() but optimized for archive.go. -// It will assume that the inputs have been preprocessed and therefore the function -// doesn't need to do as much error checking and clean-up. This was done to avoid -// repeating these steps on each file being checked during the archive process. -// The more generic fileutils.Matches() can't make these assumptions. -func OptimizedMatches(file string, patterns []string, patDirs [][]string) (bool, error) { +// Matches matches path against all the patterns. Matches is not safe to be +// called concurrently +func (pm *PatternMatcher) Matches(file string) (bool, error) { matched := false file = filepath.FromSlash(file) parentPath := filepath.Dir(file) parentPathDirs := strings.Split(parentPath, string(os.PathSeparator)) - for i, pattern := range patterns { + for _, pattern := range pm.patterns { negative := false - if exclusion(pattern) { + if pattern.exclusion { negative = true - pattern = pattern[1:] } - match, err := regexpMatch(pattern, file) + match, err := pattern.match(file) if err != nil { - return false, fmt.Errorf("Error in pattern (%s): %s", pattern, err) + return false, err } if !match && parentPath != "." { // Check to see if the pattern matches one of our parent dirs. - if len(patDirs[i]) <= len(parentPathDirs) { - match, _ = regexpMatch(strings.Join(patDirs[i], string(os.PathSeparator)), - strings.Join(parentPathDirs[:len(patDirs[i])], string(os.PathSeparator))) + if len(pattern.dirs) <= len(parentPathDirs) { + match, _ = pattern.match(strings.Join(parentPathDirs[:len(pattern.dirs)], string(os.PathSeparator))) } } @@ -120,28 +96,49 @@ func OptimizedMatches(file string, patterns []string, patDirs [][]string) (bool, return matched, nil } -// regexpMatch tries to match the logic of filepath.Match but -// does so using regexp logic. We do this so that we can expand the -// wildcard set to include other things, like "**" to mean any number -// of directories. This means that we should be backwards compatible -// with filepath.Match(). We'll end up supporting more stuff, due to -// the fact that we're using regexp, but that's ok - it does no harm. -// -// As per the comment in golangs filepath.Match, on Windows, escaping -// is disabled. Instead, '\\' is treated as path separator. -func regexpMatch(pattern, path string) (bool, error) { - regStr := "^" +// Exclusions returns true if any of the patterns define exclusions +func (pm *PatternMatcher) Exclusions() bool { + return pm.exclusions +} - // Do some syntax checking on the pattern. - // filepath's Match() has some really weird rules that are inconsistent - // so instead of trying to dup their logic, just call Match() for its - // error state and if there is an error in the pattern return it. - // If this becomes an issue we can remove this since its really only - // needed in the error (syntax) case - which isn't really critical. - if _, err := filepath.Match(pattern, path); err != nil { - return false, err +// Patterns returns array of active patterns +func (pm *PatternMatcher) Patterns() []*Pattern { + return pm.patterns +} + +// Pattern defines a single regexp used used to filter file paths. +type Pattern struct { + cleanedPattern string + dirs []string + regexp *regexp.Regexp + exclusion bool +} + +func (p *Pattern) String() string { + return p.cleanedPattern +} + +// Exclusion returns true if this pattern defines exclusion +func (p *Pattern) Exclusion() bool { + return p.exclusion +} + +func (p *Pattern) match(path string) (bool, error) { + + if p.regexp == nil { + if err := p.compile(); err != nil { + return false, filepath.ErrBadPattern + } } + b := p.regexp.MatchString(path) + + return b, nil +} + +func (p *Pattern) compile() error { + regStr := "^" + pattern := p.cleanedPattern // Go through the pattern and convert it to a regexp. // We use a scanner so we can support utf-8 chars. var scan scanner.Scanner @@ -161,17 +158,19 @@ func regexpMatch(pattern, path string) (bool, error) { // is some flavor of "**" scan.Next() + // Treat **/ as ** so eat the "/" + if string(scan.Peek()) == sl { + scan.Next() + } + if scan.Peek() == scanner.EOF { // is "**EOF" - to align with .gitignore just accept all regStr += ".*" } else { // is "**" - regStr += "((.*" + escSL + ")|([^" + escSL + "]*))" - } - - // Treat **/ as ** so eat the "/" - if string(scan.Peek()) == sl { - scan.Next() + // Note that this allows for any # of /'s (even 0) because + // the .* will eat everything, even /'s + regStr += "(.*" + escSL + ")?" } } else { // is "*" so map it to anything but "/" @@ -180,7 +179,7 @@ func regexpMatch(pattern, path string) (bool, error) { } else if ch == '?' { // "?" is any char except "/" regStr += "[^" + escSL + "]" - } else if strings.Index(".$", string(ch)) != -1 { + } else if ch == '.' || ch == '$' { // Escape some regexp special chars that have no meaning // in golang's filepath.Match regStr += `\` + string(ch) @@ -206,14 +205,30 @@ func regexpMatch(pattern, path string) (bool, error) { regStr += "$" - res, err := regexp.MatchString(regStr, path) - - // Map regexp's error to filepath's so no one knows we're not using filepath + re, err := regexp.Compile(regStr) if err != nil { - err = filepath.ErrBadPattern + return err } - return res, err + p.regexp = re + return nil +} + +// Matches returns true if file matches any of the patterns +// and isn't excluded by any of the subsequent patterns. +func Matches(file string, patterns []string) (bool, error) { + pm, err := NewPatternMatcher(patterns) + if err != nil { + return false, err + } + file = filepath.Clean(file) + + if file == "." { + // Don't let them exclude everything, kind of silly. + return false, nil + } + + return pm.Matches(file) } // CopyFile copies from src to dst until either EOF is reached diff --git a/vendor/github.com/docker/docker/pkg/fileutils/fileutils_darwin.go b/vendor/github.com/docker/docker/pkg/fileutils/fileutils_darwin.go new file mode 100644 index 000000000..e40cc271b --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/fileutils/fileutils_darwin.go @@ -0,0 +1,27 @@ +package fileutils // import "github.com/docker/docker/pkg/fileutils" + +import ( + "os" + "os/exec" + "strconv" + "strings" +) + +// GetTotalUsedFds returns the number of used File Descriptors by +// executing `lsof -p PID` +func GetTotalUsedFds() int { + pid := os.Getpid() + + cmd := exec.Command("lsof", "-p", strconv.Itoa(pid)) + + output, err := cmd.CombinedOutput() + if err != nil { + return -1 + } + + outputStr := strings.TrimSpace(string(output)) + + fds := strings.Split(outputStr, "\n") + + return len(fds) - 1 +} diff --git a/vendor/github.com/docker/docker/pkg/fileutils/fileutils_solaris.go b/vendor/github.com/docker/docker/pkg/fileutils/fileutils_solaris.go deleted file mode 100644 index 0f2cb7ab9..000000000 --- a/vendor/github.com/docker/docker/pkg/fileutils/fileutils_solaris.go +++ /dev/null @@ -1,7 +0,0 @@ -package fileutils - -// GetTotalUsedFds Returns the number of used File Descriptors. -// On Solaris these limits are per process and not systemwide -func GetTotalUsedFds() int { - return -1 -} diff --git a/vendor/github.com/docker/docker/pkg/fileutils/fileutils_unix.go b/vendor/github.com/docker/docker/pkg/fileutils/fileutils_unix.go index d5c3abf56..565396f1c 100644 --- a/vendor/github.com/docker/docker/pkg/fileutils/fileutils_unix.go +++ b/vendor/github.com/docker/docker/pkg/fileutils/fileutils_unix.go @@ -1,13 +1,13 @@ // +build linux freebsd -package fileutils +package fileutils // import "github.com/docker/docker/pkg/fileutils" import ( "fmt" "io/ioutil" "os" - "github.com/Sirupsen/logrus" + "github.com/sirupsen/logrus" ) // GetTotalUsedFds Returns the number of used File Descriptors by diff --git a/vendor/github.com/docker/docker/pkg/fileutils/fileutils_windows.go b/vendor/github.com/docker/docker/pkg/fileutils/fileutils_windows.go index 5ec21cace..3f1ebb656 100644 --- a/vendor/github.com/docker/docker/pkg/fileutils/fileutils_windows.go +++ b/vendor/github.com/docker/docker/pkg/fileutils/fileutils_windows.go @@ -1,4 +1,4 @@ -package fileutils +package fileutils // import "github.com/docker/docker/pkg/fileutils" // GetTotalUsedFds Returns the number of used File Descriptors. Not supported // on Windows. diff --git a/vendor/github.com/docker/docker/pkg/homedir/homedir_linux.go b/vendor/github.com/docker/docker/pkg/homedir/homedir_linux.go new file mode 100644 index 000000000..ee15ed52b --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/homedir/homedir_linux.go @@ -0,0 +1,21 @@ +package homedir // import "github.com/docker/docker/pkg/homedir" + +import ( + "os" + + "github.com/docker/docker/pkg/idtools" +) + +// GetStatic returns the home directory for the current user without calling +// os/user.Current(). This is useful for static-linked binary on glibc-based +// system, because a call to os/user.Current() in a static binary leads to +// segfault due to a glibc issue that won't be fixed in a short term. +// (#29344, golang/go#13470, https://sourceware.org/bugzilla/show_bug.cgi?id=19341) +func GetStatic() (string, error) { + uid := os.Getuid() + usr, err := idtools.LookupUID(uid) + if err != nil { + return "", err + } + return usr.Home, nil +} diff --git a/vendor/github.com/docker/docker/pkg/homedir/homedir_others.go b/vendor/github.com/docker/docker/pkg/homedir/homedir_others.go new file mode 100644 index 000000000..75ada2fe5 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/homedir/homedir_others.go @@ -0,0 +1,13 @@ +// +build !linux + +package homedir // import "github.com/docker/docker/pkg/homedir" + +import ( + "errors" +) + +// GetStatic is not needed for non-linux systems. +// (Precisely, it is needed only for glibc-based linux systems.) +func GetStatic() (string, error) { + return "", errors.New("homedir.GetStatic() is not supported on this system") +} diff --git a/vendor/github.com/docker/docker/pkg/homedir/homedir.go b/vendor/github.com/docker/docker/pkg/homedir/homedir_unix.go similarity index 75% rename from vendor/github.com/docker/docker/pkg/homedir/homedir.go rename to vendor/github.com/docker/docker/pkg/homedir/homedir_unix.go index 8154e83f0..d85e12448 100644 --- a/vendor/github.com/docker/docker/pkg/homedir/homedir.go +++ b/vendor/github.com/docker/docker/pkg/homedir/homedir_unix.go @@ -1,8 +1,9 @@ -package homedir +// +build !windows + +package homedir // import "github.com/docker/docker/pkg/homedir" import ( "os" - "runtime" "github.com/opencontainers/runc/libcontainer/user" ) @@ -10,9 +11,6 @@ import ( // Key returns the env var name for the user's home dir based on // the platform being run on func Key() string { - if runtime.GOOS == "windows" { - return "USERPROFILE" - } return "HOME" } @@ -21,7 +19,7 @@ func Key() string { // Returned path should be used with "path/filepath" to form new paths. func Get() string { home := os.Getenv(Key()) - if home == "" && runtime.GOOS != "windows" { + if home == "" { if u, err := user.CurrentUser(); err == nil { return u.Home } @@ -32,8 +30,5 @@ func Get() string { // GetShortcutString returns the string that is shortcut to user's home directory // in the native shell of the platform running on. func GetShortcutString() string { - if runtime.GOOS == "windows" { - return "%USERPROFILE%" // be careful while using in format functions - } return "~" } diff --git a/vendor/github.com/docker/docker/pkg/homedir/homedir_windows.go b/vendor/github.com/docker/docker/pkg/homedir/homedir_windows.go new file mode 100644 index 000000000..2f81813b2 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/homedir/homedir_windows.go @@ -0,0 +1,24 @@ +package homedir // import "github.com/docker/docker/pkg/homedir" + +import ( + "os" +) + +// Key returns the env var name for the user's home dir based on +// the platform being run on +func Key() string { + return "USERPROFILE" +} + +// Get returns the home directory of the current user with the help of +// environment variables depending on the target operating system. +// Returned path should be used with "path/filepath" to form new paths. +func Get() string { + return os.Getenv(Key()) +} + +// GetShortcutString returns the string that is shortcut to user's home directory +// in the native shell of the platform running on. +func GetShortcutString() string { + return "%USERPROFILE%" // be careful while using in format functions +} diff --git a/vendor/github.com/docker/docker/pkg/idtools/idtools.go b/vendor/github.com/docker/docker/pkg/idtools/idtools.go index 6bca46628..d1f173a31 100644 --- a/vendor/github.com/docker/docker/pkg/idtools/idtools.go +++ b/vendor/github.com/docker/docker/pkg/idtools/idtools.go @@ -1,4 +1,4 @@ -package idtools +package idtools // import "github.com/docker/docker/pkg/idtools" import ( "bufio" @@ -30,56 +30,50 @@ func (e ranges) Swap(i, j int) { e[i], e[j] = e[j], e[i] } func (e ranges) Less(i, j int) bool { return e[i].Start < e[j].Start } const ( - subuidFileName string = "/etc/subuid" - subgidFileName string = "/etc/subgid" + subuidFileName = "/etc/subuid" + subgidFileName = "/etc/subgid" ) -// MkdirAllAs creates a directory (include any along the path) and then modifies +// MkdirAllAndChown creates a directory (include any along the path) and then modifies // ownership to the requested uid/gid. If the directory already exists, this // function will still change ownership to the requested uid/gid pair. -func MkdirAllAs(path string, mode os.FileMode, ownerUID, ownerGID int) error { - return mkdirAs(path, mode, ownerUID, ownerGID, true, true) +func MkdirAllAndChown(path string, mode os.FileMode, owner IDPair) error { + return mkdirAs(path, mode, owner.UID, owner.GID, true, true) } -// MkdirAllNewAs creates a directory (include any along the path) and then modifies +// MkdirAndChown creates a directory and then modifies ownership to the requested uid/gid. +// If the directory already exists, this function still changes ownership. +// Note that unlike os.Mkdir(), this function does not return IsExist error +// in case path already exists. +func MkdirAndChown(path string, mode os.FileMode, owner IDPair) error { + return mkdirAs(path, mode, owner.UID, owner.GID, false, true) +} + +// MkdirAllAndChownNew creates a directory (include any along the path) and then modifies // ownership ONLY of newly created directories to the requested uid/gid. If the // directories along the path exist, no change of ownership will be performed -func MkdirAllNewAs(path string, mode os.FileMode, ownerUID, ownerGID int) error { - return mkdirAs(path, mode, ownerUID, ownerGID, true, false) -} - -// MkdirAs creates a directory and then modifies ownership to the requested uid/gid. -// If the directory already exists, this function still changes ownership -func MkdirAs(path string, mode os.FileMode, ownerUID, ownerGID int) error { - return mkdirAs(path, mode, ownerUID, ownerGID, false, true) +func MkdirAllAndChownNew(path string, mode os.FileMode, owner IDPair) error { + return mkdirAs(path, mode, owner.UID, owner.GID, true, false) } // GetRootUIDGID retrieves the remapped root uid/gid pair from the set of maps. // If the maps are empty, then the root uid/gid will default to "real" 0/0 func GetRootUIDGID(uidMap, gidMap []IDMap) (int, int, error) { - var uid, gid int - - if uidMap != nil { - xUID, err := ToHost(0, uidMap) - if err != nil { - return -1, -1, err - } - uid = xUID + uid, err := toHost(0, uidMap) + if err != nil { + return -1, -1, err } - if gidMap != nil { - xGID, err := ToHost(0, gidMap) - if err != nil { - return -1, -1, err - } - gid = xGID + gid, err := toHost(0, gidMap) + if err != nil { + return -1, -1, err } return uid, gid, nil } -// ToContainer takes an id mapping, and uses it to translate a +// toContainer takes an id mapping, and uses it to translate a // host ID to the remapped ID. If no map is provided, then the translation // assumes a 1-to-1 mapping and returns the passed in id -func ToContainer(hostID int, idMap []IDMap) (int, error) { +func toContainer(hostID int, idMap []IDMap) (int, error) { if idMap == nil { return hostID, nil } @@ -92,10 +86,10 @@ func ToContainer(hostID int, idMap []IDMap) (int, error) { return -1, fmt.Errorf("Host ID %d cannot be mapped to a container ID", hostID) } -// ToHost takes an id mapping and a remapped ID, and translates the +// toHost takes an id mapping and a remapped ID, and translates the // ID to the mapped host ID. If no map is provided, then the translation // assumes a 1-to-1 mapping and returns the passed in id # -func ToHost(contID int, idMap []IDMap) (int, error) { +func toHost(contID int, idMap []IDMap) (int, error) { if idMap == nil { return contID, nil } @@ -108,26 +102,101 @@ func ToHost(contID int, idMap []IDMap) (int, error) { return -1, fmt.Errorf("Container ID %d cannot be mapped to a host ID", contID) } -// CreateIDMappings takes a requested user and group name and +// IDPair is a UID and GID pair +type IDPair struct { + UID int + GID int +} + +// IDMappings contains a mappings of UIDs and GIDs +type IDMappings struct { + uids []IDMap + gids []IDMap +} + +// NewIDMappings takes a requested user and group name and // using the data from /etc/sub{uid,gid} ranges, creates the // proper uid and gid remapping ranges for that user/group pair -func CreateIDMappings(username, groupname string) ([]IDMap, []IDMap, error) { +func NewIDMappings(username, groupname string) (*IDMappings, error) { subuidRanges, err := parseSubuid(username) if err != nil { - return nil, nil, err + return nil, err } subgidRanges, err := parseSubgid(groupname) if err != nil { - return nil, nil, err + return nil, err } if len(subuidRanges) == 0 { - return nil, nil, fmt.Errorf("No subuid ranges found for user %q", username) + return nil, fmt.Errorf("No subuid ranges found for user %q", username) } if len(subgidRanges) == 0 { - return nil, nil, fmt.Errorf("No subgid ranges found for group %q", groupname) + return nil, fmt.Errorf("No subgid ranges found for group %q", groupname) } - return createIDMap(subuidRanges), createIDMap(subgidRanges), nil + return &IDMappings{ + uids: createIDMap(subuidRanges), + gids: createIDMap(subgidRanges), + }, nil +} + +// NewIDMappingsFromMaps creates a new mapping from two slices +// Deprecated: this is a temporary shim while transitioning to IDMapping +func NewIDMappingsFromMaps(uids []IDMap, gids []IDMap) *IDMappings { + return &IDMappings{uids: uids, gids: gids} +} + +// RootPair returns a uid and gid pair for the root user. The error is ignored +// because a root user always exists, and the defaults are correct when the uid +// and gid maps are empty. +func (i *IDMappings) RootPair() IDPair { + uid, gid, _ := GetRootUIDGID(i.uids, i.gids) + return IDPair{UID: uid, GID: gid} +} + +// ToHost returns the host UID and GID for the container uid, gid. +// Remapping is only performed if the ids aren't already the remapped root ids +func (i *IDMappings) ToHost(pair IDPair) (IDPair, error) { + var err error + target := i.RootPair() + + if pair.UID != target.UID { + target.UID, err = toHost(pair.UID, i.uids) + if err != nil { + return target, err + } + } + + if pair.GID != target.GID { + target.GID, err = toHost(pair.GID, i.gids) + } + return target, err +} + +// ToContainer returns the container UID and GID for the host uid and gid +func (i *IDMappings) ToContainer(pair IDPair) (int, int, error) { + uid, err := toContainer(pair.UID, i.uids) + if err != nil { + return -1, -1, err + } + gid, err := toContainer(pair.GID, i.gids) + return uid, gid, err +} + +// Empty returns true if there are no id mappings +func (i *IDMappings) Empty() bool { + return len(i.uids) == 0 && len(i.gids) == 0 +} + +// UIDs return the UID mapping +// TODO: remove this once everything has been refactored to use pairs +func (i *IDMappings) UIDs() []IDMap { + return i.uids +} + +// GIDs return the UID mapping +// TODO: remove this once everything has been refactored to use pairs +func (i *IDMappings) GIDs() []IDMap { + return i.gids } func createIDMap(subidRanges ranges) []IDMap { diff --git a/vendor/github.com/docker/docker/pkg/idtools/idtools_unix.go b/vendor/github.com/docker/docker/pkg/idtools/idtools_unix.go index b57d6ef12..1d87ea3bc 100644 --- a/vendor/github.com/docker/docker/pkg/idtools/idtools_unix.go +++ b/vendor/github.com/docker/docker/pkg/idtools/idtools_unix.go @@ -1,12 +1,24 @@ // +build !windows -package idtools +package idtools // import "github.com/docker/docker/pkg/idtools" import ( + "bytes" + "fmt" + "io" "os" "path/filepath" + "strings" + "sync" + "syscall" "github.com/docker/docker/pkg/system" + "github.com/opencontainers/runc/libcontainer/user" +) + +var ( + entOnce sync.Once + getentCmd string ) func mkdirAs(path string, mode os.FileMode, ownerUID, ownerGID int, mkAll, chownExisting bool) error { @@ -15,17 +27,22 @@ func mkdirAs(path string, mode os.FileMode, ownerUID, ownerGID int, mkAll, chown // so that we can chown all of them properly at the end. If chownExisting is false, we won't // chown the full directory path if it exists var paths []string - if _, err := os.Stat(path); err != nil && os.IsNotExist(err) { - paths = []string{path} - } else if err == nil && chownExisting { - if err := os.Chown(path, ownerUID, ownerGID); err != nil { - return err + + stat, err := system.Stat(path) + if err == nil { + if !stat.IsDir() { + return &os.PathError{Op: "mkdir", Path: path, Err: syscall.ENOTDIR} } + if !chownExisting { + return nil + } + // short-circuit--we were called with an existing directory and chown was requested - return nil - } else if err == nil { - // nothing to do; directory path fully exists already and chown was NOT requested - return nil + return lazyChown(path, ownerUID, ownerGID, stat) + } + + if os.IsNotExist(err) { + paths = []string{path} } if mkAll { @@ -41,7 +58,7 @@ func mkdirAs(path string, mode os.FileMode, ownerUID, ownerGID int, mkAll, chown paths = append(paths, dirPath) } } - if err := system.MkdirAll(path, mode); err != nil && !os.IsExist(err) { + if err := system.MkdirAll(path, mode, ""); err != nil { return err } } else { @@ -52,9 +69,162 @@ func mkdirAs(path string, mode os.FileMode, ownerUID, ownerGID int, mkAll, chown // even if it existed, we will chown the requested path + any subpaths that // didn't exist when we called MkdirAll for _, pathComponent := range paths { - if err := os.Chown(pathComponent, ownerUID, ownerGID); err != nil { + if err := lazyChown(pathComponent, ownerUID, ownerGID, nil); err != nil { return err } } return nil } + +// CanAccess takes a valid (existing) directory and a uid, gid pair and determines +// if that uid, gid pair has access (execute bit) to the directory +func CanAccess(path string, pair IDPair) bool { + statInfo, err := system.Stat(path) + if err != nil { + return false + } + fileMode := os.FileMode(statInfo.Mode()) + permBits := fileMode.Perm() + return accessible(statInfo.UID() == uint32(pair.UID), + statInfo.GID() == uint32(pair.GID), permBits) +} + +func accessible(isOwner, isGroup bool, perms os.FileMode) bool { + if isOwner && (perms&0100 == 0100) { + return true + } + if isGroup && (perms&0010 == 0010) { + return true + } + if perms&0001 == 0001 { + return true + } + return false +} + +// LookupUser uses traditional local system files lookup (from libcontainer/user) on a username, +// followed by a call to `getent` for supporting host configured non-files passwd and group dbs +func LookupUser(username string) (user.User, error) { + // first try a local system files lookup using existing capabilities + usr, err := user.LookupUser(username) + if err == nil { + return usr, nil + } + // local files lookup failed; attempt to call `getent` to query configured passwd dbs + usr, err = getentUser(fmt.Sprintf("%s %s", "passwd", username)) + if err != nil { + return user.User{}, err + } + return usr, nil +} + +// LookupUID uses traditional local system files lookup (from libcontainer/user) on a uid, +// followed by a call to `getent` for supporting host configured non-files passwd and group dbs +func LookupUID(uid int) (user.User, error) { + // first try a local system files lookup using existing capabilities + usr, err := user.LookupUid(uid) + if err == nil { + return usr, nil + } + // local files lookup failed; attempt to call `getent` to query configured passwd dbs + return getentUser(fmt.Sprintf("%s %d", "passwd", uid)) +} + +func getentUser(args string) (user.User, error) { + reader, err := callGetent(args) + if err != nil { + return user.User{}, err + } + users, err := user.ParsePasswd(reader) + if err != nil { + return user.User{}, err + } + if len(users) == 0 { + return user.User{}, fmt.Errorf("getent failed to find passwd entry for %q", strings.Split(args, " ")[1]) + } + return users[0], nil +} + +// LookupGroup uses traditional local system files lookup (from libcontainer/user) on a group name, +// followed by a call to `getent` for supporting host configured non-files passwd and group dbs +func LookupGroup(groupname string) (user.Group, error) { + // first try a local system files lookup using existing capabilities + group, err := user.LookupGroup(groupname) + if err == nil { + return group, nil + } + // local files lookup failed; attempt to call `getent` to query configured group dbs + return getentGroup(fmt.Sprintf("%s %s", "group", groupname)) +} + +// LookupGID uses traditional local system files lookup (from libcontainer/user) on a group ID, +// followed by a call to `getent` for supporting host configured non-files passwd and group dbs +func LookupGID(gid int) (user.Group, error) { + // first try a local system files lookup using existing capabilities + group, err := user.LookupGid(gid) + if err == nil { + return group, nil + } + // local files lookup failed; attempt to call `getent` to query configured group dbs + return getentGroup(fmt.Sprintf("%s %d", "group", gid)) +} + +func getentGroup(args string) (user.Group, error) { + reader, err := callGetent(args) + if err != nil { + return user.Group{}, err + } + groups, err := user.ParseGroup(reader) + if err != nil { + return user.Group{}, err + } + if len(groups) == 0 { + return user.Group{}, fmt.Errorf("getent failed to find groups entry for %q", strings.Split(args, " ")[1]) + } + return groups[0], nil +} + +func callGetent(args string) (io.Reader, error) { + entOnce.Do(func() { getentCmd, _ = resolveBinary("getent") }) + // if no `getent` command on host, can't do anything else + if getentCmd == "" { + return nil, fmt.Errorf("") + } + out, err := execCmd(getentCmd, args) + if err != nil { + exitCode, errC := system.GetExitCode(err) + if errC != nil { + return nil, err + } + switch exitCode { + case 1: + return nil, fmt.Errorf("getent reported invalid parameters/database unknown") + case 2: + terms := strings.Split(args, " ") + return nil, fmt.Errorf("getent unable to find entry %q in %s database", terms[1], terms[0]) + case 3: + return nil, fmt.Errorf("getent database doesn't support enumeration") + default: + return nil, err + } + + } + return bytes.NewReader(out), nil +} + +// lazyChown performs a chown only if the uid/gid don't match what's requested +// Normally a Chown is a no-op if uid/gid match, but in some cases this can still cause an error, e.g. if the +// dir is on an NFS share, so don't call chown unless we absolutely must. +func lazyChown(p string, uid, gid int, stat *system.StatT) error { + if stat == nil { + var err error + stat, err = system.Stat(p) + if err != nil { + return err + } + } + if stat.UID() == uint32(uid) && stat.GID() == uint32(gid) { + return nil + } + return os.Chown(p, uid, gid) +} diff --git a/vendor/github.com/docker/docker/pkg/idtools/idtools_windows.go b/vendor/github.com/docker/docker/pkg/idtools/idtools_windows.go index c9e3c937c..d72cc2892 100644 --- a/vendor/github.com/docker/docker/pkg/idtools/idtools_windows.go +++ b/vendor/github.com/docker/docker/pkg/idtools/idtools_windows.go @@ -1,6 +1,4 @@ -// +build windows - -package idtools +package idtools // import "github.com/docker/docker/pkg/idtools" import ( "os" @@ -11,8 +9,15 @@ import ( // Platforms such as Windows do not support the UID/GID concept. So make this // just a wrapper around system.MkdirAll. func mkdirAs(path string, mode os.FileMode, ownerUID, ownerGID int, mkAll, chownExisting bool) error { - if err := system.MkdirAll(path, mode); err != nil && !os.IsExist(err) { + if err := system.MkdirAll(path, mode, ""); err != nil { return err } return nil } + +// CanAccess takes a valid (existing) directory and a uid, gid pair and determines +// if that uid, gid pair has access (execute bit) to the directory +// Windows does not require/support this function, so always return true +func CanAccess(path string, pair IDPair) bool { + return true +} diff --git a/vendor/github.com/docker/docker/pkg/idtools/usergroupadd_linux.go b/vendor/github.com/docker/docker/pkg/idtools/usergroupadd_linux.go index 4a4aaed04..6272c5a40 100644 --- a/vendor/github.com/docker/docker/pkg/idtools/usergroupadd_linux.go +++ b/vendor/github.com/docker/docker/pkg/idtools/usergroupadd_linux.go @@ -1,9 +1,7 @@ -package idtools +package idtools // import "github.com/docker/docker/pkg/idtools" import ( "fmt" - "os/exec" - "path/filepath" "regexp" "sort" "strconv" @@ -33,23 +31,6 @@ var ( userMod = "usermod" ) -func resolveBinary(binname string) (string, error) { - binaryPath, err := exec.LookPath(binname) - if err != nil { - return "", err - } - resolvedPath, err := filepath.EvalSymlinks(binaryPath) - if err != nil { - return "", err - } - //only return no error if the final resolved binary basename - //matches what was searched for - if filepath.Base(resolvedPath) == binname { - return resolvedPath, nil - } - return "", fmt.Errorf("Binary %q does not resolve to a binary of that name in $PATH (%q)", binname, resolvedPath) -} - // AddNamespaceRangesUser takes a username and uses the standard system // utility to create a system user/group pair used to hold the // /etc/sub{uid,gid} ranges which will be used for user namespace @@ -181,8 +162,3 @@ func wouldOverlap(arange subIDRange, ID int) bool { } return false } - -func execCmd(cmd, args string) ([]byte, error) { - execCmd := exec.Command(cmd, strings.Split(args, " ")...) - return execCmd.CombinedOutput() -} diff --git a/vendor/github.com/docker/docker/pkg/idtools/usergroupadd_unsupported.go b/vendor/github.com/docker/docker/pkg/idtools/usergroupadd_unsupported.go index d98b354cb..e7c4d6311 100644 --- a/vendor/github.com/docker/docker/pkg/idtools/usergroupadd_unsupported.go +++ b/vendor/github.com/docker/docker/pkg/idtools/usergroupadd_unsupported.go @@ -1,6 +1,6 @@ // +build !linux -package idtools +package idtools // import "github.com/docker/docker/pkg/idtools" import "fmt" diff --git a/vendor/github.com/docker/docker/pkg/idtools/utils_unix.go b/vendor/github.com/docker/docker/pkg/idtools/utils_unix.go new file mode 100644 index 000000000..903ac4501 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/idtools/utils_unix.go @@ -0,0 +1,32 @@ +// +build !windows + +package idtools // import "github.com/docker/docker/pkg/idtools" + +import ( + "fmt" + "os/exec" + "path/filepath" + "strings" +) + +func resolveBinary(binname string) (string, error) { + binaryPath, err := exec.LookPath(binname) + if err != nil { + return "", err + } + resolvedPath, err := filepath.EvalSymlinks(binaryPath) + if err != nil { + return "", err + } + //only return no error if the final resolved binary basename + //matches what was searched for + if filepath.Base(resolvedPath) == binname { + return resolvedPath, nil + } + return "", fmt.Errorf("Binary %q does not resolve to a binary of that name in $PATH (%q)", binname, resolvedPath) +} + +func execCmd(cmd, args string) ([]byte, error) { + execCmd := exec.Command(cmd, strings.Split(args, " ")...) + return execCmd.CombinedOutput() +} diff --git a/vendor/github.com/docker/docker/pkg/ioutils/buffer.go b/vendor/github.com/docker/docker/pkg/ioutils/buffer.go index 3d737b3e1..466f79294 100644 --- a/vendor/github.com/docker/docker/pkg/ioutils/buffer.go +++ b/vendor/github.com/docker/docker/pkg/ioutils/buffer.go @@ -1,4 +1,4 @@ -package ioutils +package ioutils // import "github.com/docker/docker/pkg/ioutils" import ( "errors" diff --git a/vendor/github.com/docker/docker/pkg/ioutils/bytespipe.go b/vendor/github.com/docker/docker/pkg/ioutils/bytespipe.go index eca129be3..d4bbf3c9d 100644 --- a/vendor/github.com/docker/docker/pkg/ioutils/bytespipe.go +++ b/vendor/github.com/docker/docker/pkg/ioutils/bytespipe.go @@ -1,4 +1,4 @@ -package ioutils +package ioutils // import "github.com/docker/docker/pkg/ioutils" import ( "errors" @@ -133,8 +133,9 @@ func (bp *BytesPipe) Read(p []byte) (n int, err error) { } bp.wait.Wait() if bp.bufLen == 0 && bp.closeErr != nil { + err := bp.closeErr bp.mu.Unlock() - return 0, bp.closeErr + return 0, err } } diff --git a/vendor/github.com/docker/docker/pkg/ioutils/fmt.go b/vendor/github.com/docker/docker/pkg/ioutils/fmt.go deleted file mode 100644 index 0b04b0ba3..000000000 --- a/vendor/github.com/docker/docker/pkg/ioutils/fmt.go +++ /dev/null @@ -1,22 +0,0 @@ -package ioutils - -import ( - "fmt" - "io" -) - -// FprintfIfNotEmpty prints the string value if it's not empty -func FprintfIfNotEmpty(w io.Writer, format, value string) (int, error) { - if value != "" { - return fmt.Fprintf(w, format, value) - } - return 0, nil -} - -// FprintfIfTrue prints the boolean value if it's true -func FprintfIfTrue(w io.Writer, format string, ok bool) (int, error) { - if ok { - return fmt.Fprintf(w, format, ok) - } - return 0, nil -} diff --git a/vendor/github.com/docker/docker/pkg/ioutils/fswriters.go b/vendor/github.com/docker/docker/pkg/ioutils/fswriters.go index ca9767072..534d66ac2 100644 --- a/vendor/github.com/docker/docker/pkg/ioutils/fswriters.go +++ b/vendor/github.com/docker/docker/pkg/ioutils/fswriters.go @@ -1,4 +1,4 @@ -package ioutils +package ioutils // import "github.com/docker/docker/pkg/ioutils" import ( "io" @@ -15,13 +15,15 @@ func NewAtomicFileWriter(filename string, perm os.FileMode) (io.WriteCloser, err if err != nil { return nil, err } + abspath, err := filepath.Abs(filename) if err != nil { return nil, err } return &atomicFileWriter{ - f: f, - fn: abspath, + f: f, + fn: abspath, + perm: perm, }, nil } @@ -32,6 +34,84 @@ func AtomicWriteFile(filename string, data []byte, perm os.FileMode) error { return err } n, err := f.Write(data) + if err == nil && n < len(data) { + err = io.ErrShortWrite + f.(*atomicFileWriter).writeErr = err + } + if err1 := f.Close(); err == nil { + err = err1 + } + return err +} + +type atomicFileWriter struct { + f *os.File + fn string + writeErr error + perm os.FileMode +} + +func (w *atomicFileWriter) Write(dt []byte) (int, error) { + n, err := w.f.Write(dt) + if err != nil { + w.writeErr = err + } + return n, err +} + +func (w *atomicFileWriter) Close() (retErr error) { + defer func() { + if retErr != nil || w.writeErr != nil { + os.Remove(w.f.Name()) + } + }() + if err := w.f.Sync(); err != nil { + w.f.Close() + return err + } + if err := w.f.Close(); err != nil { + return err + } + if err := os.Chmod(w.f.Name(), w.perm); err != nil { + return err + } + if w.writeErr == nil { + return os.Rename(w.f.Name(), w.fn) + } + return nil +} + +// AtomicWriteSet is used to atomically write a set +// of files and ensure they are visible at the same time. +// Must be committed to a new directory. +type AtomicWriteSet struct { + root string +} + +// NewAtomicWriteSet creates a new atomic write set to +// atomically create a set of files. The given directory +// is used as the base directory for storing files before +// commit. If no temporary directory is given the system +// default is used. +func NewAtomicWriteSet(tmpDir string) (*AtomicWriteSet, error) { + td, err := ioutil.TempDir(tmpDir, "write-set-") + if err != nil { + return nil, err + } + + return &AtomicWriteSet{ + root: td, + }, nil +} + +// WriteFile writes a file to the set, guaranteeing the file +// has been synced. +func (ws *AtomicWriteSet) WriteFile(filename string, data []byte, perm os.FileMode) error { + f, err := ws.FileWriter(filename, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, perm) + if err != nil { + return err + } + n, err := f.Write(data) if err == nil && n < len(data) { err = io.ErrShortWrite } @@ -41,35 +121,42 @@ func AtomicWriteFile(filename string, data []byte, perm os.FileMode) error { return err } -type atomicFileWriter struct { - f *os.File - fn string - writeErr error +type syncFileCloser struct { + *os.File } -func (w *atomicFileWriter) Write(dt []byte) (int, error) { - n, err := w.f.Write(dt) +func (w syncFileCloser) Close() error { + err := w.File.Sync() + if err1 := w.File.Close(); err == nil { + err = err1 + } + return err +} + +// FileWriter opens a file writer inside the set. The file +// should be synced and closed before calling commit. +func (ws *AtomicWriteSet) FileWriter(name string, flag int, perm os.FileMode) (io.WriteCloser, error) { + f, err := os.OpenFile(filepath.Join(ws.root, name), flag, perm) if err != nil { - w.writeErr = err + return nil, err } - return n, err + return syncFileCloser{f}, nil } -func (w *atomicFileWriter) Close() (retErr error) { - defer func() { - if retErr != nil { - os.Remove(w.f.Name()) - } - }() - if err := w.f.Sync(); err != nil { - w.f.Close() - return err - } - if err := w.f.Close(); err != nil { - return err - } - if w.writeErr == nil { - return os.Rename(w.f.Name(), w.fn) - } - return nil +// Cancel cancels the set and removes all temporary data +// created in the set. +func (ws *AtomicWriteSet) Cancel() error { + return os.RemoveAll(ws.root) +} + +// Commit moves all created files to the target directory. The +// target directory must not exist and the parent of the target +// directory must exist. +func (ws *AtomicWriteSet) Commit(target string) error { + return os.Rename(ws.root, target) +} + +// String returns the location the set is writing to. +func (ws *AtomicWriteSet) String() string { + return ws.root } diff --git a/vendor/github.com/docker/docker/pkg/ioutils/multireader.go b/vendor/github.com/docker/docker/pkg/ioutils/multireader.go deleted file mode 100644 index 0d2d76b47..000000000 --- a/vendor/github.com/docker/docker/pkg/ioutils/multireader.go +++ /dev/null @@ -1,226 +0,0 @@ -package ioutils - -import ( - "bytes" - "fmt" - "io" - "os" -) - -type pos struct { - idx int - offset int64 -} - -type multiReadSeeker struct { - readers []io.ReadSeeker - pos *pos - posIdx map[io.ReadSeeker]int -} - -func (r *multiReadSeeker) Seek(offset int64, whence int) (int64, error) { - var tmpOffset int64 - switch whence { - case os.SEEK_SET: - for i, rdr := range r.readers { - // get size of the current reader - s, err := rdr.Seek(0, os.SEEK_END) - if err != nil { - return -1, err - } - - if offset > tmpOffset+s { - if i == len(r.readers)-1 { - rdrOffset := s + (offset - tmpOffset) - if _, err := rdr.Seek(rdrOffset, os.SEEK_SET); err != nil { - return -1, err - } - r.pos = &pos{i, rdrOffset} - return offset, nil - } - - tmpOffset += s - continue - } - - rdrOffset := offset - tmpOffset - idx := i - - rdr.Seek(rdrOffset, os.SEEK_SET) - // make sure all following readers are at 0 - for _, rdr := range r.readers[i+1:] { - rdr.Seek(0, os.SEEK_SET) - } - - if rdrOffset == s && i != len(r.readers)-1 { - idx++ - rdrOffset = 0 - } - r.pos = &pos{idx, rdrOffset} - return offset, nil - } - case os.SEEK_END: - for _, rdr := range r.readers { - s, err := rdr.Seek(0, os.SEEK_END) - if err != nil { - return -1, err - } - tmpOffset += s - } - r.Seek(tmpOffset+offset, os.SEEK_SET) - return tmpOffset + offset, nil - case os.SEEK_CUR: - if r.pos == nil { - return r.Seek(offset, os.SEEK_SET) - } - // Just return the current offset - if offset == 0 { - return r.getCurOffset() - } - - curOffset, err := r.getCurOffset() - if err != nil { - return -1, err - } - rdr, rdrOffset, err := r.getReaderForOffset(curOffset + offset) - if err != nil { - return -1, err - } - - r.pos = &pos{r.posIdx[rdr], rdrOffset} - return curOffset + offset, nil - default: - return -1, fmt.Errorf("Invalid whence: %d", whence) - } - - return -1, fmt.Errorf("Error seeking for whence: %d, offset: %d", whence, offset) -} - -func (r *multiReadSeeker) getReaderForOffset(offset int64) (io.ReadSeeker, int64, error) { - var rdr io.ReadSeeker - var rdrOffset int64 - - for i, rdr := range r.readers { - offsetTo, err := r.getOffsetToReader(rdr) - if err != nil { - return nil, -1, err - } - if offsetTo > offset { - rdr = r.readers[i-1] - rdrOffset = offsetTo - offset - break - } - - if rdr == r.readers[len(r.readers)-1] { - rdrOffset = offsetTo + offset - break - } - } - - return rdr, rdrOffset, nil -} - -func (r *multiReadSeeker) getCurOffset() (int64, error) { - var totalSize int64 - for _, rdr := range r.readers[:r.pos.idx+1] { - if r.posIdx[rdr] == r.pos.idx { - totalSize += r.pos.offset - break - } - - size, err := getReadSeekerSize(rdr) - if err != nil { - return -1, fmt.Errorf("error getting seeker size: %v", err) - } - totalSize += size - } - return totalSize, nil -} - -func (r *multiReadSeeker) getOffsetToReader(rdr io.ReadSeeker) (int64, error) { - var offset int64 - for _, r := range r.readers { - if r == rdr { - break - } - - size, err := getReadSeekerSize(rdr) - if err != nil { - return -1, err - } - offset += size - } - return offset, nil -} - -func (r *multiReadSeeker) Read(b []byte) (int, error) { - if r.pos == nil { - r.pos = &pos{0, 0} - } - - bCap := int64(cap(b)) - buf := bytes.NewBuffer(nil) - var rdr io.ReadSeeker - - for _, rdr = range r.readers[r.pos.idx:] { - readBytes, err := io.CopyN(buf, rdr, bCap) - if err != nil && err != io.EOF { - return -1, err - } - bCap -= readBytes - - if bCap == 0 { - break - } - } - - rdrPos, err := rdr.Seek(0, os.SEEK_CUR) - if err != nil { - return -1, err - } - r.pos = &pos{r.posIdx[rdr], rdrPos} - return buf.Read(b) -} - -func getReadSeekerSize(rdr io.ReadSeeker) (int64, error) { - // save the current position - pos, err := rdr.Seek(0, os.SEEK_CUR) - if err != nil { - return -1, err - } - - // get the size - size, err := rdr.Seek(0, os.SEEK_END) - if err != nil { - return -1, err - } - - // reset the position - if _, err := rdr.Seek(pos, os.SEEK_SET); err != nil { - return -1, err - } - return size, nil -} - -// MultiReadSeeker returns a ReadSeeker that's the logical concatenation of the provided -// input readseekers. After calling this method the initial position is set to the -// beginning of the first ReadSeeker. At the end of a ReadSeeker, Read always advances -// to the beginning of the next ReadSeeker and returns EOF at the end of the last ReadSeeker. -// Seek can be used over the sum of lengths of all readseekers. -// -// When a MultiReadSeeker is used, no Read and Seek operations should be made on -// its ReadSeeker components. Also, users should make no assumption on the state -// of individual readseekers while the MultiReadSeeker is used. -func MultiReadSeeker(readers ...io.ReadSeeker) io.ReadSeeker { - if len(readers) == 1 { - return readers[0] - } - idx := make(map[io.ReadSeeker]int) - for i, rdr := range readers { - idx[rdr] = i - } - return &multiReadSeeker{ - readers: readers, - posIdx: idx, - } -} diff --git a/vendor/github.com/docker/docker/pkg/ioutils/readers.go b/vendor/github.com/docker/docker/pkg/ioutils/readers.go index e73b02bbf..1f657bd3d 100644 --- a/vendor/github.com/docker/docker/pkg/ioutils/readers.go +++ b/vendor/github.com/docker/docker/pkg/ioutils/readers.go @@ -1,25 +1,28 @@ -package ioutils +package ioutils // import "github.com/docker/docker/pkg/ioutils" import ( + "context" "crypto/sha256" "encoding/hex" "io" - - "golang.org/x/net/context" ) -type readCloserWrapper struct { +// ReadCloserWrapper wraps an io.Reader, and implements an io.ReadCloser +// It calls the given callback function when closed. It should be constructed +// with NewReadCloserWrapper +type ReadCloserWrapper struct { io.Reader closer func() error } -func (r *readCloserWrapper) Close() error { +// Close calls back the passed closer function +func (r *ReadCloserWrapper) Close() error { return r.closer() } // NewReadCloserWrapper returns a new io.ReadCloser. func NewReadCloserWrapper(r io.Reader, closer func() error) io.ReadCloser { - return &readCloserWrapper{ + return &ReadCloserWrapper{ Reader: r, closer: closer, } @@ -55,7 +58,7 @@ func HashData(src io.Reader) (string, error) { return "sha256:" + hex.EncodeToString(h.Sum(nil)), nil } -// OnEOFReader wraps a io.ReadCloser and a function +// OnEOFReader wraps an io.ReadCloser and a function // the function will run at the end of file or close the file. type OnEOFReader struct { Rc io.ReadCloser diff --git a/vendor/github.com/docker/docker/pkg/ioutils/temp_unix.go b/vendor/github.com/docker/docker/pkg/ioutils/temp_unix.go index 1539ad21b..dc894f913 100644 --- a/vendor/github.com/docker/docker/pkg/ioutils/temp_unix.go +++ b/vendor/github.com/docker/docker/pkg/ioutils/temp_unix.go @@ -1,6 +1,6 @@ // +build !windows -package ioutils +package ioutils // import "github.com/docker/docker/pkg/ioutils" import "io/ioutil" diff --git a/vendor/github.com/docker/docker/pkg/ioutils/temp_windows.go b/vendor/github.com/docker/docker/pkg/ioutils/temp_windows.go index c258e5fdd..ecaba2e36 100644 --- a/vendor/github.com/docker/docker/pkg/ioutils/temp_windows.go +++ b/vendor/github.com/docker/docker/pkg/ioutils/temp_windows.go @@ -1,6 +1,4 @@ -// +build windows - -package ioutils +package ioutils // import "github.com/docker/docker/pkg/ioutils" import ( "io/ioutil" diff --git a/vendor/github.com/docker/docker/pkg/ioutils/writeflusher.go b/vendor/github.com/docker/docker/pkg/ioutils/writeflusher.go index 52a4901ad..91b8d1826 100644 --- a/vendor/github.com/docker/docker/pkg/ioutils/writeflusher.go +++ b/vendor/github.com/docker/docker/pkg/ioutils/writeflusher.go @@ -1,4 +1,4 @@ -package ioutils +package ioutils // import "github.com/docker/docker/pkg/ioutils" import ( "io" diff --git a/vendor/github.com/docker/docker/pkg/ioutils/writers.go b/vendor/github.com/docker/docker/pkg/ioutils/writers.go index ccc7f9c23..61c679497 100644 --- a/vendor/github.com/docker/docker/pkg/ioutils/writers.go +++ b/vendor/github.com/docker/docker/pkg/ioutils/writers.go @@ -1,4 +1,4 @@ -package ioutils +package ioutils // import "github.com/docker/docker/pkg/ioutils" import "io" diff --git a/vendor/github.com/docker/docker/pkg/jsonlog/jsonlog.go b/vendor/github.com/docker/docker/pkg/jsonlog/jsonlog.go deleted file mode 100644 index 4734c3111..000000000 --- a/vendor/github.com/docker/docker/pkg/jsonlog/jsonlog.go +++ /dev/null @@ -1,42 +0,0 @@ -package jsonlog - -import ( - "encoding/json" - "fmt" - "time" -) - -// JSONLog represents a log message, typically a single entry from a given log stream. -// JSONLogs can be easily serialized to and from JSON and support custom formatting. -type JSONLog struct { - // Log is the log message - Log string `json:"log,omitempty"` - // Stream is the log source - Stream string `json:"stream,omitempty"` - // Created is the created timestamp of log - Created time.Time `json:"time"` - // Attrs is the list of extra attributes provided by the user - Attrs map[string]string `json:"attrs,omitempty"` -} - -// Format returns the log formatted according to format -// If format is nil, returns the log message -// If format is json, returns the log marshaled in json format -// By default, returns the log with the log time formatted according to format. -func (jl *JSONLog) Format(format string) (string, error) { - if format == "" { - return jl.Log, nil - } - if format == "json" { - m, err := json.Marshal(jl) - return string(m), err - } - return fmt.Sprintf("%s %s", jl.Created.Format(format), jl.Log), nil -} - -// Reset resets the log to nil. -func (jl *JSONLog) Reset() { - jl.Log = "" - jl.Stream = "" - jl.Created = time.Time{} -} diff --git a/vendor/github.com/docker/docker/pkg/jsonlog/jsonlog_marshalling.go b/vendor/github.com/docker/docker/pkg/jsonlog/jsonlog_marshalling.go deleted file mode 100644 index 83ce684a8..000000000 --- a/vendor/github.com/docker/docker/pkg/jsonlog/jsonlog_marshalling.go +++ /dev/null @@ -1,178 +0,0 @@ -// This code was initially generated by ffjson -// This code was generated via the following steps: -// $ go get -u github.com/pquerna/ffjson -// $ make BIND_DIR=. shell -// $ ffjson pkg/jsonlog/jsonlog.go -// $ mv pkg/jsonglog/jsonlog_ffjson.go pkg/jsonlog/jsonlog_marshalling.go -// -// It has been modified to improve the performance of time marshalling to JSON -// and to clean it up. -// Should this code need to be regenerated when the JSONLog struct is changed, -// the relevant changes which have been made are: -// import ( -// "bytes" -//- -// "unicode/utf8" -// ) -// -// func (mj *JSONLog) MarshalJSON() ([]byte, error) { -//@@ -20,13 +16,13 @@ func (mj *JSONLog) MarshalJSON() ([]byte, error) { -// } -// return buf.Bytes(), nil -// } -//+ -// func (mj *JSONLog) MarshalJSONBuf(buf *bytes.Buffer) error { -//- var err error -//- var obj []byte -//- var first bool = true -//- _ = obj -//- _ = err -//- _ = first -//+ var ( -//+ err error -//+ timestamp string -//+ first bool = true -//+ ) -// buf.WriteString(`{`) -// if len(mj.Log) != 0 { -// if first == true { -//@@ -52,11 +48,11 @@ func (mj *JSONLog) MarshalJSONBuf(buf *bytes.Buffer) error { -// buf.WriteString(`,`) -// } -// buf.WriteString(`"time":`) -//- obj, err = mj.Created.MarshalJSON() -//+ timestamp, err = FastTimeMarshalJSON(mj.Created) -// if err != nil { -// return err -// } -//- buf.Write(obj) -//+ buf.WriteString(timestamp) -// buf.WriteString(`}`) -// return nil -// } -// @@ -81,9 +81,10 @@ func (mj *JSONLog) MarshalJSONBuf(buf *bytes.Buffer) error { -// if len(mj.Log) != 0 { -// - if first == true { -// - first = false -// - } else { -// - buf.WriteString(`,`) -// - } -// + first = false -// buf.WriteString(`"log":`) -// ffjsonWriteJSONString(buf, mj.Log) -// } - -package jsonlog - -import ( - "bytes" - "unicode/utf8" -) - -// MarshalJSON marshals the JSONLog. -func (mj *JSONLog) MarshalJSON() ([]byte, error) { - var buf bytes.Buffer - buf.Grow(1024) - if err := mj.MarshalJSONBuf(&buf); err != nil { - return nil, err - } - return buf.Bytes(), nil -} - -// MarshalJSONBuf marshals the JSONLog and stores the result to a bytes.Buffer. -func (mj *JSONLog) MarshalJSONBuf(buf *bytes.Buffer) error { - var ( - err error - timestamp string - first = true - ) - buf.WriteString(`{`) - if len(mj.Log) != 0 { - first = false - buf.WriteString(`"log":`) - ffjsonWriteJSONString(buf, mj.Log) - } - if len(mj.Stream) != 0 { - if first { - first = false - } else { - buf.WriteString(`,`) - } - buf.WriteString(`"stream":`) - ffjsonWriteJSONString(buf, mj.Stream) - } - if !first { - buf.WriteString(`,`) - } - buf.WriteString(`"time":`) - timestamp, err = FastTimeMarshalJSON(mj.Created) - if err != nil { - return err - } - buf.WriteString(timestamp) - buf.WriteString(`}`) - return nil -} - -func ffjsonWriteJSONString(buf *bytes.Buffer, s string) { - const hex = "0123456789abcdef" - - buf.WriteByte('"') - start := 0 - for i := 0; i < len(s); { - if b := s[i]; b < utf8.RuneSelf { - if 0x20 <= b && b != '\\' && b != '"' && b != '<' && b != '>' && b != '&' { - i++ - continue - } - if start < i { - buf.WriteString(s[start:i]) - } - switch b { - case '\\', '"': - buf.WriteByte('\\') - buf.WriteByte(b) - case '\n': - buf.WriteByte('\\') - buf.WriteByte('n') - case '\r': - buf.WriteByte('\\') - buf.WriteByte('r') - default: - - buf.WriteString(`\u00`) - buf.WriteByte(hex[b>>4]) - buf.WriteByte(hex[b&0xF]) - } - i++ - start = i - continue - } - c, size := utf8.DecodeRuneInString(s[i:]) - if c == utf8.RuneError && size == 1 { - if start < i { - buf.WriteString(s[start:i]) - } - buf.WriteString(`\ufffd`) - i += size - start = i - continue - } - - if c == '\u2028' || c == '\u2029' { - if start < i { - buf.WriteString(s[start:i]) - } - buf.WriteString(`\u202`) - buf.WriteByte(hex[c&0xF]) - i += size - start = i - continue - } - i += size - } - if start < len(s) { - buf.WriteString(s[start:]) - } - buf.WriteByte('"') -} diff --git a/vendor/github.com/docker/docker/pkg/jsonlog/jsonlogbytes.go b/vendor/github.com/docker/docker/pkg/jsonlog/jsonlogbytes.go deleted file mode 100644 index df522c0d6..000000000 --- a/vendor/github.com/docker/docker/pkg/jsonlog/jsonlogbytes.go +++ /dev/null @@ -1,122 +0,0 @@ -package jsonlog - -import ( - "bytes" - "encoding/json" - "unicode/utf8" -) - -// JSONLogs is based on JSONLog. -// It allows marshalling JSONLog from Log as []byte -// and an already marshalled Created timestamp. -type JSONLogs struct { - Log []byte `json:"log,omitempty"` - Stream string `json:"stream,omitempty"` - Created string `json:"time"` - - // json-encoded bytes - RawAttrs json.RawMessage `json:"attrs,omitempty"` -} - -// MarshalJSONBuf is based on the same method from JSONLog -// It has been modified to take into account the necessary changes. -func (mj *JSONLogs) MarshalJSONBuf(buf *bytes.Buffer) error { - var first = true - - buf.WriteString(`{`) - if len(mj.Log) != 0 { - first = false - buf.WriteString(`"log":`) - ffjsonWriteJSONBytesAsString(buf, mj.Log) - } - if len(mj.Stream) != 0 { - if first == true { - first = false - } else { - buf.WriteString(`,`) - } - buf.WriteString(`"stream":`) - ffjsonWriteJSONString(buf, mj.Stream) - } - if len(mj.RawAttrs) > 0 { - if first { - first = false - } else { - buf.WriteString(`,`) - } - buf.WriteString(`"attrs":`) - buf.Write(mj.RawAttrs) - } - if !first { - buf.WriteString(`,`) - } - buf.WriteString(`"time":`) - buf.WriteString(mj.Created) - buf.WriteString(`}`) - return nil -} - -// This is based on ffjsonWriteJSONBytesAsString. It has been changed -// to accept a string passed as a slice of bytes. -func ffjsonWriteJSONBytesAsString(buf *bytes.Buffer, s []byte) { - const hex = "0123456789abcdef" - - buf.WriteByte('"') - start := 0 - for i := 0; i < len(s); { - if b := s[i]; b < utf8.RuneSelf { - if 0x20 <= b && b != '\\' && b != '"' && b != '<' && b != '>' && b != '&' { - i++ - continue - } - if start < i { - buf.Write(s[start:i]) - } - switch b { - case '\\', '"': - buf.WriteByte('\\') - buf.WriteByte(b) - case '\n': - buf.WriteByte('\\') - buf.WriteByte('n') - case '\r': - buf.WriteByte('\\') - buf.WriteByte('r') - default: - - buf.WriteString(`\u00`) - buf.WriteByte(hex[b>>4]) - buf.WriteByte(hex[b&0xF]) - } - i++ - start = i - continue - } - c, size := utf8.DecodeRune(s[i:]) - if c == utf8.RuneError && size == 1 { - if start < i { - buf.Write(s[start:i]) - } - buf.WriteString(`\ufffd`) - i += size - start = i - continue - } - - if c == '\u2028' || c == '\u2029' { - if start < i { - buf.Write(s[start:i]) - } - buf.WriteString(`\u202`) - buf.WriteByte(hex[c&0xF]) - i += size - start = i - continue - } - i += size - } - if start < len(s) { - buf.Write(s[start:]) - } - buf.WriteByte('"') -} diff --git a/vendor/github.com/docker/docker/pkg/jsonlog/time_marshalling.go b/vendor/github.com/docker/docker/pkg/jsonlog/time_marshalling.go deleted file mode 100644 index 211733814..000000000 --- a/vendor/github.com/docker/docker/pkg/jsonlog/time_marshalling.go +++ /dev/null @@ -1,27 +0,0 @@ -// Package jsonlog provides helper functions to parse and print time (time.Time) as JSON. -package jsonlog - -import ( - "errors" - "time" -) - -const ( - // RFC3339NanoFixed is our own version of RFC339Nano because we want one - // that pads the nano seconds part with zeros to ensure - // the timestamps are aligned in the logs. - RFC3339NanoFixed = "2006-01-02T15:04:05.000000000Z07:00" - // JSONFormat is the format used by FastMarshalJSON - JSONFormat = `"` + time.RFC3339Nano + `"` -) - -// FastTimeMarshalJSON avoids one of the extra allocations that -// time.MarshalJSON is making. -func FastTimeMarshalJSON(t time.Time) (string, error) { - if y := t.Year(); y < 0 || y >= 10000 { - // RFC 3339 is clear that years are 4 digits exactly. - // See golang.org/issue/4556#c15 for more discussion. - return "", errors.New("time.MarshalJSON: year outside of range [0,9999]") - } - return t.Format(JSONFormat), nil -} diff --git a/vendor/github.com/docker/docker/pkg/jsonmessage/jsonmessage.go b/vendor/github.com/docker/docker/pkg/jsonmessage/jsonmessage.go index c6ad345ce..dd95f3670 100644 --- a/vendor/github.com/docker/docker/pkg/jsonmessage/jsonmessage.go +++ b/vendor/github.com/docker/docker/pkg/jsonmessage/jsonmessage.go @@ -1,4 +1,4 @@ -package jsonmessage +package jsonmessage // import "github.com/docker/docker/pkg/jsonmessage" import ( "encoding/json" @@ -9,12 +9,14 @@ import ( "time" "github.com/Nvveen/Gotty" - - "github.com/docker/docker/pkg/jsonlog" "github.com/docker/docker/pkg/term" "github.com/docker/go-units" ) +// RFC3339NanoFixed is time.RFC3339Nano with nanoseconds padded using zeros to +// ensure the formatted time isalways the same number of characters. +const RFC3339NanoFixed = "2006-01-02T15:04:05.000000000Z07:00" + // JSONError wraps a concrete Code and Message, `Code` is // is an integer error code, `Message` is the error message. type JSONError struct { @@ -35,29 +37,33 @@ type JSONProgress struct { Current int64 `json:"current,omitempty"` Total int64 `json:"total,omitempty"` Start int64 `json:"start,omitempty"` + // If true, don't show xB/yB + HideCounts bool `json:"hidecounts,omitempty"` + Units string `json:"units,omitempty"` + nowFunc func() time.Time + winSize int } func (p *JSONProgress) String() string { var ( - width = 200 + width = p.width() pbBox string numbersBox string timeLeftBox string ) - - ws, err := term.GetWinsize(p.terminalFd) - if err == nil { - width = int(ws.Width) - } - if p.Current <= 0 && p.Total <= 0 { return "" } - current := units.HumanSize(float64(p.Current)) if p.Total <= 0 { - return fmt.Sprintf("%8v", current) + switch p.Units { + case "": + current := units.HumanSize(float64(p.Current)) + return fmt.Sprintf("%8v", current) + default: + return fmt.Sprintf("%d %s", p.Current, p.Units) + } } - total := units.HumanSize(float64(p.Total)) + percentage := int(float64(p.Current)/float64(p.Total)*100) / 2 if percentage > 50 { percentage = 50 @@ -71,15 +77,29 @@ func (p *JSONProgress) String() string { pbBox = fmt.Sprintf("[%s>%s] ", strings.Repeat("=", percentage), strings.Repeat(" ", numSpaces)) } - numbersBox = fmt.Sprintf("%8v/%v", current, total) + switch { + case p.HideCounts: + case p.Units == "": // no units, use bytes + current := units.HumanSize(float64(p.Current)) + total := units.HumanSize(float64(p.Total)) - if p.Current > p.Total { - // remove total display if the reported current is wonky. - numbersBox = fmt.Sprintf("%8v", current) + numbersBox = fmt.Sprintf("%8v/%v", current, total) + + if p.Current > p.Total { + // remove total display if the reported current is wonky. + numbersBox = fmt.Sprintf("%8v", current) + } + default: + numbersBox = fmt.Sprintf("%d/%d %s", p.Current, p.Total, p.Units) + + if p.Current > p.Total { + // remove total display if the reported current is wonky. + numbersBox = fmt.Sprintf("%d %s", p.Current, p.Units) + } } if p.Current > 0 && p.Start > 0 && percentage < 50 { - fromStart := time.Now().UTC().Sub(time.Unix(p.Start, 0)) + fromStart := p.now().Sub(time.Unix(p.Start, 0)) perEntry := fromStart / time.Duration(p.Current) left := time.Duration(p.Total-p.Current) * perEntry left = (left / time.Second) * time.Second @@ -91,6 +111,28 @@ func (p *JSONProgress) String() string { return pbBox + numbersBox + timeLeftBox } +// shim for testing +func (p *JSONProgress) now() time.Time { + if p.nowFunc == nil { + p.nowFunc = func() time.Time { + return time.Now().UTC() + } + } + return p.nowFunc() +} + +// shim for testing +func (p *JSONProgress) width() int { + if p.winSize != 0 { + return p.winSize + } + ws, err := term.GetWinsize(p.terminalFd) + if err == nil { + return int(ws.Width) + } + return 200 +} + // JSONMessage defines a message struct. It describes // the created time, where it from, status, ID of the // message. It's used for docker events. @@ -105,7 +147,7 @@ type JSONMessage struct { TimeNano int64 `json:"timeNano,omitempty"` Error *JSONError `json:"errorDetail,omitempty"` ErrorMessage string `json:"error,omitempty"` //deprecated - // Aux contains out-of-band data, such as digests for push signing. + // Aux contains out-of-band data, such as digests for push signing and image id after building. Aux *json.RawMessage `json:"aux,omitempty"` } @@ -165,7 +207,7 @@ func cursorDown(out io.Writer, ti termInfo, l int) { func (jm *JSONMessage) Display(out io.Writer, termInfo termInfo) error { if jm.Error != nil { if jm.Error.Code == 401 { - return fmt.Errorf("Authentication is required.") + return fmt.Errorf("authentication is required") } return jm.Error } @@ -178,9 +220,9 @@ func (jm *JSONMessage) Display(out io.Writer, termInfo termInfo) error { return nil } if jm.TimeNano != 0 { - fmt.Fprintf(out, "%s ", time.Unix(0, jm.TimeNano).Format(jsonlog.RFC3339NanoFixed)) + fmt.Fprintf(out, "%s ", time.Unix(0, jm.TimeNano).Format(RFC3339NanoFixed)) } else if jm.Time != 0 { - fmt.Fprintf(out, "%s ", time.Unix(jm.Time, 0).Format(jsonlog.RFC3339NanoFixed)) + fmt.Fprintf(out, "%s ", time.Unix(jm.Time, 0).Format(RFC3339NanoFixed)) } if jm.ID != "" { fmt.Fprintf(out, "%s: ", jm.ID) @@ -203,7 +245,7 @@ func (jm *JSONMessage) Display(out io.Writer, termInfo termInfo) error { // DisplayJSONMessagesStream displays a json message stream from `in` to `out`, `isTerminal` // describes if `out` is a terminal. If this is the case, it will print `\n` at the end of // each line and move the cursor while displaying. -func DisplayJSONMessagesStream(in io.Reader, out io.Writer, terminalFd uintptr, isTerminal bool, auxCallback func(*json.RawMessage)) error { +func DisplayJSONMessagesStream(in io.Reader, out io.Writer, terminalFd uintptr, isTerminal bool, auxCallback func(JSONMessage)) error { var ( dec = json.NewDecoder(in) ids = make(map[string]int) @@ -235,7 +277,7 @@ func DisplayJSONMessagesStream(in io.Reader, out io.Writer, terminalFd uintptr, if jm.Aux != nil { if auxCallback != nil { - auxCallback(jm.Aux) + auxCallback(jm) } continue } @@ -288,6 +330,6 @@ type stream interface { } // DisplayJSONMessagesToStream prints json messages to the output stream -func DisplayJSONMessagesToStream(in io.Reader, stream stream, auxCallback func(*json.RawMessage)) error { +func DisplayJSONMessagesToStream(in io.Reader, stream stream, auxCallback func(JSONMessage)) error { return DisplayJSONMessagesStream(in, stream, stream.FD(), stream.IsTerminal(), auxCallback) } diff --git a/vendor/github.com/docker/docker/pkg/longpath/longpath.go b/vendor/github.com/docker/docker/pkg/longpath/longpath.go index 9b15bfff4..4177affba 100644 --- a/vendor/github.com/docker/docker/pkg/longpath/longpath.go +++ b/vendor/github.com/docker/docker/pkg/longpath/longpath.go @@ -2,7 +2,7 @@ // in Windows, which are expected to be prepended with `\\?\` and followed by either // a drive letter, a UNC server\share, or a volume identifier. -package longpath +package longpath // import "github.com/docker/docker/pkg/longpath" import ( "strings" diff --git a/vendor/github.com/docker/docker/pkg/mount/flags.go b/vendor/github.com/docker/docker/pkg/mount/flags.go new file mode 100644 index 000000000..272363b68 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/flags.go @@ -0,0 +1,149 @@ +package mount // import "github.com/docker/docker/pkg/mount" + +import ( + "fmt" + "strings" +) + +var flags = map[string]struct { + clear bool + flag int +}{ + "defaults": {false, 0}, + "ro": {false, RDONLY}, + "rw": {true, RDONLY}, + "suid": {true, NOSUID}, + "nosuid": {false, NOSUID}, + "dev": {true, NODEV}, + "nodev": {false, NODEV}, + "exec": {true, NOEXEC}, + "noexec": {false, NOEXEC}, + "sync": {false, SYNCHRONOUS}, + "async": {true, SYNCHRONOUS}, + "dirsync": {false, DIRSYNC}, + "remount": {false, REMOUNT}, + "mand": {false, MANDLOCK}, + "nomand": {true, MANDLOCK}, + "atime": {true, NOATIME}, + "noatime": {false, NOATIME}, + "diratime": {true, NODIRATIME}, + "nodiratime": {false, NODIRATIME}, + "bind": {false, BIND}, + "rbind": {false, RBIND}, + "unbindable": {false, UNBINDABLE}, + "runbindable": {false, RUNBINDABLE}, + "private": {false, PRIVATE}, + "rprivate": {false, RPRIVATE}, + "shared": {false, SHARED}, + "rshared": {false, RSHARED}, + "slave": {false, SLAVE}, + "rslave": {false, RSLAVE}, + "relatime": {false, RELATIME}, + "norelatime": {true, RELATIME}, + "strictatime": {false, STRICTATIME}, + "nostrictatime": {true, STRICTATIME}, +} + +var validFlags = map[string]bool{ + "": true, + "size": true, + "mode": true, + "uid": true, + "gid": true, + "nr_inodes": true, + "nr_blocks": true, + "mpol": true, +} + +var propagationFlags = map[string]bool{ + "bind": true, + "rbind": true, + "unbindable": true, + "runbindable": true, + "private": true, + "rprivate": true, + "shared": true, + "rshared": true, + "slave": true, + "rslave": true, +} + +// MergeTmpfsOptions merge mount options to make sure there is no duplicate. +func MergeTmpfsOptions(options []string) ([]string, error) { + // We use collisions maps to remove duplicates. + // For flag, the key is the flag value (the key for propagation flag is -1) + // For data=value, the key is the data + flagCollisions := map[int]bool{} + dataCollisions := map[string]bool{} + + var newOptions []string + // We process in reverse order + for i := len(options) - 1; i >= 0; i-- { + option := options[i] + if option == "defaults" { + continue + } + if f, ok := flags[option]; ok && f.flag != 0 { + // There is only one propagation mode + key := f.flag + if propagationFlags[option] { + key = -1 + } + // Check to see if there is collision for flag + if !flagCollisions[key] { + // We prepend the option and add to collision map + newOptions = append([]string{option}, newOptions...) + flagCollisions[key] = true + } + continue + } + opt := strings.SplitN(option, "=", 2) + if len(opt) != 2 || !validFlags[opt[0]] { + return nil, fmt.Errorf("Invalid tmpfs option %q", opt) + } + if !dataCollisions[opt[0]] { + // We prepend the option and add to collision map + newOptions = append([]string{option}, newOptions...) + dataCollisions[opt[0]] = true + } + } + + return newOptions, nil +} + +// Parse fstab type mount options into mount() flags +// and device specific data +func parseOptions(options string) (int, string) { + var ( + flag int + data []string + ) + + for _, o := range strings.Split(options, ",") { + // If the option does not exist in the flags table or the flag + // is not supported on the platform, + // then it is a data value for a specific fs type + if f, exists := flags[o]; exists && f.flag != 0 { + if f.clear { + flag &= ^f.flag + } else { + flag |= f.flag + } + } else { + data = append(data, o) + } + } + return flag, strings.Join(data, ",") +} + +// ParseTmpfsOptions parse fstab type mount options into flags and data +func ParseTmpfsOptions(options string) (int, string, error) { + flags, data := parseOptions(options) + for _, o := range strings.Split(data, ",") { + opt := strings.SplitN(o, "=", 2) + if !validFlags[opt[0]] { + return 0, "", fmt.Errorf("Invalid tmpfs option %q", opt) + } + } + return flags, data, nil +} diff --git a/vendor/github.com/docker/docker/pkg/mount/flags_freebsd.go b/vendor/github.com/docker/docker/pkg/mount/flags_freebsd.go new file mode 100644 index 000000000..ef35ef905 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/flags_freebsd.go @@ -0,0 +1,49 @@ +// +build freebsd,cgo + +package mount // import "github.com/docker/docker/pkg/mount" + +/* +#include +*/ +import "C" + +const ( + // RDONLY will mount the filesystem as read-only. + RDONLY = C.MNT_RDONLY + + // NOSUID will not allow set-user-identifier or set-group-identifier bits to + // take effect. + NOSUID = C.MNT_NOSUID + + // NOEXEC will not allow execution of any binaries on the mounted file system. + NOEXEC = C.MNT_NOEXEC + + // SYNCHRONOUS will allow any I/O to the file system to be done synchronously. + SYNCHRONOUS = C.MNT_SYNCHRONOUS + + // NOATIME will not update the file access time when reading from a file. + NOATIME = C.MNT_NOATIME +) + +// These flags are unsupported. +const ( + BIND = 0 + DIRSYNC = 0 + MANDLOCK = 0 + NODEV = 0 + NODIRATIME = 0 + UNBINDABLE = 0 + RUNBINDABLE = 0 + PRIVATE = 0 + RPRIVATE = 0 + SHARED = 0 + RSHARED = 0 + SLAVE = 0 + RSLAVE = 0 + RBIND = 0 + RELATIVE = 0 + RELATIME = 0 + REMOUNT = 0 + STRICTATIME = 0 + mntDetach = 0 +) diff --git a/vendor/github.com/docker/docker/pkg/mount/flags_linux.go b/vendor/github.com/docker/docker/pkg/mount/flags_linux.go new file mode 100644 index 000000000..a1b199a31 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/flags_linux.go @@ -0,0 +1,87 @@ +package mount // import "github.com/docker/docker/pkg/mount" + +import ( + "golang.org/x/sys/unix" +) + +const ( + // RDONLY will mount the file system read-only. + RDONLY = unix.MS_RDONLY + + // NOSUID will not allow set-user-identifier or set-group-identifier bits to + // take effect. + NOSUID = unix.MS_NOSUID + + // NODEV will not interpret character or block special devices on the file + // system. + NODEV = unix.MS_NODEV + + // NOEXEC will not allow execution of any binaries on the mounted file system. + NOEXEC = unix.MS_NOEXEC + + // SYNCHRONOUS will allow I/O to the file system to be done synchronously. + SYNCHRONOUS = unix.MS_SYNCHRONOUS + + // DIRSYNC will force all directory updates within the file system to be done + // synchronously. This affects the following system calls: create, link, + // unlink, symlink, mkdir, rmdir, mknod and rename. + DIRSYNC = unix.MS_DIRSYNC + + // REMOUNT will attempt to remount an already-mounted file system. This is + // commonly used to change the mount flags for a file system, especially to + // make a readonly file system writeable. It does not change device or mount + // point. + REMOUNT = unix.MS_REMOUNT + + // MANDLOCK will force mandatory locks on a filesystem. + MANDLOCK = unix.MS_MANDLOCK + + // NOATIME will not update the file access time when reading from a file. + NOATIME = unix.MS_NOATIME + + // NODIRATIME will not update the directory access time. + NODIRATIME = unix.MS_NODIRATIME + + // BIND remounts a subtree somewhere else. + BIND = unix.MS_BIND + + // RBIND remounts a subtree and all possible submounts somewhere else. + RBIND = unix.MS_BIND | unix.MS_REC + + // UNBINDABLE creates a mount which cannot be cloned through a bind operation. + UNBINDABLE = unix.MS_UNBINDABLE + + // RUNBINDABLE marks the entire mount tree as UNBINDABLE. + RUNBINDABLE = unix.MS_UNBINDABLE | unix.MS_REC + + // PRIVATE creates a mount which carries no propagation abilities. + PRIVATE = unix.MS_PRIVATE + + // RPRIVATE marks the entire mount tree as PRIVATE. + RPRIVATE = unix.MS_PRIVATE | unix.MS_REC + + // SLAVE creates a mount which receives propagation from its master, but not + // vice versa. + SLAVE = unix.MS_SLAVE + + // RSLAVE marks the entire mount tree as SLAVE. + RSLAVE = unix.MS_SLAVE | unix.MS_REC + + // SHARED creates a mount which provides the ability to create mirrors of + // that mount such that mounts and unmounts within any of the mirrors + // propagate to the other mirrors. + SHARED = unix.MS_SHARED + + // RSHARED marks the entire mount tree as SHARED. + RSHARED = unix.MS_SHARED | unix.MS_REC + + // RELATIME updates inode access times relative to modify or change time. + RELATIME = unix.MS_RELATIME + + // STRICTATIME allows to explicitly request full atime updates. This makes + // it possible for the kernel to default to relatime or noatime but still + // allow userspace to override it. + STRICTATIME = unix.MS_STRICTATIME + + mntDetach = unix.MNT_DETACH +) diff --git a/vendor/github.com/docker/docker/pkg/mount/flags_unsupported.go b/vendor/github.com/docker/docker/pkg/mount/flags_unsupported.go new file mode 100644 index 000000000..cc6c47590 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/flags_unsupported.go @@ -0,0 +1,31 @@ +// +build !linux,!freebsd freebsd,!cgo + +package mount // import "github.com/docker/docker/pkg/mount" + +// These flags are unsupported. +const ( + BIND = 0 + DIRSYNC = 0 + MANDLOCK = 0 + NOATIME = 0 + NODEV = 0 + NODIRATIME = 0 + NOEXEC = 0 + NOSUID = 0 + UNBINDABLE = 0 + RUNBINDABLE = 0 + PRIVATE = 0 + RPRIVATE = 0 + SHARED = 0 + RSHARED = 0 + SLAVE = 0 + RSLAVE = 0 + RBIND = 0 + RELATIME = 0 + RELATIVE = 0 + REMOUNT = 0 + STRICTATIME = 0 + SYNCHRONOUS = 0 + RDONLY = 0 + mntDetach = 0 +) diff --git a/vendor/github.com/docker/docker/pkg/mount/mount.go b/vendor/github.com/docker/docker/pkg/mount/mount.go new file mode 100644 index 000000000..874aff654 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/mount.go @@ -0,0 +1,141 @@ +package mount // import "github.com/docker/docker/pkg/mount" + +import ( + "sort" + "strings" + "syscall" + + "github.com/sirupsen/logrus" +) + +// FilterFunc is a type defining a callback function +// to filter out unwanted entries. It takes a pointer +// to an Info struct (not fully populated, currently +// only Mountpoint is filled in), and returns two booleans: +// - skip: true if the entry should be skipped +// - stop: true if parsing should be stopped after the entry +type FilterFunc func(*Info) (skip, stop bool) + +// PrefixFilter discards all entries whose mount points +// do not start with a prefix specified +func PrefixFilter(prefix string) FilterFunc { + return func(m *Info) (bool, bool) { + skip := !strings.HasPrefix(m.Mountpoint, prefix) + return skip, false + } +} + +// SingleEntryFilter looks for a specific entry +func SingleEntryFilter(mp string) FilterFunc { + return func(m *Info) (bool, bool) { + if m.Mountpoint == mp { + return false, true // don't skip, stop now + } + return true, false // skip, keep going + } +} + +// ParentsFilter returns all entries whose mount points +// can be parents of a path specified, discarding others. +// For example, given `/var/lib/docker/something`, entries +// like `/var/lib/docker`, `/var` and `/` are returned. +func ParentsFilter(path string) FilterFunc { + return func(m *Info) (bool, bool) { + skip := !strings.HasPrefix(path, m.Mountpoint) + return skip, false + } +} + +// GetMounts retrieves a list of mounts for the current running process, +// with an optional filter applied (use nil for no filter). +func GetMounts(f FilterFunc) ([]*Info, error) { + return parseMountTable(f) +} + +// Mounted determines if a specified mountpoint has been mounted. +// On Linux it looks at /proc/self/mountinfo. +func Mounted(mountpoint string) (bool, error) { + entries, err := GetMounts(SingleEntryFilter(mountpoint)) + if err != nil { + return false, err + } + + return len(entries) > 0, nil +} + +// Mount will mount filesystem according to the specified configuration, on the +// condition that the target path is *not* already mounted. Options must be +// specified like the mount or fstab unix commands: "opt1=val1,opt2=val2". See +// flags.go for supported option flags. +func Mount(device, target, mType, options string) error { + flag, _ := parseOptions(options) + if flag&REMOUNT != REMOUNT { + if mounted, err := Mounted(target); err != nil || mounted { + return err + } + } + return ForceMount(device, target, mType, options) +} + +// ForceMount will mount a filesystem according to the specified configuration, +// *regardless* if the target path is not already mounted. Options must be +// specified like the mount or fstab unix commands: "opt1=val1,opt2=val2". See +// flags.go for supported option flags. +func ForceMount(device, target, mType, options string) error { + flag, data := parseOptions(options) + return mount(device, target, mType, uintptr(flag), data) +} + +// Unmount lazily unmounts a filesystem on supported platforms, otherwise +// does a normal unmount. +func Unmount(target string) error { + err := unmount(target, mntDetach) + if err == syscall.EINVAL { + // ignore "not mounted" error + err = nil + } + return err +} + +// RecursiveUnmount unmounts the target and all mounts underneath, starting with +// the deepsest mount first. +func RecursiveUnmount(target string) error { + mounts, err := parseMountTable(PrefixFilter(target)) + if err != nil { + return err + } + + // Make the deepest mount be first + sort.Slice(mounts, func(i, j int) bool { + return len(mounts[i].Mountpoint) > len(mounts[j].Mountpoint) + }) + + for i, m := range mounts { + logrus.Debugf("Trying to unmount %s", m.Mountpoint) + err = unmount(m.Mountpoint, mntDetach) + if err != nil { + // If the error is EINVAL either this whole package is wrong (invalid flags passed to unmount(2)) or this is + // not a mountpoint (which is ok in this case). + // Meanwhile calling `Mounted()` is very expensive. + // + // We've purposefully used `syscall.EINVAL` here instead of `unix.EINVAL` to avoid platform branching + // Since `EINVAL` is defined for both Windows and Linux in the `syscall` package (and other platforms), + // this is nicer than defining a custom value that we can refer to in each platform file. + if err == syscall.EINVAL { + continue + } + if i == len(mounts)-1 { + if mounted, e := Mounted(m.Mountpoint); e != nil || mounted { + return err + } + continue + } + // This is some submount, we can ignore this error for now, the final unmount will fail if this is a real problem + logrus.WithError(err).Warnf("Failed to unmount submount %s", m.Mountpoint) + continue + } + + logrus.Debugf("Unmounted %s", m.Mountpoint) + } + return nil +} diff --git a/vendor/github.com/docker/docker/pkg/mount/mounter_freebsd.go b/vendor/github.com/docker/docker/pkg/mount/mounter_freebsd.go new file mode 100644 index 000000000..b6ab83a23 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/mounter_freebsd.go @@ -0,0 +1,60 @@ +package mount // import "github.com/docker/docker/pkg/mount" + +/* +#include +#include +#include +#include +#include +#include +*/ +import "C" + +import ( + "fmt" + "strings" + "unsafe" + + "golang.org/x/sys/unix" +) + +func allocateIOVecs(options []string) []C.struct_iovec { + out := make([]C.struct_iovec, len(options)) + for i, option := range options { + out[i].iov_base = unsafe.Pointer(C.CString(option)) + out[i].iov_len = C.size_t(len(option) + 1) + } + return out +} + +func mount(device, target, mType string, flag uintptr, data string) error { + isNullFS := false + + xs := strings.Split(data, ",") + for _, x := range xs { + if x == "bind" { + isNullFS = true + } + } + + options := []string{"fspath", target} + if isNullFS { + options = append(options, "fstype", "nullfs", "target", device) + } else { + options = append(options, "fstype", mType, "from", device) + } + rawOptions := allocateIOVecs(options) + for _, rawOption := range rawOptions { + defer C.free(rawOption.iov_base) + } + + if errno := C.nmount(&rawOptions[0], C.uint(len(options)), C.int(flag)); errno != 0 { + reason := C.GoString(C.strerror(*C.__error())) + return fmt.Errorf("Failed to call nmount: %s", reason) + } + return nil +} + +func unmount(target string, flag int) error { + return unix.Unmount(target, flag) +} diff --git a/vendor/github.com/docker/docker/pkg/mount/mounter_linux.go b/vendor/github.com/docker/docker/pkg/mount/mounter_linux.go new file mode 100644 index 000000000..631daf10a --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/mounter_linux.go @@ -0,0 +1,57 @@ +package mount // import "github.com/docker/docker/pkg/mount" + +import ( + "golang.org/x/sys/unix" +) + +const ( + // ptypes is the set propagation types. + ptypes = unix.MS_SHARED | unix.MS_PRIVATE | unix.MS_SLAVE | unix.MS_UNBINDABLE + + // pflags is the full set valid flags for a change propagation call. + pflags = ptypes | unix.MS_REC | unix.MS_SILENT + + // broflags is the combination of bind and read only + broflags = unix.MS_BIND | unix.MS_RDONLY +) + +// isremount returns true if either device name or flags identify a remount request, false otherwise. +func isremount(device string, flags uintptr) bool { + switch { + // We treat device "" and "none" as a remount request to provide compatibility with + // requests that don't explicitly set MS_REMOUNT such as those manipulating bind mounts. + case flags&unix.MS_REMOUNT != 0, device == "", device == "none": + return true + default: + return false + } +} + +func mount(device, target, mType string, flags uintptr, data string) error { + oflags := flags &^ ptypes + if !isremount(device, flags) || data != "" { + // Initial call applying all non-propagation flags for mount + // or remount with changed data + if err := unix.Mount(device, target, mType, oflags, data); err != nil { + return err + } + } + + if flags&ptypes != 0 { + // Change the propagation type. + if err := unix.Mount("", target, "", flags&pflags, ""); err != nil { + return err + } + } + + if oflags&broflags == broflags { + // Remount the bind to apply read only. + return unix.Mount("", target, "", oflags|unix.MS_REMOUNT, "") + } + + return nil +} + +func unmount(target string, flag int) error { + return unix.Unmount(target, flag) +} diff --git a/vendor/github.com/docker/docker/pkg/mount/mounter_unsupported.go b/vendor/github.com/docker/docker/pkg/mount/mounter_unsupported.go new file mode 100644 index 000000000..1428dffa5 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/mounter_unsupported.go @@ -0,0 +1,11 @@ +// +build !linux,!freebsd freebsd,!cgo + +package mount // import "github.com/docker/docker/pkg/mount" + +func mount(device, target, mType string, flag uintptr, data string) error { + panic("Not implemented") +} + +func unmount(target string, flag int) error { + panic("Not implemented") +} diff --git a/vendor/github.com/docker/docker/pkg/mount/mountinfo.go b/vendor/github.com/docker/docker/pkg/mount/mountinfo.go new file mode 100644 index 000000000..ecd03fc02 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/mountinfo.go @@ -0,0 +1,40 @@ +package mount // import "github.com/docker/docker/pkg/mount" + +// Info reveals information about a particular mounted filesystem. This +// struct is populated from the content in the /proc//mountinfo file. +type Info struct { + // ID is a unique identifier of the mount (may be reused after umount). + ID int + + // Parent indicates the ID of the mount parent (or of self for the top of the + // mount tree). + Parent int + + // Major indicates one half of the device ID which identifies the device class. + Major int + + // Minor indicates one half of the device ID which identifies a specific + // instance of device. + Minor int + + // Root of the mount within the filesystem. + Root string + + // Mountpoint indicates the mount point relative to the process's root. + Mountpoint string + + // Opts represents mount-specific options. + Opts string + + // Optional represents optional fields. + Optional string + + // Fstype indicates the type of filesystem, such as EXT3. + Fstype string + + // Source indicates filesystem specific information or "none". + Source string + + // VfsOpts represents per super block options. + VfsOpts string +} diff --git a/vendor/github.com/docker/docker/pkg/mount/mountinfo_freebsd.go b/vendor/github.com/docker/docker/pkg/mount/mountinfo_freebsd.go new file mode 100644 index 000000000..36c89dc1a --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/mountinfo_freebsd.go @@ -0,0 +1,55 @@ +package mount // import "github.com/docker/docker/pkg/mount" + +/* +#include +#include +#include +*/ +import "C" + +import ( + "fmt" + "reflect" + "unsafe" +) + +// Parse /proc/self/mountinfo because comparing Dev and ino does not work from +// bind mounts. +func parseMountTable(filter FilterFunc) ([]*Info, error) { + var rawEntries *C.struct_statfs + + count := int(C.getmntinfo(&rawEntries, C.MNT_WAIT)) + if count == 0 { + return nil, fmt.Errorf("Failed to call getmntinfo") + } + + var entries []C.struct_statfs + header := (*reflect.SliceHeader)(unsafe.Pointer(&entries)) + header.Cap = count + header.Len = count + header.Data = uintptr(unsafe.Pointer(rawEntries)) + + var out []*Info + for _, entry := range entries { + var mountinfo Info + var skip, stop bool + mountinfo.Mountpoint = C.GoString(&entry.f_mntonname[0]) + + if filter != nil { + // filter out entries we're not interested in + skip, stop = filter(p) + if skip { + continue + } + } + + mountinfo.Source = C.GoString(&entry.f_mntfromname[0]) + mountinfo.Fstype = C.GoString(&entry.f_fstypename[0]) + + out = append(out, &mountinfo) + if stop { + break + } + } + return out, nil +} diff --git a/vendor/github.com/docker/docker/pkg/mount/mountinfo_linux.go b/vendor/github.com/docker/docker/pkg/mount/mountinfo_linux.go new file mode 100644 index 000000000..c1dba01fc --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/mountinfo_linux.go @@ -0,0 +1,132 @@ +package mount // import "github.com/docker/docker/pkg/mount" + +import ( + "bufio" + "fmt" + "io" + "os" + "strconv" + "strings" +) + +func parseInfoFile(r io.Reader, filter FilterFunc) ([]*Info, error) { + s := bufio.NewScanner(r) + out := []*Info{} + for s.Scan() { + if err := s.Err(); err != nil { + return nil, err + } + /* + 36 35 98:0 /mnt1 /mnt2 rw,noatime master:1 - ext3 /dev/root rw,errors=continue + (1)(2)(3) (4) (5) (6) (7) (8) (9) (10) (11) + + (1) mount ID: unique identifier of the mount (may be reused after umount) + (2) parent ID: ID of parent (or of self for the top of the mount tree) + (3) major:minor: value of st_dev for files on filesystem + (4) root: root of the mount within the filesystem + (5) mount point: mount point relative to the process's root + (6) mount options: per mount options + (7) optional fields: zero or more fields of the form "tag[:value]" + (8) separator: marks the end of the optional fields + (9) filesystem type: name of filesystem of the form "type[.subtype]" + (10) mount source: filesystem specific information or "none" + (11) super options: per super block options + */ + + text := s.Text() + fields := strings.Split(text, " ") + numFields := len(fields) + if numFields < 10 { + // should be at least 10 fields + return nil, fmt.Errorf("Parsing '%s' failed: not enough fields (%d)", text, numFields) + } + + p := &Info{} + // ignore any numbers parsing errors, as there should not be any + p.ID, _ = strconv.Atoi(fields[0]) + p.Parent, _ = strconv.Atoi(fields[1]) + mm := strings.Split(fields[2], ":") + if len(mm) != 2 { + return nil, fmt.Errorf("Parsing '%s' failed: unexpected minor:major pair %s", text, mm) + } + p.Major, _ = strconv.Atoi(mm[0]) + p.Minor, _ = strconv.Atoi(mm[1]) + + p.Root = fields[3] + p.Mountpoint = fields[4] + p.Opts = fields[5] + + var skip, stop bool + if filter != nil { + // filter out entries we're not interested in + skip, stop = filter(p) + if skip { + continue + } + } + + // one or more optional fields, when a separator (-) + i := 6 + for ; i < numFields && fields[i] != "-"; i++ { + switch i { + case 6: + p.Optional = fields[6] + default: + /* NOTE there might be more optional fields before the such as + fields[7]...fields[N] (where N < sepIndex), although + as of Linux kernel 4.15 the only known ones are + mount propagation flags in fields[6]. The correct + behavior is to ignore any unknown optional fields. + */ + break + } + } + if i == numFields { + return nil, fmt.Errorf("Parsing '%s' failed: missing separator ('-')", text) + } + + // There should be 3 fields after the separator... + if i+4 > numFields { + return nil, fmt.Errorf("Parsing '%s' failed: not enough fields after a separator", text) + } + // ... but in Linux <= 3.9 mounting a cifs with spaces in a share name + // (like "//serv/My Documents") _may_ end up having a space in the last field + // of mountinfo (like "unc=//serv/My Documents"). Since kernel 3.10-rc1, cifs + // option unc= is ignored, so a space should not appear. In here we ignore + // those "extra" fields caused by extra spaces. + p.Fstype = fields[i+1] + p.Source = fields[i+2] + p.VfsOpts = fields[i+3] + + out = append(out, p) + if stop { + break + } + } + return out, nil +} + +// Parse /proc/self/mountinfo because comparing Dev and ino does not work from +// bind mounts +func parseMountTable(filter FilterFunc) ([]*Info, error) { + f, err := os.Open("/proc/self/mountinfo") + if err != nil { + return nil, err + } + defer f.Close() + + return parseInfoFile(f, filter) +} + +// PidMountInfo collects the mounts for a specific process ID. If the process +// ID is unknown, it is better to use `GetMounts` which will inspect +// "/proc/self/mountinfo" instead. +func PidMountInfo(pid int) ([]*Info, error) { + f, err := os.Open(fmt.Sprintf("/proc/%d/mountinfo", pid)) + if err != nil { + return nil, err + } + defer f.Close() + + return parseInfoFile(f, nil) +} diff --git a/vendor/github.com/docker/docker/pkg/mount/mountinfo_unsupported.go b/vendor/github.com/docker/docker/pkg/mount/mountinfo_unsupported.go new file mode 100644 index 000000000..fd16d3ed6 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/mountinfo_unsupported.go @@ -0,0 +1,12 @@ +// +build !windows,!linux,!freebsd freebsd,!cgo + +package mount // import "github.com/docker/docker/pkg/mount" + +import ( + "fmt" + "runtime" +) + +func parseMountTable(f FilterFunc) ([]*Info, error) { + return nil, fmt.Errorf("mount.parseMountTable is not implemented on %s/%s", runtime.GOOS, runtime.GOARCH) +} diff --git a/vendor/github.com/docker/docker/pkg/mount/mountinfo_windows.go b/vendor/github.com/docker/docker/pkg/mount/mountinfo_windows.go new file mode 100644 index 000000000..27e0f6976 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/mountinfo_windows.go @@ -0,0 +1,6 @@ +package mount // import "github.com/docker/docker/pkg/mount" + +func parseMountTable(f FilterFunc) ([]*Info, error) { + // Do NOT return an error! + return nil, nil +} diff --git a/vendor/github.com/docker/docker/pkg/mount/sharedsubtree_linux.go b/vendor/github.com/docker/docker/pkg/mount/sharedsubtree_linux.go new file mode 100644 index 000000000..538f6637a --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/mount/sharedsubtree_linux.go @@ -0,0 +1,67 @@ +package mount // import "github.com/docker/docker/pkg/mount" + +// MakeShared ensures a mounted filesystem has the SHARED mount option enabled. +// See the supported options in flags.go for further reference. +func MakeShared(mountPoint string) error { + return ensureMountedAs(mountPoint, "shared") +} + +// MakeRShared ensures a mounted filesystem has the RSHARED mount option enabled. +// See the supported options in flags.go for further reference. +func MakeRShared(mountPoint string) error { + return ensureMountedAs(mountPoint, "rshared") +} + +// MakePrivate ensures a mounted filesystem has the PRIVATE mount option enabled. +// See the supported options in flags.go for further reference. +func MakePrivate(mountPoint string) error { + return ensureMountedAs(mountPoint, "private") +} + +// MakeRPrivate ensures a mounted filesystem has the RPRIVATE mount option +// enabled. See the supported options in flags.go for further reference. +func MakeRPrivate(mountPoint string) error { + return ensureMountedAs(mountPoint, "rprivate") +} + +// MakeSlave ensures a mounted filesystem has the SLAVE mount option enabled. +// See the supported options in flags.go for further reference. +func MakeSlave(mountPoint string) error { + return ensureMountedAs(mountPoint, "slave") +} + +// MakeRSlave ensures a mounted filesystem has the RSLAVE mount option enabled. +// See the supported options in flags.go for further reference. +func MakeRSlave(mountPoint string) error { + return ensureMountedAs(mountPoint, "rslave") +} + +// MakeUnbindable ensures a mounted filesystem has the UNBINDABLE mount option +// enabled. See the supported options in flags.go for further reference. +func MakeUnbindable(mountPoint string) error { + return ensureMountedAs(mountPoint, "unbindable") +} + +// MakeRUnbindable ensures a mounted filesystem has the RUNBINDABLE mount +// option enabled. See the supported options in flags.go for further reference. +func MakeRUnbindable(mountPoint string) error { + return ensureMountedAs(mountPoint, "runbindable") +} + +func ensureMountedAs(mountPoint, options string) error { + mounted, err := Mounted(mountPoint) + if err != nil { + return err + } + + if !mounted { + if err := Mount(mountPoint, mountPoint, "none", "bind,rw"); err != nil { + return err + } + } + if _, err = Mounted(mountPoint); err != nil { + return err + } + + return ForceMount("", mountPoint, "none", options) +} diff --git a/vendor/github.com/docker/docker/pkg/pools/pools.go b/vendor/github.com/docker/docker/pkg/pools/pools.go index 76e84f9d7..46339c282 100644 --- a/vendor/github.com/docker/docker/pkg/pools/pools.go +++ b/vendor/github.com/docker/docker/pkg/pools/pools.go @@ -7,7 +7,7 @@ // // Utility functions which operate on pools should be added to this // package to allow them to be reused. -package pools +package pools // import "github.com/docker/docker/pkg/pools" import ( "bufio" @@ -17,32 +17,29 @@ import ( "github.com/docker/docker/pkg/ioutils" ) +const buffer32K = 32 * 1024 + var ( // BufioReader32KPool is a pool which returns bufio.Reader with a 32K buffer. - BufioReader32KPool *BufioReaderPool + BufioReader32KPool = newBufioReaderPoolWithSize(buffer32K) // BufioWriter32KPool is a pool which returns bufio.Writer with a 32K buffer. - BufioWriter32KPool *BufioWriterPool + BufioWriter32KPool = newBufioWriterPoolWithSize(buffer32K) + buffer32KPool = newBufferPoolWithSize(buffer32K) ) -const buffer32K = 32 * 1024 - // BufioReaderPool is a bufio reader that uses sync.Pool. type BufioReaderPool struct { pool sync.Pool } -func init() { - BufioReader32KPool = newBufioReaderPoolWithSize(buffer32K) - BufioWriter32KPool = newBufioWriterPoolWithSize(buffer32K) -} - // newBufioReaderPoolWithSize is unexported because new pools should be // added here to be shared where required. func newBufioReaderPoolWithSize(size int) *BufioReaderPool { - pool := sync.Pool{ - New: func() interface{} { return bufio.NewReaderSize(nil, size) }, + return &BufioReaderPool{ + pool: sync.Pool{ + New: func() interface{} { return bufio.NewReaderSize(nil, size) }, + }, } - return &BufioReaderPool{pool: pool} } // Get returns a bufio.Reader which reads from r. The buffer size is that of the pool. @@ -58,11 +55,31 @@ func (bufPool *BufioReaderPool) Put(b *bufio.Reader) { bufPool.pool.Put(b) } +type bufferPool struct { + pool sync.Pool +} + +func newBufferPoolWithSize(size int) *bufferPool { + return &bufferPool{ + pool: sync.Pool{ + New: func() interface{} { return make([]byte, size) }, + }, + } +} + +func (bp *bufferPool) Get() []byte { + return bp.pool.Get().([]byte) +} + +func (bp *bufferPool) Put(b []byte) { + bp.pool.Put(b) +} + // Copy is a convenience wrapper which uses a buffer to avoid allocation in io.Copy. func Copy(dst io.Writer, src io.Reader) (written int64, err error) { - buf := BufioReader32KPool.Get(src) - written, err = io.Copy(dst, buf) - BufioReader32KPool.Put(buf) + buf := buffer32KPool.Get() + written, err = io.CopyBuffer(dst, src, buf) + buffer32KPool.Put(buf) return } @@ -86,10 +103,11 @@ type BufioWriterPool struct { // newBufioWriterPoolWithSize is unexported because new pools should be // added here to be shared where required. func newBufioWriterPoolWithSize(size int) *BufioWriterPool { - pool := sync.Pool{ - New: func() interface{} { return bufio.NewWriterSize(nil, size) }, + return &BufioWriterPool{ + pool: sync.Pool{ + New: func() interface{} { return bufio.NewWriterSize(nil, size) }, + }, } - return &BufioWriterPool{pool: pool} } // Get returns a bufio.Writer which writes to w. The buffer size is that of the pool. diff --git a/vendor/github.com/docker/docker/pkg/promise/promise.go b/vendor/github.com/docker/docker/pkg/promise/promise.go deleted file mode 100644 index dd52b9082..000000000 --- a/vendor/github.com/docker/docker/pkg/promise/promise.go +++ /dev/null @@ -1,11 +0,0 @@ -package promise - -// Go is a basic promise implementation: it wraps calls a function in a goroutine, -// and returns a channel which will later return the function's return value. -func Go(f func() error) chan error { - ch := make(chan error, 1) - go func() { - ch <- f() - }() - return ch -} diff --git a/vendor/github.com/docker/docker/pkg/random/random.go b/vendor/github.com/docker/docker/pkg/random/random.go deleted file mode 100644 index 70de4d130..000000000 --- a/vendor/github.com/docker/docker/pkg/random/random.go +++ /dev/null @@ -1,71 +0,0 @@ -package random - -import ( - cryptorand "crypto/rand" - "io" - "math" - "math/big" - "math/rand" - "sync" - "time" -) - -// Rand is a global *rand.Rand instance, which initialized with NewSource() source. -var Rand = rand.New(NewSource()) - -// Reader is a global, shared instance of a pseudorandom bytes generator. -// It doesn't consume entropy. -var Reader io.Reader = &reader{rnd: Rand} - -// copypaste from standard math/rand -type lockedSource struct { - lk sync.Mutex - src rand.Source -} - -func (r *lockedSource) Int63() (n int64) { - r.lk.Lock() - n = r.src.Int63() - r.lk.Unlock() - return -} - -func (r *lockedSource) Seed(seed int64) { - r.lk.Lock() - r.src.Seed(seed) - r.lk.Unlock() -} - -// NewSource returns math/rand.Source safe for concurrent use and initialized -// with current unix-nano timestamp -func NewSource() rand.Source { - var seed int64 - if cryptoseed, err := cryptorand.Int(cryptorand.Reader, big.NewInt(math.MaxInt64)); err != nil { - // This should not happen, but worst-case fallback to time-based seed. - seed = time.Now().UnixNano() - } else { - seed = cryptoseed.Int64() - } - return &lockedSource{ - src: rand.NewSource(seed), - } -} - -type reader struct { - rnd *rand.Rand -} - -func (r *reader) Read(b []byte) (int, error) { - i := 0 - for { - val := r.rnd.Int63() - for val > 0 { - b[i] = byte(val) - i++ - if i == len(b) { - return i, nil - } - val >>= 8 - } - } -} diff --git a/vendor/github.com/docker/docker/pkg/stdcopy/stdcopy.go b/vendor/github.com/docker/docker/pkg/stdcopy/stdcopy.go index 8f67ece94..8f6e0a737 100644 --- a/vendor/github.com/docker/docker/pkg/stdcopy/stdcopy.go +++ b/vendor/github.com/docker/docker/pkg/stdcopy/stdcopy.go @@ -1,4 +1,4 @@ -package stdcopy +package stdcopy // import "github.com/docker/docker/pkg/stdcopy" import ( "bytes" @@ -7,8 +7,6 @@ import ( "fmt" "io" "sync" - - "github.com/Sirupsen/logrus" ) // StdType is the type of standard stream @@ -22,6 +20,9 @@ const ( Stdout // Stderr represents standard error steam type. Stderr + // Systemerr represents errors originating from the system that make it + // into the multiplexed stream. + Systemerr stdWriterPrefixLen = 8 stdWriterFdIndex = 0 @@ -108,19 +109,18 @@ func StdCopy(dstout, dsterr io.Writer, src io.Reader) (written int64, err error) nr += nr2 if er == io.EOF { if nr < stdWriterPrefixLen { - logrus.Debugf("Corrupted prefix: %v", buf[:nr]) return written, nil } break } if er != nil { - logrus.Debugf("Error reading header: %s", er) return 0, er } } + stream := StdType(buf[stdWriterFdIndex]) // Check the first byte to know where to write - switch StdType(buf[stdWriterFdIndex]) { + switch stream { case Stdin: fallthrough case Stdout: @@ -129,19 +129,21 @@ func StdCopy(dstout, dsterr io.Writer, src io.Reader) (written int64, err error) case Stderr: // Write on stderr out = dsterr + case Systemerr: + // If we're on Systemerr, we won't write anywhere. + // NB: if this code changes later, make sure you don't try to write + // to outstream if Systemerr is the stream + out = nil default: - logrus.Debugf("Error selecting output fd: (%d)", buf[stdWriterFdIndex]) return 0, fmt.Errorf("Unrecognized input header: %d", buf[stdWriterFdIndex]) } // Retrieve the size of the frame frameSize = int(binary.BigEndian.Uint32(buf[stdWriterSizeIndex : stdWriterSizeIndex+4])) - logrus.Debugf("framesize: %d", frameSize) // Check if the buffer is big enough to read the frame. // Extend it if necessary. if frameSize+stdWriterPrefixLen > bufLen { - logrus.Debugf("Extending buffer cap by %d (was %d)", frameSize+stdWriterPrefixLen-bufLen+1, len(buf)) buf = append(buf, make([]byte, frameSize+stdWriterPrefixLen-bufLen+1)...) bufLen = len(buf) } @@ -153,26 +155,29 @@ func StdCopy(dstout, dsterr io.Writer, src io.Reader) (written int64, err error) nr += nr2 if er == io.EOF { if nr < frameSize+stdWriterPrefixLen { - logrus.Debugf("Corrupted frame: %v", buf[stdWriterPrefixLen:nr]) return written, nil } break } if er != nil { - logrus.Debugf("Error reading frame: %s", er) return 0, er } } + // we might have an error from the source mixed up in our multiplexed + // stream. if we do, return it. + if stream == Systemerr { + return written, fmt.Errorf("error from daemon in stream: %s", string(buf[stdWriterPrefixLen:frameSize+stdWriterPrefixLen])) + } + // Write the retrieved frame (without header) nw, ew = out.Write(buf[stdWriterPrefixLen : frameSize+stdWriterPrefixLen]) if ew != nil { - logrus.Debugf("Error writing frame: %s", ew) return 0, ew } + // If the frame has not been fully written: error if nw != frameSize { - logrus.Debugf("Error Short Write: (%d on %d)", nw, frameSize) return 0, io.ErrShortWrite } written += int64(nw) diff --git a/vendor/github.com/docker/docker/pkg/stringid/stringid.go b/vendor/github.com/docker/docker/pkg/stringid/stringid.go index 82f85596c..fa7d9166e 100644 --- a/vendor/github.com/docker/docker/pkg/stringid/stringid.go +++ b/vendor/github.com/docker/docker/pkg/stringid/stringid.go @@ -1,16 +1,18 @@ // Package stringid provides helper functions for dealing with string identifiers -package stringid +package stringid // import "github.com/docker/docker/pkg/stringid" import ( - "crypto/rand" + cryptorand "crypto/rand" "encoding/hex" "fmt" "io" + "math" + "math/big" + "math/rand" "regexp" "strconv" "strings" - - "github.com/docker/docker/pkg/random" + "time" ) const shortLen = 12 @@ -39,12 +41,8 @@ func TruncateID(id string) string { return id } -func generateID(crypto bool) string { +func generateID(r io.Reader) string { b := make([]byte, 32) - r := random.Reader - if crypto { - r = rand.Reader - } for { if _, err := io.ReadFull(r, b); err != nil { panic(err) // This shouldn't happen @@ -62,14 +60,14 @@ func generateID(crypto bool) string { // GenerateRandomID returns a unique id. func GenerateRandomID() string { - return generateID(true) + return generateID(cryptorand.Reader) } // GenerateNonCryptoID generates unique id without using cryptographically // secure sources of random. // It helps you to save entropy. func GenerateNonCryptoID() string { - return generateID(false) + return generateID(readerFunc(rand.Read)) } // ValidateID checks whether an ID string is a valid image ID. @@ -79,3 +77,23 @@ func ValidateID(id string) error { } return nil } + +func init() { + // safely set the seed globally so we generate random ids. Tries to use a + // crypto seed before falling back to time. + var seed int64 + if cryptoseed, err := cryptorand.Int(cryptorand.Reader, big.NewInt(math.MaxInt64)); err != nil { + // This should not happen, but worst-case fallback to time-based seed. + seed = time.Now().UnixNano() + } else { + seed = cryptoseed.Int64() + } + + rand.Seed(seed) +} + +type readerFunc func(p []byte) (int, error) + +func (fn readerFunc) Read(p []byte) (int, error) { + return fn(p) +} diff --git a/vendor/github.com/docker/docker/pkg/system/chtimes.go b/vendor/github.com/docker/docker/pkg/system/chtimes.go index 7637f12e1..c26a4e24b 100644 --- a/vendor/github.com/docker/docker/pkg/system/chtimes.go +++ b/vendor/github.com/docker/docker/pkg/system/chtimes.go @@ -1,27 +1,10 @@ -package system +package system // import "github.com/docker/docker/pkg/system" import ( "os" - "syscall" "time" - "unsafe" ) -var ( - maxTime time.Time -) - -func init() { - if unsafe.Sizeof(syscall.Timespec{}.Nsec) == 8 { - // This is a 64 bit timespec - // os.Chtimes limits time to the following - maxTime = time.Unix(0, 1<<63-1) - } else { - // This is a 32 bit timespec - maxTime = time.Unix(1<<31-1, 0) - } -} - // Chtimes changes the access time and modified time of a file at the given path func Chtimes(name string, atime time.Time, mtime time.Time) error { unixMinTime := time.Unix(0, 0) @@ -44,9 +27,5 @@ func Chtimes(name string, atime time.Time, mtime time.Time) error { } // Take platform specific action for setting create time. - if err := setCTime(name, mtime); err != nil { - return err - } - - return nil + return setCTime(name, mtime) } diff --git a/vendor/github.com/docker/docker/pkg/system/chtimes_unix.go b/vendor/github.com/docker/docker/pkg/system/chtimes_unix.go index 09d58bcbf..259138a45 100644 --- a/vendor/github.com/docker/docker/pkg/system/chtimes_unix.go +++ b/vendor/github.com/docker/docker/pkg/system/chtimes_unix.go @@ -1,6 +1,6 @@ // +build !windows -package system +package system // import "github.com/docker/docker/pkg/system" import ( "time" diff --git a/vendor/github.com/docker/docker/pkg/system/chtimes_windows.go b/vendor/github.com/docker/docker/pkg/system/chtimes_windows.go index 294586846..d3a115ff4 100644 --- a/vendor/github.com/docker/docker/pkg/system/chtimes_windows.go +++ b/vendor/github.com/docker/docker/pkg/system/chtimes_windows.go @@ -1,27 +1,26 @@ -// +build windows - -package system +package system // import "github.com/docker/docker/pkg/system" import ( - "syscall" "time" + + "golang.org/x/sys/windows" ) //setCTime will set the create time on a file. On Windows, this requires //calling SetFileTime and explicitly including the create time. func setCTime(path string, ctime time.Time) error { - ctimespec := syscall.NsecToTimespec(ctime.UnixNano()) - pathp, e := syscall.UTF16PtrFromString(path) + ctimespec := windows.NsecToTimespec(ctime.UnixNano()) + pathp, e := windows.UTF16PtrFromString(path) if e != nil { return e } - h, e := syscall.CreateFile(pathp, - syscall.FILE_WRITE_ATTRIBUTES, syscall.FILE_SHARE_WRITE, nil, - syscall.OPEN_EXISTING, syscall.FILE_FLAG_BACKUP_SEMANTICS, 0) + h, e := windows.CreateFile(pathp, + windows.FILE_WRITE_ATTRIBUTES, windows.FILE_SHARE_WRITE, nil, + windows.OPEN_EXISTING, windows.FILE_FLAG_BACKUP_SEMANTICS, 0) if e != nil { return e } - defer syscall.Close(h) - c := syscall.NsecToFiletime(syscall.TimespecToNsec(ctimespec)) - return syscall.SetFileTime(h, &c, nil, nil) + defer windows.Close(h) + c := windows.NsecToFiletime(windows.TimespecToNsec(ctimespec)) + return windows.SetFileTime(h, &c, nil, nil) } diff --git a/vendor/github.com/docker/docker/pkg/system/errors.go b/vendor/github.com/docker/docker/pkg/system/errors.go index 288318985..2573d7162 100644 --- a/vendor/github.com/docker/docker/pkg/system/errors.go +++ b/vendor/github.com/docker/docker/pkg/system/errors.go @@ -1,4 +1,4 @@ -package system +package system // import "github.com/docker/docker/pkg/system" import ( "errors" @@ -7,4 +7,7 @@ import ( var ( // ErrNotSupportedPlatform means the platform is not supported. ErrNotSupportedPlatform = errors.New("platform and architecture is not supported") + + // ErrNotSupportedOperatingSystem means the operating system is not supported. + ErrNotSupportedOperatingSystem = errors.New("operating system is not supported") ) diff --git a/vendor/github.com/docker/docker/pkg/system/events_windows.go b/vendor/github.com/docker/docker/pkg/system/events_windows.go deleted file mode 100644 index 3ec6d2215..000000000 --- a/vendor/github.com/docker/docker/pkg/system/events_windows.go +++ /dev/null @@ -1,85 +0,0 @@ -package system - -// This file implements syscalls for Win32 events which are not implemented -// in golang. - -import ( - "syscall" - "unsafe" - - "golang.org/x/sys/windows" -) - -var ( - procCreateEvent = modkernel32.NewProc("CreateEventW") - procOpenEvent = modkernel32.NewProc("OpenEventW") - procSetEvent = modkernel32.NewProc("SetEvent") - procResetEvent = modkernel32.NewProc("ResetEvent") - procPulseEvent = modkernel32.NewProc("PulseEvent") -) - -// CreateEvent implements win32 CreateEventW func in golang. It will create an event object. -func CreateEvent(eventAttributes *syscall.SecurityAttributes, manualReset bool, initialState bool, name string) (handle syscall.Handle, err error) { - namep, _ := syscall.UTF16PtrFromString(name) - var _p1 uint32 - if manualReset { - _p1 = 1 - } - var _p2 uint32 - if initialState { - _p2 = 1 - } - r0, _, e1 := procCreateEvent.Call(uintptr(unsafe.Pointer(eventAttributes)), uintptr(_p1), uintptr(_p2), uintptr(unsafe.Pointer(namep))) - use(unsafe.Pointer(namep)) - handle = syscall.Handle(r0) - if handle == syscall.InvalidHandle { - err = e1 - } - return -} - -// OpenEvent implements win32 OpenEventW func in golang. It opens an event object. -func OpenEvent(desiredAccess uint32, inheritHandle bool, name string) (handle syscall.Handle, err error) { - namep, _ := syscall.UTF16PtrFromString(name) - var _p1 uint32 - if inheritHandle { - _p1 = 1 - } - r0, _, e1 := procOpenEvent.Call(uintptr(desiredAccess), uintptr(_p1), uintptr(unsafe.Pointer(namep))) - use(unsafe.Pointer(namep)) - handle = syscall.Handle(r0) - if handle == syscall.InvalidHandle { - err = e1 - } - return -} - -// SetEvent implements win32 SetEvent func in golang. -func SetEvent(handle syscall.Handle) (err error) { - return setResetPulse(handle, procSetEvent) -} - -// ResetEvent implements win32 ResetEvent func in golang. -func ResetEvent(handle syscall.Handle) (err error) { - return setResetPulse(handle, procResetEvent) -} - -// PulseEvent implements win32 PulseEvent func in golang. -func PulseEvent(handle syscall.Handle) (err error) { - return setResetPulse(handle, procPulseEvent) -} - -func setResetPulse(handle syscall.Handle, proc *windows.LazyProc) (err error) { - r0, _, _ := proc.Call(uintptr(handle)) - if r0 != 0 { - err = syscall.Errno(r0) - } - return -} - -var temp unsafe.Pointer - -// use ensures a variable is kept alive without the GC freeing while still needed -func use(p unsafe.Pointer) { - temp = p -} diff --git a/vendor/github.com/docker/docker/pkg/system/exitcode.go b/vendor/github.com/docker/docker/pkg/system/exitcode.go index 60f0514b1..4ba8fe35b 100644 --- a/vendor/github.com/docker/docker/pkg/system/exitcode.go +++ b/vendor/github.com/docker/docker/pkg/system/exitcode.go @@ -1,4 +1,4 @@ -package system +package system // import "github.com/docker/docker/pkg/system" import ( "fmt" @@ -17,17 +17,3 @@ func GetExitCode(err error) (int, error) { } return exitCode, fmt.Errorf("failed to get exit code") } - -// ProcessExitCode process the specified error and returns the exit status code -// if the error was of type exec.ExitError, returns nothing otherwise. -func ProcessExitCode(err error) (exitCode int) { - if err != nil { - var exiterr error - if exitCode, exiterr = GetExitCode(err); exiterr != nil { - // TODO: Fix this so we check the error's text. - // we've failed to retrieve exit code, so we set it to 127 - exitCode = 127 - } - } - return -} diff --git a/vendor/github.com/docker/docker/pkg/system/filesys.go b/vendor/github.com/docker/docker/pkg/system/filesys.go index 7aa920de1..adeb16305 100644 --- a/vendor/github.com/docker/docker/pkg/system/filesys.go +++ b/vendor/github.com/docker/docker/pkg/system/filesys.go @@ -1,6 +1,6 @@ // +build !windows -package system +package system // import "github.com/docker/docker/pkg/system" import ( "io/ioutil" @@ -8,15 +8,14 @@ import ( "path/filepath" ) -// MkdirAllWithACL is a wrapper for MkdirAll that creates a directory -// ACL'd for Builtin Administrators and Local System. -func MkdirAllWithACL(path string, perm os.FileMode) error { - return MkdirAll(path, perm) +// MkdirAllWithACL is a wrapper for MkdirAll on unix systems. +func MkdirAllWithACL(path string, perm os.FileMode, sddl string) error { + return MkdirAll(path, perm, sddl) } // MkdirAll creates a directory named path along with any necessary parents, // with permission specified by attribute perm for all dir created. -func MkdirAll(path string, perm os.FileMode) error { +func MkdirAll(path string, perm os.FileMode, sddl string) error { return os.MkdirAll(path, perm) } diff --git a/vendor/github.com/docker/docker/pkg/system/filesys_windows.go b/vendor/github.com/docker/docker/pkg/system/filesys_windows.go index 626d2ad88..a1f6013f1 100644 --- a/vendor/github.com/docker/docker/pkg/system/filesys_windows.go +++ b/vendor/github.com/docker/docker/pkg/system/filesys_windows.go @@ -1,6 +1,4 @@ -// +build windows - -package system +package system // import "github.com/docker/docker/pkg/system" import ( "os" @@ -14,23 +12,31 @@ import ( "unsafe" winio "github.com/Microsoft/go-winio" + "golang.org/x/sys/windows" +) + +const ( + // SddlAdministratorsLocalSystem is local administrators plus NT AUTHORITY\System + SddlAdministratorsLocalSystem = "D:P(A;OICI;GA;;;BA)(A;OICI;GA;;;SY)" + // SddlNtvmAdministratorsLocalSystem is NT VIRTUAL MACHINE\Virtual Machines plus local administrators plus NT AUTHORITY\System + SddlNtvmAdministratorsLocalSystem = "D:P(A;OICI;GA;;;S-1-5-83-0)(A;OICI;GA;;;BA)(A;OICI;GA;;;SY)" ) // MkdirAllWithACL is a wrapper for MkdirAll that creates a directory -// ACL'd for Builtin Administrators and Local System. -func MkdirAllWithACL(path string, perm os.FileMode) error { - return mkdirall(path, true) +// with an appropriate SDDL defined ACL. +func MkdirAllWithACL(path string, perm os.FileMode, sddl string) error { + return mkdirall(path, true, sddl) } // MkdirAll implementation that is volume path aware for Windows. -func MkdirAll(path string, _ os.FileMode) error { - return mkdirall(path, false) +func MkdirAll(path string, _ os.FileMode, sddl string) error { + return mkdirall(path, false, sddl) } // mkdirall is a custom version of os.MkdirAll modified for use on Windows // so that it is both volume path aware, and can create a directory with // a DACL. -func mkdirall(path string, adminAndLocalSystem bool) error { +func mkdirall(path string, applyACL bool, sddl string) error { if re := regexp.MustCompile(`^\\\\\?\\Volume{[a-z0-9-]+}$`); re.MatchString(path) { return nil } @@ -64,15 +70,15 @@ func mkdirall(path string, adminAndLocalSystem bool) error { if j > 1 { // Create parent - err = mkdirall(path[0:j-1], false) + err = mkdirall(path[0:j-1], false, sddl) if err != nil { return err } } // Parent now exists; invoke os.Mkdir or mkdirWithACL and use its result. - if adminAndLocalSystem { - err = mkdirWithACL(path) + if applyACL { + err = mkdirWithACL(path, sddl) } else { err = os.Mkdir(path, 0) } @@ -92,13 +98,12 @@ func mkdirall(path string, adminAndLocalSystem bool) error { // mkdirWithACL creates a new directory. If there is an error, it will be of // type *PathError. . // -// This is a modified and combined version of os.Mkdir and syscall.Mkdir +// This is a modified and combined version of os.Mkdir and windows.Mkdir // in golang to cater for creating a directory am ACL permitting full // access, with inheritance, to any subfolder/file for Built-in Administrators // and Local System. -func mkdirWithACL(name string) error { - sa := syscall.SecurityAttributes{Length: 0} - sddl := "D:P(A;OICI;GA;;;BA)(A;OICI;GA;;;SY)" +func mkdirWithACL(name string, sddl string) error { + sa := windows.SecurityAttributes{Length: 0} sd, err := winio.SddlToSecurityDescriptor(sddl) if err != nil { return &os.PathError{Op: "mkdir", Path: name, Err: err} @@ -107,12 +112,12 @@ func mkdirWithACL(name string) error { sa.InheritHandle = 1 sa.SecurityDescriptor = uintptr(unsafe.Pointer(&sd[0])) - namep, err := syscall.UTF16PtrFromString(name) + namep, err := windows.UTF16PtrFromString(name) if err != nil { return &os.PathError{Op: "mkdir", Path: name, Err: err} } - e := syscall.CreateDirectory(namep, &sa) + e := windows.CreateDirectory(namep, &sa) if e != nil { return &os.PathError{Op: "mkdir", Path: name, Err: e} } @@ -135,7 +140,7 @@ func IsAbs(path string) bool { return true } -// The origin of the functions below here are the golang OS and syscall packages, +// The origin of the functions below here are the golang OS and windows packages, // slightly modified to only cope with files, not directories due to the // specific use case. // @@ -167,74 +172,74 @@ func OpenFileSequential(name string, flag int, _ os.FileMode) (*os.File, error) if name == "" { return nil, &os.PathError{Op: "open", Path: name, Err: syscall.ENOENT} } - r, errf := syscallOpenFileSequential(name, flag, 0) + r, errf := windowsOpenFileSequential(name, flag, 0) if errf == nil { return r, nil } return nil, &os.PathError{Op: "open", Path: name, Err: errf} } -func syscallOpenFileSequential(name string, flag int, _ os.FileMode) (file *os.File, err error) { - r, e := syscallOpenSequential(name, flag|syscall.O_CLOEXEC, 0) +func windowsOpenFileSequential(name string, flag int, _ os.FileMode) (file *os.File, err error) { + r, e := windowsOpenSequential(name, flag|windows.O_CLOEXEC, 0) if e != nil { return nil, e } return os.NewFile(uintptr(r), name), nil } -func makeInheritSa() *syscall.SecurityAttributes { - var sa syscall.SecurityAttributes +func makeInheritSa() *windows.SecurityAttributes { + var sa windows.SecurityAttributes sa.Length = uint32(unsafe.Sizeof(sa)) sa.InheritHandle = 1 return &sa } -func syscallOpenSequential(path string, mode int, _ uint32) (fd syscall.Handle, err error) { +func windowsOpenSequential(path string, mode int, _ uint32) (fd windows.Handle, err error) { if len(path) == 0 { - return syscall.InvalidHandle, syscall.ERROR_FILE_NOT_FOUND + return windows.InvalidHandle, windows.ERROR_FILE_NOT_FOUND } - pathp, err := syscall.UTF16PtrFromString(path) + pathp, err := windows.UTF16PtrFromString(path) if err != nil { - return syscall.InvalidHandle, err + return windows.InvalidHandle, err } var access uint32 - switch mode & (syscall.O_RDONLY | syscall.O_WRONLY | syscall.O_RDWR) { - case syscall.O_RDONLY: - access = syscall.GENERIC_READ - case syscall.O_WRONLY: - access = syscall.GENERIC_WRITE - case syscall.O_RDWR: - access = syscall.GENERIC_READ | syscall.GENERIC_WRITE + switch mode & (windows.O_RDONLY | windows.O_WRONLY | windows.O_RDWR) { + case windows.O_RDONLY: + access = windows.GENERIC_READ + case windows.O_WRONLY: + access = windows.GENERIC_WRITE + case windows.O_RDWR: + access = windows.GENERIC_READ | windows.GENERIC_WRITE } - if mode&syscall.O_CREAT != 0 { - access |= syscall.GENERIC_WRITE + if mode&windows.O_CREAT != 0 { + access |= windows.GENERIC_WRITE } - if mode&syscall.O_APPEND != 0 { - access &^= syscall.GENERIC_WRITE - access |= syscall.FILE_APPEND_DATA + if mode&windows.O_APPEND != 0 { + access &^= windows.GENERIC_WRITE + access |= windows.FILE_APPEND_DATA } - sharemode := uint32(syscall.FILE_SHARE_READ | syscall.FILE_SHARE_WRITE) - var sa *syscall.SecurityAttributes - if mode&syscall.O_CLOEXEC == 0 { + sharemode := uint32(windows.FILE_SHARE_READ | windows.FILE_SHARE_WRITE) + var sa *windows.SecurityAttributes + if mode&windows.O_CLOEXEC == 0 { sa = makeInheritSa() } var createmode uint32 switch { - case mode&(syscall.O_CREAT|syscall.O_EXCL) == (syscall.O_CREAT | syscall.O_EXCL): - createmode = syscall.CREATE_NEW - case mode&(syscall.O_CREAT|syscall.O_TRUNC) == (syscall.O_CREAT | syscall.O_TRUNC): - createmode = syscall.CREATE_ALWAYS - case mode&syscall.O_CREAT == syscall.O_CREAT: - createmode = syscall.OPEN_ALWAYS - case mode&syscall.O_TRUNC == syscall.O_TRUNC: - createmode = syscall.TRUNCATE_EXISTING + case mode&(windows.O_CREAT|windows.O_EXCL) == (windows.O_CREAT | windows.O_EXCL): + createmode = windows.CREATE_NEW + case mode&(windows.O_CREAT|windows.O_TRUNC) == (windows.O_CREAT | windows.O_TRUNC): + createmode = windows.CREATE_ALWAYS + case mode&windows.O_CREAT == windows.O_CREAT: + createmode = windows.OPEN_ALWAYS + case mode&windows.O_TRUNC == windows.O_TRUNC: + createmode = windows.TRUNCATE_EXISTING default: - createmode = syscall.OPEN_EXISTING + createmode = windows.OPEN_EXISTING } // Use FILE_FLAG_SEQUENTIAL_SCAN rather than FILE_ATTRIBUTE_NORMAL as implemented in golang. //https://msdn.microsoft.com/en-us/library/windows/desktop/aa363858(v=vs.85).aspx const fileFlagSequentialScan = 0x08000000 // FILE_FLAG_SEQUENTIAL_SCAN - h, e := syscall.CreateFile(pathp, access, sharemode, sa, createmode, fileFlagSequentialScan, 0) + h, e := windows.CreateFile(pathp, access, sharemode, sa, createmode, fileFlagSequentialScan, 0) return h, e } diff --git a/vendor/github.com/docker/docker/pkg/system/init.go b/vendor/github.com/docker/docker/pkg/system/init.go new file mode 100644 index 000000000..a17597aab --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/system/init.go @@ -0,0 +1,22 @@ +package system // import "github.com/docker/docker/pkg/system" + +import ( + "syscall" + "time" + "unsafe" +) + +// Used by chtimes +var maxTime time.Time + +func init() { + // chtimes initialization + if unsafe.Sizeof(syscall.Timespec{}.Nsec) == 8 { + // This is a 64 bit timespec + // os.Chtimes limits time to the following + maxTime = time.Unix(0, 1<<63-1) + } else { + // This is a 32 bit timespec + maxTime = time.Unix(1<<31-1, 0) + } +} diff --git a/vendor/github.com/docker/docker/pkg/system/init_unix.go b/vendor/github.com/docker/docker/pkg/system/init_unix.go new file mode 100644 index 000000000..4996a67c1 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/system/init_unix.go @@ -0,0 +1,7 @@ +// +build !windows + +package system // import "github.com/docker/docker/pkg/system" + +// InitLCOW does nothing since LCOW is a windows only feature +func InitLCOW(experimental bool) { +} diff --git a/vendor/github.com/docker/docker/pkg/system/init_windows.go b/vendor/github.com/docker/docker/pkg/system/init_windows.go new file mode 100644 index 000000000..4910ff69d --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/system/init_windows.go @@ -0,0 +1,12 @@ +package system // import "github.com/docker/docker/pkg/system" + +// lcowSupported determines if Linux Containers on Windows are supported. +var lcowSupported = false + +// InitLCOW sets whether LCOW is supported or not +func InitLCOW(experimental bool) { + v := GetOSVersion() + if experimental && v.Build >= 16299 { + lcowSupported = true + } +} diff --git a/vendor/github.com/docker/docker/pkg/system/lcow.go b/vendor/github.com/docker/docker/pkg/system/lcow.go new file mode 100644 index 000000000..5be3e2182 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/system/lcow.go @@ -0,0 +1,32 @@ +package system // import "github.com/docker/docker/pkg/system" + +import ( + "runtime" + "strings" + + specs "github.com/opencontainers/image-spec/specs-go/v1" + "github.com/pkg/errors" +) + +// IsOSSupported determines if an operating system is supported by the host +func IsOSSupported(os string) bool { + if strings.EqualFold(runtime.GOOS, os) { + return true + } + if LCOWSupported() && strings.EqualFold(os, "linux") { + return true + } + return false +} + +// ValidatePlatform determines if a platform structure is valid. +// TODO This is a temporary windows-only function, should be replaced by +// comparison of worker capabilities +func ValidatePlatform(platform specs.Platform) error { + if runtime.GOOS == "windows" { + if !(platform.OS == runtime.GOOS || (LCOWSupported() && platform.OS == "linux")) { + return errors.Errorf("unsupported os %s", platform.OS) + } + } + return nil +} diff --git a/vendor/github.com/docker/docker/pkg/system/lcow_unix.go b/vendor/github.com/docker/docker/pkg/system/lcow_unix.go new file mode 100644 index 000000000..26397fb8a --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/system/lcow_unix.go @@ -0,0 +1,8 @@ +// +build !windows + +package system // import "github.com/docker/docker/pkg/system" + +// LCOWSupported returns true if Linux containers on Windows are supported. +func LCOWSupported() bool { + return false +} diff --git a/vendor/github.com/docker/docker/pkg/system/lcow_windows.go b/vendor/github.com/docker/docker/pkg/system/lcow_windows.go new file mode 100644 index 000000000..f0139df8f --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/system/lcow_windows.go @@ -0,0 +1,6 @@ +package system // import "github.com/docker/docker/pkg/system" + +// LCOWSupported returns true if Linux containers on Windows are supported. +func LCOWSupported() bool { + return lcowSupported +} diff --git a/vendor/github.com/docker/docker/pkg/system/lstat.go b/vendor/github.com/docker/docker/pkg/system/lstat_unix.go similarity index 84% rename from vendor/github.com/docker/docker/pkg/system/lstat.go rename to vendor/github.com/docker/docker/pkg/system/lstat_unix.go index bd23c4d50..7477995f1 100644 --- a/vendor/github.com/docker/docker/pkg/system/lstat.go +++ b/vendor/github.com/docker/docker/pkg/system/lstat_unix.go @@ -1,6 +1,6 @@ // +build !windows -package system +package system // import "github.com/docker/docker/pkg/system" import ( "syscall" diff --git a/vendor/github.com/docker/docker/pkg/system/lstat_windows.go b/vendor/github.com/docker/docker/pkg/system/lstat_windows.go index 49e87eb40..359c791d9 100644 --- a/vendor/github.com/docker/docker/pkg/system/lstat_windows.go +++ b/vendor/github.com/docker/docker/pkg/system/lstat_windows.go @@ -1,25 +1,14 @@ -// +build windows +package system // import "github.com/docker/docker/pkg/system" -package system - -import ( - "os" -) +import "os" // Lstat calls os.Lstat to get a fileinfo interface back. // This is then copied into our own locally defined structure. -// Note the Linux version uses fromStatT to do the copy back, -// but that not strictly necessary when already in an OS specific module. func Lstat(path string) (*StatT, error) { fi, err := os.Lstat(path) if err != nil { return nil, err } - return &StatT{ - name: fi.Name(), - size: fi.Size(), - mode: fi.Mode(), - modTime: fi.ModTime(), - isDir: fi.IsDir()}, nil + return fromStatT(&fi) } diff --git a/vendor/github.com/docker/docker/pkg/system/meminfo.go b/vendor/github.com/docker/docker/pkg/system/meminfo.go index 3b6e947e6..6667eb84d 100644 --- a/vendor/github.com/docker/docker/pkg/system/meminfo.go +++ b/vendor/github.com/docker/docker/pkg/system/meminfo.go @@ -1,4 +1,4 @@ -package system +package system // import "github.com/docker/docker/pkg/system" // MemInfo contains memory statistics of the host system. type MemInfo struct { diff --git a/vendor/github.com/docker/docker/pkg/system/meminfo_linux.go b/vendor/github.com/docker/docker/pkg/system/meminfo_linux.go index 385f1d5e7..d79e8b076 100644 --- a/vendor/github.com/docker/docker/pkg/system/meminfo_linux.go +++ b/vendor/github.com/docker/docker/pkg/system/meminfo_linux.go @@ -1,4 +1,4 @@ -package system +package system // import "github.com/docker/docker/pkg/system" import ( "bufio" diff --git a/vendor/github.com/docker/docker/pkg/system/meminfo_solaris.go b/vendor/github.com/docker/docker/pkg/system/meminfo_solaris.go deleted file mode 100644 index 925776e78..000000000 --- a/vendor/github.com/docker/docker/pkg/system/meminfo_solaris.go +++ /dev/null @@ -1,129 +0,0 @@ -// +build solaris,cgo - -package system - -import ( - "fmt" - "unsafe" -) - -// #cgo CFLAGS: -std=c99 -// #cgo LDFLAGS: -lkstat -// #include -// #include -// #include -// #include -// #include -// #include -// struct swaptable *allocSwaptable(int num) { -// struct swaptable *st; -// struct swapent *swapent; -// st = (struct swaptable *)malloc(num * sizeof(swapent_t) + sizeof (int)); -// swapent = st->swt_ent; -// for (int i = 0; i < num; i++,swapent++) { -// swapent->ste_path = (char *)malloc(MAXPATHLEN * sizeof (char)); -// } -// st->swt_n = num; -// return st; -//} -// void freeSwaptable (struct swaptable *st) { -// struct swapent *swapent = st->swt_ent; -// for (int i = 0; i < st->swt_n; i++,swapent++) { -// free(swapent->ste_path); -// } -// free(st); -// } -// swapent_t getSwapEnt(swapent_t *ent, int i) { -// return ent[i]; -// } -// int64_t getPpKernel() { -// int64_t pp_kernel = 0; -// kstat_ctl_t *ksc; -// kstat_t *ks; -// kstat_named_t *knp; -// kid_t kid; -// -// if ((ksc = kstat_open()) == NULL) { -// return -1; -// } -// if ((ks = kstat_lookup(ksc, "unix", 0, "system_pages")) == NULL) { -// return -1; -// } -// if (((kid = kstat_read(ksc, ks, NULL)) == -1) || -// ((knp = kstat_data_lookup(ks, "pp_kernel")) == NULL)) { -// return -1; -// } -// switch (knp->data_type) { -// case KSTAT_DATA_UINT64: -// pp_kernel = knp->value.ui64; -// break; -// case KSTAT_DATA_UINT32: -// pp_kernel = knp->value.ui32; -// break; -// } -// pp_kernel *= sysconf(_SC_PAGESIZE); -// return (pp_kernel > 0 ? pp_kernel : -1); -// } -import "C" - -// Get the system memory info using sysconf same as prtconf -func getTotalMem() int64 { - pagesize := C.sysconf(C._SC_PAGESIZE) - npages := C.sysconf(C._SC_PHYS_PAGES) - return int64(pagesize * npages) -} - -func getFreeMem() int64 { - pagesize := C.sysconf(C._SC_PAGESIZE) - npages := C.sysconf(C._SC_AVPHYS_PAGES) - return int64(pagesize * npages) -} - -// ReadMemInfo retrieves memory statistics of the host system and returns a -// MemInfo type. -func ReadMemInfo() (*MemInfo, error) { - - ppKernel := C.getPpKernel() - MemTotal := getTotalMem() - MemFree := getFreeMem() - SwapTotal, SwapFree, err := getSysSwap() - - if ppKernel < 0 || MemTotal < 0 || MemFree < 0 || SwapTotal < 0 || - SwapFree < 0 { - return nil, fmt.Errorf("error getting system memory info %v\n", err) - } - - meminfo := &MemInfo{} - // Total memory is total physical memory less than memory locked by kernel - meminfo.MemTotal = MemTotal - int64(ppKernel) - meminfo.MemFree = MemFree - meminfo.SwapTotal = SwapTotal - meminfo.SwapFree = SwapFree - - return meminfo, nil -} - -func getSysSwap() (int64, int64, error) { - var tSwap int64 - var fSwap int64 - var diskblksPerPage int64 - num, err := C.swapctl(C.SC_GETNSWP, nil) - if err != nil { - return -1, -1, err - } - st := C.allocSwaptable(num) - _, err = C.swapctl(C.SC_LIST, unsafe.Pointer(st)) - if err != nil { - C.freeSwaptable(st) - return -1, -1, err - } - - diskblksPerPage = int64(C.sysconf(C._SC_PAGESIZE) >> C.DEV_BSHIFT) - for i := 0; i < int(num); i++ { - swapent := C.getSwapEnt(&st.swt_ent[0], C.int(i)) - tSwap += int64(swapent.ste_pages) * diskblksPerPage - fSwap += int64(swapent.ste_free) * diskblksPerPage - } - C.freeSwaptable(st) - return tSwap, fSwap, nil -} diff --git a/vendor/github.com/docker/docker/pkg/system/meminfo_unsupported.go b/vendor/github.com/docker/docker/pkg/system/meminfo_unsupported.go index 3ce019dff..56f449426 100644 --- a/vendor/github.com/docker/docker/pkg/system/meminfo_unsupported.go +++ b/vendor/github.com/docker/docker/pkg/system/meminfo_unsupported.go @@ -1,6 +1,6 @@ -// +build !linux,!windows,!solaris +// +build !linux,!windows -package system +package system // import "github.com/docker/docker/pkg/system" // ReadMemInfo is not supported on platforms other than linux and windows. func ReadMemInfo() (*MemInfo, error) { diff --git a/vendor/github.com/docker/docker/pkg/system/meminfo_windows.go b/vendor/github.com/docker/docker/pkg/system/meminfo_windows.go index 883944a4c..6ed93f2fe 100644 --- a/vendor/github.com/docker/docker/pkg/system/meminfo_windows.go +++ b/vendor/github.com/docker/docker/pkg/system/meminfo_windows.go @@ -1,4 +1,4 @@ -package system +package system // import "github.com/docker/docker/pkg/system" import ( "unsafe" diff --git a/vendor/github.com/docker/docker/pkg/system/mknod.go b/vendor/github.com/docker/docker/pkg/system/mknod.go index 73958182b..b132482e0 100644 --- a/vendor/github.com/docker/docker/pkg/system/mknod.go +++ b/vendor/github.com/docker/docker/pkg/system/mknod.go @@ -1,15 +1,15 @@ // +build !windows -package system +package system // import "github.com/docker/docker/pkg/system" import ( - "syscall" + "golang.org/x/sys/unix" ) // Mknod creates a filesystem node (file, device special file or named pipe) named path // with attributes specified by mode and dev. func Mknod(path string, mode uint32, dev int) error { - return syscall.Mknod(path, mode, dev) + return unix.Mknod(path, mode, dev) } // Mkdev is used to build the value of linux devices (in /dev/) which specifies major @@ -18,5 +18,5 @@ func Mknod(path string, mode uint32, dev int) error { // They are, from low to high: the lower 8 bits of the minor, then 12 bits of the major, // then the top 12 bits of the minor. func Mkdev(major int64, minor int64) uint32 { - return uint32(((minor & 0xfff00) << 12) | ((major & 0xfff) << 8) | (minor & 0xff)) + return uint32(unix.Mkdev(uint32(major), uint32(minor))) } diff --git a/vendor/github.com/docker/docker/pkg/system/mknod_windows.go b/vendor/github.com/docker/docker/pkg/system/mknod_windows.go index 2e863c021..ec89d7a15 100644 --- a/vendor/github.com/docker/docker/pkg/system/mknod_windows.go +++ b/vendor/github.com/docker/docker/pkg/system/mknod_windows.go @@ -1,6 +1,4 @@ -// +build windows - -package system +package system // import "github.com/docker/docker/pkg/system" // Mknod is not implemented on Windows. func Mknod(path string, mode uint32, dev int) error { diff --git a/vendor/github.com/docker/docker/pkg/system/path.go b/vendor/github.com/docker/docker/pkg/system/path.go new file mode 100644 index 000000000..a3d957afa --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/system/path.go @@ -0,0 +1,60 @@ +package system // import "github.com/docker/docker/pkg/system" + +import ( + "fmt" + "path/filepath" + "runtime" + "strings" + + "github.com/containerd/continuity/pathdriver" +) + +const defaultUnixPathEnv = "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin" + +// DefaultPathEnv is unix style list of directories to search for +// executables. Each directory is separated from the next by a colon +// ':' character . +func DefaultPathEnv(os string) string { + if runtime.GOOS == "windows" { + if os != runtime.GOOS { + return defaultUnixPathEnv + } + // Deliberately empty on Windows containers on Windows as the default path will be set by + // the container. Docker has no context of what the default path should be. + return "" + } + return defaultUnixPathEnv + +} + +// CheckSystemDriveAndRemoveDriveLetter verifies that a path, if it includes a drive letter, +// is the system drive. +// On Linux: this is a no-op. +// On Windows: this does the following> +// CheckSystemDriveAndRemoveDriveLetter verifies and manipulates a Windows path. +// This is used, for example, when validating a user provided path in docker cp. +// If a drive letter is supplied, it must be the system drive. The drive letter +// is always removed. Also, it translates it to OS semantics (IOW / to \). We +// need the path in this syntax so that it can ultimately be concatenated with +// a Windows long-path which doesn't support drive-letters. Examples: +// C: --> Fail +// C:\ --> \ +// a --> a +// /a --> \a +// d:\ --> Fail +func CheckSystemDriveAndRemoveDriveLetter(path string, driver pathdriver.PathDriver) (string, error) { + if runtime.GOOS != "windows" || LCOWSupported() { + return path, nil + } + + if len(path) == 2 && string(path[1]) == ":" { + return "", fmt.Errorf("No relative path specified in %q", path) + } + if !driver.IsAbs(path) || len(path) < 2 { + return filepath.FromSlash(path), nil + } + if string(path[1]) == ":" && !strings.EqualFold(string(path[0]), "c") { + return "", fmt.Errorf("The specified path is not on the system drive (C:)") + } + return filepath.FromSlash(path[2:]), nil +} diff --git a/vendor/github.com/docker/docker/pkg/system/path_unix.go b/vendor/github.com/docker/docker/pkg/system/path_unix.go deleted file mode 100644 index c607c4db0..000000000 --- a/vendor/github.com/docker/docker/pkg/system/path_unix.go +++ /dev/null @@ -1,14 +0,0 @@ -// +build !windows - -package system - -// DefaultPathEnv is unix style list of directories to search for -// executables. Each directory is separated from the next by a colon -// ':' character . -const DefaultPathEnv = "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin" - -// CheckSystemDriveAndRemoveDriveLetter verifies that a path, if it includes a drive letter, -// is the system drive. This is a no-op on Linux. -func CheckSystemDriveAndRemoveDriveLetter(path string) (string, error) { - return path, nil -} diff --git a/vendor/github.com/docker/docker/pkg/system/path_windows.go b/vendor/github.com/docker/docker/pkg/system/path_windows.go deleted file mode 100644 index cbfe2c157..000000000 --- a/vendor/github.com/docker/docker/pkg/system/path_windows.go +++ /dev/null @@ -1,37 +0,0 @@ -// +build windows - -package system - -import ( - "fmt" - "path/filepath" - "strings" -) - -// DefaultPathEnv is deliberately empty on Windows as the default path will be set by -// the container. Docker has no context of what the default path should be. -const DefaultPathEnv = "" - -// CheckSystemDriveAndRemoveDriveLetter verifies and manipulates a Windows path. -// This is used, for example, when validating a user provided path in docker cp. -// If a drive letter is supplied, it must be the system drive. The drive letter -// is always removed. Also, it translates it to OS semantics (IOW / to \). We -// need the path in this syntax so that it can ultimately be contatenated with -// a Windows long-path which doesn't support drive-letters. Examples: -// C: --> Fail -// C:\ --> \ -// a --> a -// /a --> \a -// d:\ --> Fail -func CheckSystemDriveAndRemoveDriveLetter(path string) (string, error) { - if len(path) == 2 && string(path[1]) == ":" { - return "", fmt.Errorf("No relative path specified in %q", path) - } - if !filepath.IsAbs(path) || len(path) < 2 { - return filepath.FromSlash(path), nil - } - if string(path[1]) == ":" && !strings.EqualFold(string(path[0]), "c") { - return "", fmt.Errorf("The specified path is not on the system drive (C:)") - } - return filepath.FromSlash(path[2:]), nil -} diff --git a/vendor/github.com/docker/docker/pkg/system/process_unix.go b/vendor/github.com/docker/docker/pkg/system/process_unix.go index c99d796d1..0195a891b 100644 --- a/vendor/github.com/docker/docker/pkg/system/process_unix.go +++ b/vendor/github.com/docker/docker/pkg/system/process_unix.go @@ -1,15 +1,17 @@ -// +build linux freebsd solaris darwin +// +build linux freebsd darwin -package system +package system // import "github.com/docker/docker/pkg/system" import ( "syscall" + + "golang.org/x/sys/unix" ) // IsProcessAlive returns true if process with a given pid is running. func IsProcessAlive(pid int) bool { - err := syscall.Kill(pid, syscall.Signal(0)) - if err == nil || err == syscall.EPERM { + err := unix.Kill(pid, syscall.Signal(0)) + if err == nil || err == unix.EPERM { return true } @@ -18,5 +20,5 @@ func IsProcessAlive(pid int) bool { // KillProcess force-stops a process. func KillProcess(pid int) { - syscall.Kill(pid, syscall.SIGKILL) + unix.Kill(pid, unix.SIGKILL) } diff --git a/vendor/github.com/docker/docker/pkg/system/process_windows.go b/vendor/github.com/docker/docker/pkg/system/process_windows.go index e8113f97e..4e70c97b1 100644 --- a/vendor/github.com/docker/docker/pkg/system/process_windows.go +++ b/vendor/github.com/docker/docker/pkg/system/process_windows.go @@ -1,20 +1,18 @@ -package system +package system // import "github.com/docker/docker/pkg/system" + +import "os" // IsProcessAlive returns true if process with a given pid is running. func IsProcessAlive(pid int) bool { - // TODO Windows containerd. Not sure this is needed - // p, err := os.FindProcess(pid) - // if err == nil { - // return true - // } - return false + _, err := os.FindProcess(pid) + + return err == nil } // KillProcess force-stops a process. func KillProcess(pid int) { - // TODO Windows containerd. Not sure this is needed - // p, err := os.FindProcess(pid) - // if err == nil { - // p.Kill() - // } + p, err := os.FindProcess(pid) + if err == nil { + p.Kill() + } } diff --git a/vendor/github.com/docker/docker/pkg/system/rm.go b/vendor/github.com/docker/docker/pkg/system/rm.go new file mode 100644 index 000000000..02e4d2622 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/system/rm.go @@ -0,0 +1,80 @@ +package system // import "github.com/docker/docker/pkg/system" + +import ( + "os" + "syscall" + "time" + + "github.com/docker/docker/pkg/mount" + "github.com/pkg/errors" +) + +// EnsureRemoveAll wraps `os.RemoveAll` to check for specific errors that can +// often be remedied. +// Only use `EnsureRemoveAll` if you really want to make every effort to remove +// a directory. +// +// Because of the way `os.Remove` (and by extension `os.RemoveAll`) works, there +// can be a race between reading directory entries and then actually attempting +// to remove everything in the directory. +// These types of errors do not need to be returned since it's ok for the dir to +// be gone we can just retry the remove operation. +// +// This should not return a `os.ErrNotExist` kind of error under any circumstances +func EnsureRemoveAll(dir string) error { + notExistErr := make(map[string]bool) + + // track retries + exitOnErr := make(map[string]int) + maxRetry := 50 + + // Attempt to unmount anything beneath this dir first + mount.RecursiveUnmount(dir) + + for { + err := os.RemoveAll(dir) + if err == nil { + return err + } + + pe, ok := err.(*os.PathError) + if !ok { + return err + } + + if os.IsNotExist(err) { + if notExistErr[pe.Path] { + return err + } + notExistErr[pe.Path] = true + + // There is a race where some subdir can be removed but after the parent + // dir entries have been read. + // So the path could be from `os.Remove(subdir)` + // If the reported non-existent path is not the passed in `dir` we + // should just retry, but otherwise return with no error. + if pe.Path == dir { + return nil + } + continue + } + + if pe.Err != syscall.EBUSY { + return err + } + + if mounted, _ := mount.Mounted(pe.Path); mounted { + if e := mount.Unmount(pe.Path); e != nil { + if mounted, _ := mount.Mounted(pe.Path); mounted { + return errors.Wrapf(e, "error while removing %s", dir) + } + } + } + + if exitOnErr[pe.Path] == maxRetry { + return err + } + exitOnErr[pe.Path]++ + time.Sleep(100 * time.Millisecond) + } +} diff --git a/vendor/github.com/docker/docker/pkg/system/stat_darwin.go b/vendor/github.com/docker/docker/pkg/system/stat_darwin.go index f0742f59e..c1c0ee9f3 100644 --- a/vendor/github.com/docker/docker/pkg/system/stat_darwin.go +++ b/vendor/github.com/docker/docker/pkg/system/stat_darwin.go @@ -1,10 +1,8 @@ -package system +package system // import "github.com/docker/docker/pkg/system" -import ( - "syscall" -) +import "syscall" -// fromStatT creates a system.StatT type from a syscall.Stat_t type +// fromStatT converts a syscall.Stat_t type to a system.Stat_t type func fromStatT(s *syscall.Stat_t) (*StatT, error) { return &StatT{size: s.Size, mode: uint32(s.Mode), @@ -13,20 +11,3 @@ func fromStatT(s *syscall.Stat_t) (*StatT, error) { rdev: uint64(s.Rdev), mtim: s.Mtimespec}, nil } - -// FromStatT loads a system.StatT from a syscall.Stat_t. -func FromStatT(s *syscall.Stat_t) (*StatT, error) { - return fromStatT(s) -} - -// Stat takes a path to a file and returns -// a system.StatT type pertaining to that file. -// -// Throws an error if the file does not exist -func Stat(path string) (*StatT, error) { - s := &syscall.Stat_t{} - if err := syscall.Stat(path, s); err != nil { - return nil, err - } - return fromStatT(s) -} diff --git a/vendor/github.com/docker/docker/pkg/system/stat_freebsd.go b/vendor/github.com/docker/docker/pkg/system/stat_freebsd.go index d0fb6f151..c1c0ee9f3 100644 --- a/vendor/github.com/docker/docker/pkg/system/stat_freebsd.go +++ b/vendor/github.com/docker/docker/pkg/system/stat_freebsd.go @@ -1,8 +1,6 @@ -package system +package system // import "github.com/docker/docker/pkg/system" -import ( - "syscall" -) +import "syscall" // fromStatT converts a syscall.Stat_t type to a system.Stat_t type func fromStatT(s *syscall.Stat_t) (*StatT, error) { @@ -13,15 +11,3 @@ func fromStatT(s *syscall.Stat_t) (*StatT, error) { rdev: uint64(s.Rdev), mtim: s.Mtimespec}, nil } - -// Stat takes a path to a file and returns -// a system.Stat_t type pertaining to that file. -// -// Throws an error if the file does not exist -func Stat(path string) (*StatT, error) { - s := &syscall.Stat_t{} - if err := syscall.Stat(path, s); err != nil { - return nil, err - } - return fromStatT(s) -} diff --git a/vendor/github.com/docker/docker/pkg/system/stat_linux.go b/vendor/github.com/docker/docker/pkg/system/stat_linux.go index 8b1eded13..98c9eb18d 100644 --- a/vendor/github.com/docker/docker/pkg/system/stat_linux.go +++ b/vendor/github.com/docker/docker/pkg/system/stat_linux.go @@ -1,8 +1,6 @@ -package system +package system // import "github.com/docker/docker/pkg/system" -import ( - "syscall" -) +import "syscall" // fromStatT converts a syscall.Stat_t type to a system.Stat_t type func fromStatT(s *syscall.Stat_t) (*StatT, error) { @@ -14,20 +12,8 @@ func fromStatT(s *syscall.Stat_t) (*StatT, error) { mtim: s.Mtim}, nil } -// FromStatT exists only on linux, and loads a system.StatT from a -// syscal.Stat_t. +// FromStatT converts a syscall.Stat_t type to a system.Stat_t type +// This is exposed on Linux as pkg/archive/changes uses it. func FromStatT(s *syscall.Stat_t) (*StatT, error) { return fromStatT(s) } - -// Stat takes a path to a file and returns -// a system.StatT type pertaining to that file. -// -// Throws an error if the file does not exist -func Stat(path string) (*StatT, error) { - s := &syscall.Stat_t{} - if err := syscall.Stat(path, s); err != nil { - return nil, err - } - return fromStatT(s) -} diff --git a/vendor/github.com/docker/docker/pkg/system/stat_openbsd.go b/vendor/github.com/docker/docker/pkg/system/stat_openbsd.go index 3c3b71fb2..756b92d1e 100644 --- a/vendor/github.com/docker/docker/pkg/system/stat_openbsd.go +++ b/vendor/github.com/docker/docker/pkg/system/stat_openbsd.go @@ -1,10 +1,8 @@ -package system +package system // import "github.com/docker/docker/pkg/system" -import ( - "syscall" -) +import "syscall" -// fromStatT creates a system.StatT type from a syscall.Stat_t type +// fromStatT converts a syscall.Stat_t type to a system.Stat_t type func fromStatT(s *syscall.Stat_t) (*StatT, error) { return &StatT{size: s.Size, mode: uint32(s.Mode), diff --git a/vendor/github.com/docker/docker/pkg/system/stat_solaris.go b/vendor/github.com/docker/docker/pkg/system/stat_solaris.go index 0216985a2..756b92d1e 100644 --- a/vendor/github.com/docker/docker/pkg/system/stat_solaris.go +++ b/vendor/github.com/docker/docker/pkg/system/stat_solaris.go @@ -1,12 +1,8 @@ -// +build solaris +package system // import "github.com/docker/docker/pkg/system" -package system +import "syscall" -import ( - "syscall" -) - -// fromStatT creates a system.StatT type from a syscall.Stat_t type +// fromStatT converts a syscall.Stat_t type to a system.Stat_t type func fromStatT(s *syscall.Stat_t) (*StatT, error) { return &StatT{size: s.Size, mode: uint32(s.Mode), @@ -15,20 +11,3 @@ func fromStatT(s *syscall.Stat_t) (*StatT, error) { rdev: uint64(s.Rdev), mtim: s.Mtim}, nil } - -// FromStatT loads a system.StatT from a syscal.Stat_t. -func FromStatT(s *syscall.Stat_t) (*StatT, error) { - return fromStatT(s) -} - -// Stat takes a path to a file and returns -// a system.StatT type pertaining to that file. -// -// Throws an error if the file does not exist -func Stat(path string) (*StatT, error) { - s := &syscall.Stat_t{} - if err := syscall.Stat(path, s); err != nil { - return nil, err - } - return fromStatT(s) -} diff --git a/vendor/github.com/docker/docker/pkg/system/stat.go b/vendor/github.com/docker/docker/pkg/system/stat_unix.go similarity index 63% rename from vendor/github.com/docker/docker/pkg/system/stat.go rename to vendor/github.com/docker/docker/pkg/system/stat_unix.go index 087034c5e..3d7e2ebbe 100644 --- a/vendor/github.com/docker/docker/pkg/system/stat.go +++ b/vendor/github.com/docker/docker/pkg/system/stat_unix.go @@ -1,6 +1,6 @@ // +build !windows -package system +package system // import "github.com/docker/docker/pkg/system" import ( "syscall" @@ -47,7 +47,19 @@ func (s StatT) Mtim() syscall.Timespec { return s.mtim } -// GetLastModification returns file's last modification time. -func (s StatT) GetLastModification() syscall.Timespec { - return s.Mtim() +// IsDir reports whether s describes a directory. +func (s StatT) IsDir() bool { + return s.mode&syscall.S_IFDIR != 0 +} + +// Stat takes a path to a file and returns +// a system.StatT type pertaining to that file. +// +// Throws an error if the file does not exist +func Stat(path string) (*StatT, error) { + s := &syscall.Stat_t{} + if err := syscall.Stat(path, s); err != nil { + return nil, err + } + return fromStatT(s) } diff --git a/vendor/github.com/docker/docker/pkg/system/stat_unsupported.go b/vendor/github.com/docker/docker/pkg/system/stat_unsupported.go deleted file mode 100644 index 5d85f523c..000000000 --- a/vendor/github.com/docker/docker/pkg/system/stat_unsupported.go +++ /dev/null @@ -1,17 +0,0 @@ -// +build !linux,!windows,!freebsd,!solaris,!openbsd,!darwin - -package system - -import ( - "syscall" -) - -// fromStatT creates a system.StatT type from a syscall.Stat_t type -func fromStatT(s *syscall.Stat_t) (*StatT, error) { - return &StatT{size: s.Size, - mode: uint32(s.Mode), - uid: s.Uid, - gid: s.Gid, - rdev: uint64(s.Rdev), - mtim: s.Mtimespec}, nil -} diff --git a/vendor/github.com/docker/docker/pkg/system/stat_windows.go b/vendor/github.com/docker/docker/pkg/system/stat_windows.go index 39490c625..b2456cb88 100644 --- a/vendor/github.com/docker/docker/pkg/system/stat_windows.go +++ b/vendor/github.com/docker/docker/pkg/system/stat_windows.go @@ -1,6 +1,4 @@ -// +build windows - -package system +package system // import "github.com/docker/docker/pkg/system" import ( "os" @@ -8,18 +6,11 @@ import ( ) // StatT type contains status of a file. It contains metadata -// like name, permission, size, etc about a file. +// like permission, size, etc about a file. type StatT struct { - name string - size int64 - mode os.FileMode - modTime time.Time - isDir bool -} - -// Name returns file's name. -func (s StatT) Name() string { - return s.name + mode os.FileMode + size int64 + mtim time.Time } // Size returns file's size. @@ -29,15 +20,30 @@ func (s StatT) Size() int64 { // Mode returns file's permission mode. func (s StatT) Mode() os.FileMode { - return s.mode + return os.FileMode(s.mode) } -// ModTime returns file's last modification time. -func (s StatT) ModTime() time.Time { - return s.modTime +// Mtim returns file's last modification time. +func (s StatT) Mtim() time.Time { + return time.Time(s.mtim) } -// IsDir returns whether file is actually a directory. -func (s StatT) IsDir() bool { - return s.isDir +// Stat takes a path to a file and returns +// a system.StatT type pertaining to that file. +// +// Throws an error if the file does not exist +func Stat(path string) (*StatT, error) { + fi, err := os.Stat(path) + if err != nil { + return nil, err + } + return fromStatT(&fi) +} + +// fromStatT converts a os.FileInfo type to a system.StatT type +func fromStatT(fi *os.FileInfo) (*StatT, error) { + return &StatT{ + size: (*fi).Size(), + mode: (*fi).Mode(), + mtim: (*fi).ModTime()}, nil } diff --git a/vendor/github.com/docker/docker/pkg/system/syscall_unix.go b/vendor/github.com/docker/docker/pkg/system/syscall_unix.go index 3ae912846..919a412a7 100644 --- a/vendor/github.com/docker/docker/pkg/system/syscall_unix.go +++ b/vendor/github.com/docker/docker/pkg/system/syscall_unix.go @@ -1,13 +1,13 @@ // +build linux freebsd -package system +package system // import "github.com/docker/docker/pkg/system" -import "syscall" +import "golang.org/x/sys/unix" // Unmount is a platform-specific helper function to call // the unmount syscall. func Unmount(dest string) error { - return syscall.Unmount(dest, 0) + return unix.Unmount(dest, 0) } // CommandLineToArgv should not be used on Unix. diff --git a/vendor/github.com/docker/docker/pkg/system/syscall_windows.go b/vendor/github.com/docker/docker/pkg/system/syscall_windows.go index 1f311874f..ee7e0256f 100644 --- a/vendor/github.com/docker/docker/pkg/system/syscall_windows.go +++ b/vendor/github.com/docker/docker/pkg/system/syscall_windows.go @@ -1,15 +1,17 @@ -package system +package system // import "github.com/docker/docker/pkg/system" import ( - "syscall" + "fmt" "unsafe" - "github.com/Sirupsen/logrus" + "github.com/sirupsen/logrus" + "golang.org/x/sys/windows" ) var ( - ntuserApiset = syscall.NewLazyDLL("ext-ms-win-ntuser-window-l1-1-0") - procGetVersionExW = modkernel32.NewProc("GetVersionExW") + ntuserApiset = windows.NewLazyDLL("ext-ms-win-ntuser-window-l1-1-0") + procGetVersionExW = modkernel32.NewProc("GetVersionExW") + procGetProductInfo = modkernel32.NewProc("GetProductInfo") ) // OSVersion is a wrapper for Windows version information @@ -41,7 +43,7 @@ type osVersionInfoEx struct { func GetOSVersion() OSVersion { var err error osv := OSVersion{} - osv.Version, err = syscall.GetVersion() + osv.Version, err = windows.GetVersion() if err != nil { // GetVersion never fails. panic(err) @@ -52,6 +54,10 @@ func GetOSVersion() OSVersion { return osv } +func (osv OSVersion) ToString() string { + return fmt.Sprintf("%d.%d.%d", osv.MajorVersion, osv.MinorVersion, osv.Build) +} + // IsWindowsClient returns true if the SKU is client // @engine maintainers - this function should not be removed or modified as it // is used to enforce licensing restrictions on Windows. @@ -66,6 +72,22 @@ func IsWindowsClient() bool { return osviex.ProductType == verNTWorkstation } +// IsIoTCore returns true if the currently running image is based off of +// Windows 10 IoT Core. +// @engine maintainers - this function should not be removed or modified as it +// is used to enforce licensing restrictions on Windows. +func IsIoTCore() bool { + var returnedProductType uint32 + r1, _, err := procGetProductInfo.Call(6, 1, 0, 0, uintptr(unsafe.Pointer(&returnedProductType))) + if r1 == 0 { + logrus.Warnf("GetProductInfo failed - assuming this is not IoT: %v", err) + return false + } + const productIoTUAP = 0x0000007B + const productIoTUAPCommercial = 0x00000083 + return returnedProductType == productIoTUAP || returnedProductType == productIoTUAPCommercial +} + // Unmount is a platform-specific helper function to call // the unmount syscall. Not supported on Windows func Unmount(dest string) error { @@ -76,20 +98,20 @@ func Unmount(dest string) error { func CommandLineToArgv(commandLine string) ([]string, error) { var argc int32 - argsPtr, err := syscall.UTF16PtrFromString(commandLine) + argsPtr, err := windows.UTF16PtrFromString(commandLine) if err != nil { return nil, err } - argv, err := syscall.CommandLineToArgv(argsPtr, &argc) + argv, err := windows.CommandLineToArgv(argsPtr, &argc) if err != nil { return nil, err } - defer syscall.LocalFree(syscall.Handle(uintptr(unsafe.Pointer(argv)))) + defer windows.LocalFree(windows.Handle(uintptr(unsafe.Pointer(argv)))) newArgs := make([]string, argc) for i, v := range (*argv)[:argc] { - newArgs[i] = string(syscall.UTF16ToString((*v)[:])) + newArgs[i] = string(windows.UTF16ToString((*v)[:])) } return newArgs, nil diff --git a/vendor/github.com/docker/docker/pkg/system/umask.go b/vendor/github.com/docker/docker/pkg/system/umask.go index 3d0146b01..9912a2bab 100644 --- a/vendor/github.com/docker/docker/pkg/system/umask.go +++ b/vendor/github.com/docker/docker/pkg/system/umask.go @@ -1,13 +1,13 @@ // +build !windows -package system +package system // import "github.com/docker/docker/pkg/system" import ( - "syscall" + "golang.org/x/sys/unix" ) // Umask sets current process's file mode creation mask to newmask // and returns oldmask. func Umask(newmask int) (oldmask int, err error) { - return syscall.Umask(newmask), nil + return unix.Umask(newmask), nil } diff --git a/vendor/github.com/docker/docker/pkg/system/umask_windows.go b/vendor/github.com/docker/docker/pkg/system/umask_windows.go index 13f1de176..fc62388c3 100644 --- a/vendor/github.com/docker/docker/pkg/system/umask_windows.go +++ b/vendor/github.com/docker/docker/pkg/system/umask_windows.go @@ -1,6 +1,4 @@ -// +build windows - -package system +package system // import "github.com/docker/docker/pkg/system" // Umask is not supported on the windows platform. func Umask(newmask int) (oldmask int, err error) { diff --git a/vendor/github.com/docker/docker/pkg/system/utimes_freebsd.go b/vendor/github.com/docker/docker/pkg/system/utimes_freebsd.go index e2eac3b55..ed1b9fad5 100644 --- a/vendor/github.com/docker/docker/pkg/system/utimes_freebsd.go +++ b/vendor/github.com/docker/docker/pkg/system/utimes_freebsd.go @@ -1,20 +1,22 @@ -package system +package system // import "github.com/docker/docker/pkg/system" import ( "syscall" "unsafe" + + "golang.org/x/sys/unix" ) // LUtimesNano is used to change access and modification time of the specified path. -// It's used for symbol link file because syscall.UtimesNano doesn't support a NOFOLLOW flag atm. +// It's used for symbol link file because unix.UtimesNano doesn't support a NOFOLLOW flag atm. func LUtimesNano(path string, ts []syscall.Timespec) error { var _path *byte - _path, err := syscall.BytePtrFromString(path) + _path, err := unix.BytePtrFromString(path) if err != nil { return err } - if _, _, err := syscall.Syscall(syscall.SYS_LUTIMES, uintptr(unsafe.Pointer(_path)), uintptr(unsafe.Pointer(&ts[0])), 0); err != 0 && err != syscall.ENOSYS { + if _, _, err := unix.Syscall(unix.SYS_LUTIMES, uintptr(unsafe.Pointer(_path)), uintptr(unsafe.Pointer(&ts[0])), 0); err != 0 && err != unix.ENOSYS { return err } diff --git a/vendor/github.com/docker/docker/pkg/system/utimes_linux.go b/vendor/github.com/docker/docker/pkg/system/utimes_linux.go index fc8a1aba9..0afe85458 100644 --- a/vendor/github.com/docker/docker/pkg/system/utimes_linux.go +++ b/vendor/github.com/docker/docker/pkg/system/utimes_linux.go @@ -1,24 +1,23 @@ -package system +package system // import "github.com/docker/docker/pkg/system" import ( "syscall" "unsafe" + + "golang.org/x/sys/unix" ) // LUtimesNano is used to change access and modification time of the specified path. -// It's used for symbol link file because syscall.UtimesNano doesn't support a NOFOLLOW flag atm. +// It's used for symbol link file because unix.UtimesNano doesn't support a NOFOLLOW flag atm. func LUtimesNano(path string, ts []syscall.Timespec) error { - // These are not currently available in syscall - atFdCwd := -100 - atSymLinkNoFollow := 0x100 + atFdCwd := unix.AT_FDCWD var _path *byte - _path, err := syscall.BytePtrFromString(path) + _path, err := unix.BytePtrFromString(path) if err != nil { return err } - - if _, _, err := syscall.Syscall6(syscall.SYS_UTIMENSAT, uintptr(atFdCwd), uintptr(unsafe.Pointer(_path)), uintptr(unsafe.Pointer(&ts[0])), uintptr(atSymLinkNoFollow), 0, 0); err != 0 && err != syscall.ENOSYS { + if _, _, err := unix.Syscall6(unix.SYS_UTIMENSAT, uintptr(atFdCwd), uintptr(unsafe.Pointer(_path)), uintptr(unsafe.Pointer(&ts[0])), unix.AT_SYMLINK_NOFOLLOW, 0, 0); err != 0 && err != unix.ENOSYS { return err } diff --git a/vendor/github.com/docker/docker/pkg/system/utimes_unsupported.go b/vendor/github.com/docker/docker/pkg/system/utimes_unsupported.go index 139714544..095e072e1 100644 --- a/vendor/github.com/docker/docker/pkg/system/utimes_unsupported.go +++ b/vendor/github.com/docker/docker/pkg/system/utimes_unsupported.go @@ -1,6 +1,6 @@ // +build !linux,!freebsd -package system +package system // import "github.com/docker/docker/pkg/system" import "syscall" diff --git a/vendor/github.com/docker/docker/pkg/system/xattrs_linux.go b/vendor/github.com/docker/docker/pkg/system/xattrs_linux.go index d2e2c0579..66d4895b2 100644 --- a/vendor/github.com/docker/docker/pkg/system/xattrs_linux.go +++ b/vendor/github.com/docker/docker/pkg/system/xattrs_linux.go @@ -1,63 +1,29 @@ -package system +package system // import "github.com/docker/docker/pkg/system" -import ( - "syscall" - "unsafe" -) +import "golang.org/x/sys/unix" // Lgetxattr retrieves the value of the extended attribute identified by attr // and associated with the given path in the file system. // It will returns a nil slice and nil error if the xattr is not set. func Lgetxattr(path string, attr string) ([]byte, error) { - pathBytes, err := syscall.BytePtrFromString(path) - if err != nil { - return nil, err - } - attrBytes, err := syscall.BytePtrFromString(attr) - if err != nil { - return nil, err - } - dest := make([]byte, 128) - destBytes := unsafe.Pointer(&dest[0]) - sz, _, errno := syscall.Syscall6(syscall.SYS_LGETXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(attrBytes)), uintptr(destBytes), uintptr(len(dest)), 0, 0) - if errno == syscall.ENODATA { + sz, errno := unix.Lgetxattr(path, attr, dest) + if errno == unix.ENODATA { return nil, nil } - if errno == syscall.ERANGE { + if errno == unix.ERANGE { dest = make([]byte, sz) - destBytes := unsafe.Pointer(&dest[0]) - sz, _, errno = syscall.Syscall6(syscall.SYS_LGETXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(attrBytes)), uintptr(destBytes), uintptr(len(dest)), 0, 0) + sz, errno = unix.Lgetxattr(path, attr, dest) } - if errno != 0 { + if errno != nil { return nil, errno } return dest[:sz], nil } -var _zero uintptr - // Lsetxattr sets the value of the extended attribute identified by attr // and associated with the given path in the file system. func Lsetxattr(path string, attr string, data []byte, flags int) error { - pathBytes, err := syscall.BytePtrFromString(path) - if err != nil { - return err - } - attrBytes, err := syscall.BytePtrFromString(attr) - if err != nil { - return err - } - var dataBytes unsafe.Pointer - if len(data) > 0 { - dataBytes = unsafe.Pointer(&data[0]) - } else { - dataBytes = unsafe.Pointer(&_zero) - } - _, _, errno := syscall.Syscall6(syscall.SYS_LSETXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(attrBytes)), uintptr(dataBytes), uintptr(len(data)), uintptr(flags), 0) - if errno != 0 { - return errno - } - return nil + return unix.Lsetxattr(path, attr, data, flags) } diff --git a/vendor/github.com/docker/docker/pkg/system/xattrs_unsupported.go b/vendor/github.com/docker/docker/pkg/system/xattrs_unsupported.go index 0114f2227..d780a90cd 100644 --- a/vendor/github.com/docker/docker/pkg/system/xattrs_unsupported.go +++ b/vendor/github.com/docker/docker/pkg/system/xattrs_unsupported.go @@ -1,6 +1,6 @@ // +build !linux -package system +package system // import "github.com/docker/docker/pkg/system" // Lgetxattr is not supported on platforms other than linux. func Lgetxattr(path string, attr string) ([]byte, error) { diff --git a/vendor/github.com/docker/docker/pkg/tarsum/builder_context.go b/vendor/github.com/docker/docker/pkg/tarsum/builder_context.go new file mode 100644 index 000000000..bc7d84df4 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/tarsum/builder_context.go @@ -0,0 +1,21 @@ +package tarsum // import "github.com/docker/docker/pkg/tarsum" + +// BuilderContext is an interface extending TarSum by adding the Remove method. +// In general there was concern about adding this method to TarSum itself +// so instead it is being added just to "BuilderContext" which will then +// only be used during the .dockerignore file processing +// - see builder/evaluator.go +type BuilderContext interface { + TarSum + Remove(string) +} + +func (bc *tarSum) Remove(filename string) { + for i, fis := range bc.sums { + if fis.Name() == filename { + bc.sums = append(bc.sums[:i], bc.sums[i+1:]...) + // Note, we don't just return because there could be + // more than one with this name + } + } +} diff --git a/vendor/github.com/docker/docker/pkg/tarsum/fileinfosums.go b/vendor/github.com/docker/docker/pkg/tarsum/fileinfosums.go new file mode 100644 index 000000000..01d4ed59b --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/tarsum/fileinfosums.go @@ -0,0 +1,133 @@ +package tarsum // import "github.com/docker/docker/pkg/tarsum" + +import ( + "runtime" + "sort" + "strings" +) + +// FileInfoSumInterface provides an interface for accessing file checksum +// information within a tar file. This info is accessed through interface +// so the actual name and sum cannot be melded with. +type FileInfoSumInterface interface { + // File name + Name() string + // Checksum of this particular file and its headers + Sum() string + // Position of file in the tar + Pos() int64 +} + +type fileInfoSum struct { + name string + sum string + pos int64 +} + +func (fis fileInfoSum) Name() string { + return fis.name +} +func (fis fileInfoSum) Sum() string { + return fis.sum +} +func (fis fileInfoSum) Pos() int64 { + return fis.pos +} + +// FileInfoSums provides a list of FileInfoSumInterfaces. +type FileInfoSums []FileInfoSumInterface + +// GetFile returns the first FileInfoSumInterface with a matching name. +func (fis FileInfoSums) GetFile(name string) FileInfoSumInterface { + // We do case insensitive matching on Windows as c:\APP and c:\app are + // the same. See issue #33107. + for i := range fis { + if (runtime.GOOS == "windows" && strings.EqualFold(fis[i].Name(), name)) || + (runtime.GOOS != "windows" && fis[i].Name() == name) { + return fis[i] + } + } + return nil +} + +// GetAllFile returns a FileInfoSums with all matching names. +func (fis FileInfoSums) GetAllFile(name string) FileInfoSums { + f := FileInfoSums{} + for i := range fis { + if fis[i].Name() == name { + f = append(f, fis[i]) + } + } + return f +} + +// GetDuplicatePaths returns a FileInfoSums with all duplicated paths. +func (fis FileInfoSums) GetDuplicatePaths() (dups FileInfoSums) { + seen := make(map[string]int, len(fis)) // allocate earl. no need to grow this map. + for i := range fis { + f := fis[i] + if _, ok := seen[f.Name()]; ok { + dups = append(dups, f) + } else { + seen[f.Name()] = 0 + } + } + return dups +} + +// Len returns the size of the FileInfoSums. +func (fis FileInfoSums) Len() int { return len(fis) } + +// Swap swaps two FileInfoSum values if a FileInfoSums list. +func (fis FileInfoSums) Swap(i, j int) { fis[i], fis[j] = fis[j], fis[i] } + +// SortByPos sorts FileInfoSums content by position. +func (fis FileInfoSums) SortByPos() { + sort.Sort(byPos{fis}) +} + +// SortByNames sorts FileInfoSums content by name. +func (fis FileInfoSums) SortByNames() { + sort.Sort(byName{fis}) +} + +// SortBySums sorts FileInfoSums content by sums. +func (fis FileInfoSums) SortBySums() { + dups := fis.GetDuplicatePaths() + if len(dups) > 0 { + sort.Sort(bySum{fis, dups}) + } else { + sort.Sort(bySum{fis, nil}) + } +} + +// byName is a sort.Sort helper for sorting by file names. +// If names are the same, order them by their appearance in the tar archive +type byName struct{ FileInfoSums } + +func (bn byName) Less(i, j int) bool { + if bn.FileInfoSums[i].Name() == bn.FileInfoSums[j].Name() { + return bn.FileInfoSums[i].Pos() < bn.FileInfoSums[j].Pos() + } + return bn.FileInfoSums[i].Name() < bn.FileInfoSums[j].Name() +} + +// bySum is a sort.Sort helper for sorting by the sums of all the fileinfos in the tar archive +type bySum struct { + FileInfoSums + dups FileInfoSums +} + +func (bs bySum) Less(i, j int) bool { + if bs.dups != nil && bs.FileInfoSums[i].Name() == bs.FileInfoSums[j].Name() { + return bs.FileInfoSums[i].Pos() < bs.FileInfoSums[j].Pos() + } + return bs.FileInfoSums[i].Sum() < bs.FileInfoSums[j].Sum() +} + +// byPos is a sort.Sort helper for sorting by the sums of all the fileinfos by their original order +type byPos struct{ FileInfoSums } + +func (bp byPos) Less(i, j int) bool { + return bp.FileInfoSums[i].Pos() < bp.FileInfoSums[j].Pos() +} diff --git a/vendor/github.com/docker/docker/pkg/tarsum/tarsum.go b/vendor/github.com/docker/docker/pkg/tarsum/tarsum.go new file mode 100644 index 000000000..5542e1b2c --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/tarsum/tarsum.go @@ -0,0 +1,301 @@ +// Package tarsum provides algorithms to perform checksum calculation on +// filesystem layers. +// +// The transportation of filesystems, regarding Docker, is done with tar(1) +// archives. There are a variety of tar serialization formats [2], and a key +// concern here is ensuring a repeatable checksum given a set of inputs from a +// generic tar archive. Types of transportation include distribution to and from a +// registry endpoint, saving and loading through commands or Docker daemon APIs, +// transferring the build context from client to Docker daemon, and committing the +// filesystem of a container to become an image. +// +// As tar archives are used for transit, but not preserved in many situations, the +// focus of the algorithm is to ensure the integrity of the preserved filesystem, +// while maintaining a deterministic accountability. This includes neither +// constraining the ordering or manipulation of the files during the creation or +// unpacking of the archive, nor include additional metadata state about the file +// system attributes. +package tarsum // import "github.com/docker/docker/pkg/tarsum" + +import ( + "archive/tar" + "bytes" + "compress/gzip" + "crypto" + "crypto/sha256" + "encoding/hex" + "errors" + "fmt" + "hash" + "io" + "path" + "strings" +) + +const ( + buf8K = 8 * 1024 + buf16K = 16 * 1024 + buf32K = 32 * 1024 +) + +// NewTarSum creates a new interface for calculating a fixed time checksum of a +// tar archive. +// +// This is used for calculating checksums of layers of an image, in some cases +// including the byte payload of the image's json metadata as well, and for +// calculating the checksums for buildcache. +func NewTarSum(r io.Reader, dc bool, v Version) (TarSum, error) { + return NewTarSumHash(r, dc, v, DefaultTHash) +} + +// NewTarSumHash creates a new TarSum, providing a THash to use rather than +// the DefaultTHash. +func NewTarSumHash(r io.Reader, dc bool, v Version, tHash THash) (TarSum, error) { + headerSelector, err := getTarHeaderSelector(v) + if err != nil { + return nil, err + } + ts := &tarSum{Reader: r, DisableCompression: dc, tarSumVersion: v, headerSelector: headerSelector, tHash: tHash} + err = ts.initTarSum() + return ts, err +} + +// NewTarSumForLabel creates a new TarSum using the provided TarSum version+hash label. +func NewTarSumForLabel(r io.Reader, disableCompression bool, label string) (TarSum, error) { + parts := strings.SplitN(label, "+", 2) + if len(parts) != 2 { + return nil, errors.New("tarsum label string should be of the form: {tarsum_version}+{hash_name}") + } + + versionName, hashName := parts[0], parts[1] + + version, ok := tarSumVersionsByName[versionName] + if !ok { + return nil, fmt.Errorf("unknown TarSum version name: %q", versionName) + } + + hashConfig, ok := standardHashConfigs[hashName] + if !ok { + return nil, fmt.Errorf("unknown TarSum hash name: %q", hashName) + } + + tHash := NewTHash(hashConfig.name, hashConfig.hash.New) + + return NewTarSumHash(r, disableCompression, version, tHash) +} + +// TarSum is the generic interface for calculating fixed time +// checksums of a tar archive. +type TarSum interface { + io.Reader + GetSums() FileInfoSums + Sum([]byte) string + Version() Version + Hash() THash +} + +// tarSum struct is the structure for a Version0 checksum calculation. +type tarSum struct { + io.Reader + tarR *tar.Reader + tarW *tar.Writer + writer writeCloseFlusher + bufTar *bytes.Buffer + bufWriter *bytes.Buffer + bufData []byte + h hash.Hash + tHash THash + sums FileInfoSums + fileCounter int64 + currentFile string + finished bool + first bool + DisableCompression bool // false by default. When false, the output gzip compressed. + tarSumVersion Version // this field is not exported so it can not be mutated during use + headerSelector tarHeaderSelector // handles selecting and ordering headers for files in the archive +} + +func (ts tarSum) Hash() THash { + return ts.tHash +} + +func (ts tarSum) Version() Version { + return ts.tarSumVersion +} + +// THash provides a hash.Hash type generator and its name. +type THash interface { + Hash() hash.Hash + Name() string +} + +// NewTHash is a convenience method for creating a THash. +func NewTHash(name string, h func() hash.Hash) THash { + return simpleTHash{n: name, h: h} +} + +type tHashConfig struct { + name string + hash crypto.Hash +} + +var ( + // NOTE: DO NOT include MD5 or SHA1, which are considered insecure. + standardHashConfigs = map[string]tHashConfig{ + "sha256": {name: "sha256", hash: crypto.SHA256}, + "sha512": {name: "sha512", hash: crypto.SHA512}, + } +) + +// DefaultTHash is default TarSum hashing algorithm - "sha256". +var DefaultTHash = NewTHash("sha256", sha256.New) + +type simpleTHash struct { + n string + h func() hash.Hash +} + +func (sth simpleTHash) Name() string { return sth.n } +func (sth simpleTHash) Hash() hash.Hash { return sth.h() } + +func (ts *tarSum) encodeHeader(h *tar.Header) error { + for _, elem := range ts.headerSelector.selectHeaders(h) { + // Ignore these headers to be compatible with versions + // before go 1.10 + if elem[0] == "gname" || elem[0] == "uname" { + elem[1] = "" + } + if _, err := ts.h.Write([]byte(elem[0] + elem[1])); err != nil { + return err + } + } + return nil +} + +func (ts *tarSum) initTarSum() error { + ts.bufTar = bytes.NewBuffer([]byte{}) + ts.bufWriter = bytes.NewBuffer([]byte{}) + ts.tarR = tar.NewReader(ts.Reader) + ts.tarW = tar.NewWriter(ts.bufTar) + if !ts.DisableCompression { + ts.writer = gzip.NewWriter(ts.bufWriter) + } else { + ts.writer = &nopCloseFlusher{Writer: ts.bufWriter} + } + if ts.tHash == nil { + ts.tHash = DefaultTHash + } + ts.h = ts.tHash.Hash() + ts.h.Reset() + ts.first = true + ts.sums = FileInfoSums{} + return nil +} + +func (ts *tarSum) Read(buf []byte) (int, error) { + if ts.finished { + return ts.bufWriter.Read(buf) + } + if len(ts.bufData) < len(buf) { + switch { + case len(buf) <= buf8K: + ts.bufData = make([]byte, buf8K) + case len(buf) <= buf16K: + ts.bufData = make([]byte, buf16K) + case len(buf) <= buf32K: + ts.bufData = make([]byte, buf32K) + default: + ts.bufData = make([]byte, len(buf)) + } + } + buf2 := ts.bufData[:len(buf)] + + n, err := ts.tarR.Read(buf2) + if err != nil { + if err == io.EOF { + if _, err := ts.h.Write(buf2[:n]); err != nil { + return 0, err + } + if !ts.first { + ts.sums = append(ts.sums, fileInfoSum{name: ts.currentFile, sum: hex.EncodeToString(ts.h.Sum(nil)), pos: ts.fileCounter}) + ts.fileCounter++ + ts.h.Reset() + } else { + ts.first = false + } + + if _, err := ts.tarW.Write(buf2[:n]); err != nil { + return 0, err + } + + currentHeader, err := ts.tarR.Next() + if err != nil { + if err == io.EOF { + if err := ts.tarW.Close(); err != nil { + return 0, err + } + if _, err := io.Copy(ts.writer, ts.bufTar); err != nil { + return 0, err + } + if err := ts.writer.Close(); err != nil { + return 0, err + } + ts.finished = true + return ts.bufWriter.Read(buf) + } + return 0, err + } + + ts.currentFile = path.Join(".", path.Join("/", currentHeader.Name)) + if err := ts.encodeHeader(currentHeader); err != nil { + return 0, err + } + if err := ts.tarW.WriteHeader(currentHeader); err != nil { + return 0, err + } + + if _, err := io.Copy(ts.writer, ts.bufTar); err != nil { + return 0, err + } + ts.writer.Flush() + + return ts.bufWriter.Read(buf) + } + return 0, err + } + + // Filling the hash buffer + if _, err = ts.h.Write(buf2[:n]); err != nil { + return 0, err + } + + // Filling the tar writer + if _, err = ts.tarW.Write(buf2[:n]); err != nil { + return 0, err + } + + // Filling the output writer + if _, err = io.Copy(ts.writer, ts.bufTar); err != nil { + return 0, err + } + ts.writer.Flush() + + return ts.bufWriter.Read(buf) +} + +func (ts *tarSum) Sum(extra []byte) string { + ts.sums.SortBySums() + h := ts.tHash.Hash() + if extra != nil { + h.Write(extra) + } + for _, fis := range ts.sums { + h.Write([]byte(fis.Sum())) + } + checksum := ts.Version().String() + "+" + ts.tHash.Name() + ":" + hex.EncodeToString(h.Sum(nil)) + return checksum +} + +func (ts *tarSum) GetSums() FileInfoSums { + return ts.sums +} diff --git a/vendor/github.com/docker/docker/pkg/tarsum/tarsum_spec.md b/vendor/github.com/docker/docker/pkg/tarsum/tarsum_spec.md new file mode 100644 index 000000000..e7ea2abd0 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/tarsum/tarsum_spec.md @@ -0,0 +1,229 @@ +page_title: TarSum checksum specification +page_description: Documentation for algorithms used in the TarSum checksum calculation +page_keywords: docker, checksum, validation, tarsum + +# TarSum Checksum Specification + +## Abstract + +This document describes the algorithms used in performing the TarSum checksum +calculation on filesystem layers, the need for this method over existing +methods, and the versioning of this calculation. + +## Warning + +This checksum algorithm is for best-effort comparison of file trees with fuzzy logic. + +This is _not_ a cryptographic attestation, and should not be considered secure. + +## Introduction + +The transportation of filesystems, regarding Docker, is done with tar(1) +archives. There are a variety of tar serialization formats [2], and a key +concern here is ensuring a repeatable checksum given a set of inputs from a +generic tar archive. Types of transportation include distribution to and from a +registry endpoint, saving and loading through commands or Docker daemon APIs, +transferring the build context from client to Docker daemon, and committing the +filesystem of a container to become an image. + +As tar archives are used for transit, but not preserved in many situations, the +focus of the algorithm is to ensure the integrity of the preserved filesystem, +while maintaining a deterministic accountability. This includes neither +constraining the ordering or manipulation of the files during the creation or +unpacking of the archive, nor include additional metadata state about the file +system attributes. + +## Intended Audience + +This document is outlining the methods used for consistent checksum calculation +for filesystems transported via tar archives. + +Auditing these methodologies is an open and iterative process. This document +should accommodate the review of source code. Ultimately, this document should +be the starting point of further refinements to the algorithm and its future +versions. + +## Concept + +The checksum mechanism must ensure the integrity and assurance of the +filesystem payload. + +## Checksum Algorithm Profile + +A checksum mechanism must define the following operations and attributes: + +* Associated hashing cipher - used to checksum each file payload and attribute + information. +* Checksum list - each file of the filesystem archive has its checksum + calculated from the payload and attributes of the file. The final checksum is + calculated from this list, with specific ordering. +* Version - as the algorithm adapts to requirements, there are behaviors of the + algorithm to manage by versioning. +* Archive being calculated - the tar archive having its checksum calculated + +## Elements of TarSum checksum + +The calculated sum output is a text string. The elements included in the output +of the calculated sum comprise the information needed for validation of the sum +(TarSum version and hashing cipher used) and the expected checksum in hexadecimal +form. + +There are two delimiters used: +* '+' separates TarSum version from hashing cipher +* ':' separates calculation mechanics from expected hash + +Example: + +``` + "tarsum.v1+sha256:220a60ecd4a3c32c282622a625a54db9ba0ff55b5ba9c29c7064a2bc358b6a3e" + | | \ | + | | \ | + |_version_|_cipher__|__ | + | \ | + |_calculation_mechanics_|______________________expected_sum_______________________| +``` + +## Versioning + +Versioning was introduced [0] to accommodate differences in calculation needed, +and ability to maintain reverse compatibility. + +The general algorithm will be describe further in the 'Calculation'. + +### Version0 + +This is the initial version of TarSum. + +Its element in the TarSum checksum string is `tarsum`. + +### Version1 + +Its element in the TarSum checksum is `tarsum.v1`. + +The notable changes in this version: +* Exclusion of file `mtime` from the file information headers, in each file + checksum calculation +* Inclusion of extended attributes (`xattrs`. Also seen as `SCHILY.xattr.` prefixed Pax + tar file info headers) keys and values in each file checksum calculation + +### VersionDev + +*Do not use unless validating refinements to the checksum algorithm* + +Its element in the TarSum checksum is `tarsum.dev`. + +This is a floating place holder for a next version and grounds for testing +changes. The methods used for calculation are subject to change without notice, +and this version is for testing and not for production use. + +## Ciphers + +The official default and standard hashing cipher used in the calculation mechanic +is `sha256`. This refers to SHA256 hash algorithm as defined in FIPS 180-4. + +Though the TarSum algorithm itself is not exclusively bound to the single +hashing cipher `sha256`, support for alternate hashing ciphers was later added +[1]. Use cases for alternate cipher could include future-proofing TarSum +checksum format and using faster cipher hashes for tar filesystem checksums. + +## Calculation + +### Requirement + +As mentioned earlier, the calculation is such that it takes into consideration +the lifecycle of the tar archive. In that the tar archive is not an immutable, +permanent artifact. Otherwise options like relying on a known hashing cipher +checksum of the archive itself would be reliable enough. The tar archive of the +filesystem is used as a transportation medium for Docker images, and the +archive is discarded once its contents are extracted. Therefore, for consistent +validation items such as order of files in the tar archive and time stamps are +subject to change once an image is received. + +### Process + +The method is typically iterative due to reading tar info headers from the +archive stream, though this is not a strict requirement. + +#### Files + +Each file in the tar archive have their contents (headers and body) checksummed +individually using the designated associated hashing cipher. The ordered +headers of the file are written to the checksum calculation first, and then the +payload of the file body. + +The resulting checksum of the file is appended to the list of file sums. The +sum is encoded as a string of the hexadecimal digest. Additionally, the file +name and position in the archive is kept as reference for special ordering. + +#### Headers + +The following headers are read, in this +order ( and the corresponding representation of its value): +* 'name' - string +* 'mode' - string of the base10 integer +* 'uid' - string of the integer +* 'gid' - string of the integer +* 'size' - string of the integer +* 'mtime' (_Version0 only_) - string of integer of the seconds since 1970-01-01 00:00:00 UTC +* 'typeflag' - string of the char +* 'linkname' - string +* 'uname' - string +* 'gname' - string +* 'devmajor' - string of the integer +* 'devminor' - string of the integer + +For >= Version1, the extended attribute headers ("SCHILY.xattr." prefixed pax +headers) included after the above list. These xattrs key/values are first +sorted by the keys. + +#### Header Format + +The ordered headers are written to the hash in the format of + + "{.key}{.value}" + +with no newline. + +#### Body + +After the order headers of the file have been added to the checksum for the +file, the body of the file is written to the hash. + +#### List of file sums + +The list of file sums is sorted by the string of the hexadecimal digest. + +If there are two files in the tar with matching paths, the order of occurrence +for that path is reflected for the sums of the corresponding file header and +body. + +#### Final Checksum + +Begin with a fresh or initial state of the associated hash cipher. If there is +additional payload to include in the TarSum calculation for the archive, it is +written first. Then each checksum from the ordered list of file sums is written +to the hash. + +The resulting digest is formatted per the Elements of TarSum checksum, +including the TarSum version, the associated hash cipher and the hexadecimal +encoded checksum digest. + +## Security Considerations + +The initial version of TarSum has undergone one update that could invalidate +handcrafted tar archives. The tar archive format supports appending of files +with same names as prior files in the archive. The latter file will clobber the +prior file of the same path. Due to this the algorithm now accounts for files +with matching paths, and orders the list of file sums accordingly [3]. + +## Footnotes + +* [0] Versioning https://github.com/docker/docker/commit/747f89cd327db9d50251b17797c4d825162226d0 +* [1] Alternate ciphers https://github.com/docker/docker/commit/4e9925d780665149b8bc940d5ba242ada1973c4e +* [2] Tar http://en.wikipedia.org/wiki/Tar_%28computing%29 +* [3] Name collision https://github.com/docker/docker/commit/c5e6362c53cbbc09ddbabd5a7323e04438b57d31 + +## Acknowledgments + +Joffrey F (shin-) and Guillaume J. Charmes (creack) on the initial work of the +TarSum calculation. diff --git a/vendor/github.com/docker/docker/pkg/tarsum/versioning.go b/vendor/github.com/docker/docker/pkg/tarsum/versioning.go new file mode 100644 index 000000000..aa1f17186 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/tarsum/versioning.go @@ -0,0 +1,158 @@ +package tarsum // import "github.com/docker/docker/pkg/tarsum" + +import ( + "archive/tar" + "errors" + "io" + "sort" + "strconv" + "strings" +) + +// Version is used for versioning of the TarSum algorithm +// based on the prefix of the hash used +// i.e. "tarsum+sha256:e58fcf7418d4390dec8e8fb69d88c06ec07039d651fedd3aa72af9972e7d046b" +type Version int + +// Prefix of "tarsum" +const ( + Version0 Version = iota + Version1 + // VersionDev this constant will be either the latest or an unsettled next-version of the TarSum calculation + VersionDev +) + +// WriteV1Header writes a tar header to a writer in V1 tarsum format. +func WriteV1Header(h *tar.Header, w io.Writer) { + for _, elem := range v1TarHeaderSelect(h) { + w.Write([]byte(elem[0] + elem[1])) + } +} + +// VersionLabelForChecksum returns the label for the given tarsum +// checksum, i.e., everything before the first `+` character in +// the string or an empty string if no label separator is found. +func VersionLabelForChecksum(checksum string) string { + // Checksums are in the form: {versionLabel}+{hashID}:{hex} + sepIndex := strings.Index(checksum, "+") + if sepIndex < 0 { + return "" + } + return checksum[:sepIndex] +} + +// GetVersions gets a list of all known tarsum versions. +func GetVersions() []Version { + v := []Version{} + for k := range tarSumVersions { + v = append(v, k) + } + return v +} + +var ( + tarSumVersions = map[Version]string{ + Version0: "tarsum", + Version1: "tarsum.v1", + VersionDev: "tarsum.dev", + } + tarSumVersionsByName = map[string]Version{ + "tarsum": Version0, + "tarsum.v1": Version1, + "tarsum.dev": VersionDev, + } +) + +func (tsv Version) String() string { + return tarSumVersions[tsv] +} + +// GetVersionFromTarsum returns the Version from the provided string. +func GetVersionFromTarsum(tarsum string) (Version, error) { + tsv := tarsum + if strings.Contains(tarsum, "+") { + tsv = strings.SplitN(tarsum, "+", 2)[0] + } + for v, s := range tarSumVersions { + if s == tsv { + return v, nil + } + } + return -1, ErrNotVersion +} + +// Errors that may be returned by functions in this package +var ( + ErrNotVersion = errors.New("string does not include a TarSum Version") + ErrVersionNotImplemented = errors.New("TarSum Version is not yet implemented") +) + +// tarHeaderSelector is the interface which different versions +// of tarsum should use for selecting and ordering tar headers +// for each item in the archive. +type tarHeaderSelector interface { + selectHeaders(h *tar.Header) (orderedHeaders [][2]string) +} + +type tarHeaderSelectFunc func(h *tar.Header) (orderedHeaders [][2]string) + +func (f tarHeaderSelectFunc) selectHeaders(h *tar.Header) (orderedHeaders [][2]string) { + return f(h) +} + +func v0TarHeaderSelect(h *tar.Header) (orderedHeaders [][2]string) { + return [][2]string{ + {"name", h.Name}, + {"mode", strconv.FormatInt(h.Mode, 10)}, + {"uid", strconv.Itoa(h.Uid)}, + {"gid", strconv.Itoa(h.Gid)}, + {"size", strconv.FormatInt(h.Size, 10)}, + {"mtime", strconv.FormatInt(h.ModTime.UTC().Unix(), 10)}, + {"typeflag", string([]byte{h.Typeflag})}, + {"linkname", h.Linkname}, + {"uname", h.Uname}, + {"gname", h.Gname}, + {"devmajor", strconv.FormatInt(h.Devmajor, 10)}, + {"devminor", strconv.FormatInt(h.Devminor, 10)}, + } +} + +func v1TarHeaderSelect(h *tar.Header) (orderedHeaders [][2]string) { + // Get extended attributes. + xAttrKeys := make([]string, len(h.Xattrs)) + for k := range h.Xattrs { + xAttrKeys = append(xAttrKeys, k) + } + sort.Strings(xAttrKeys) + + // Make the slice with enough capacity to hold the 11 basic headers + // we want from the v0 selector plus however many xattrs we have. + orderedHeaders = make([][2]string, 0, 11+len(xAttrKeys)) + + // Copy all headers from v0 excluding the 'mtime' header (the 5th element). + v0headers := v0TarHeaderSelect(h) + orderedHeaders = append(orderedHeaders, v0headers[0:5]...) + orderedHeaders = append(orderedHeaders, v0headers[6:]...) + + // Finally, append the sorted xattrs. + for _, k := range xAttrKeys { + orderedHeaders = append(orderedHeaders, [2]string{k, h.Xattrs[k]}) + } + + return +} + +var registeredHeaderSelectors = map[Version]tarHeaderSelectFunc{ + Version0: v0TarHeaderSelect, + Version1: v1TarHeaderSelect, + VersionDev: v1TarHeaderSelect, +} + +func getTarHeaderSelector(v Version) (tarHeaderSelector, error) { + headerSelector, ok := registeredHeaderSelectors[v] + if !ok { + return nil, ErrVersionNotImplemented + } + + return headerSelector, nil +} diff --git a/vendor/github.com/docker/docker/pkg/tarsum/writercloser.go b/vendor/github.com/docker/docker/pkg/tarsum/writercloser.go new file mode 100644 index 000000000..c4c45a35e --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/tarsum/writercloser.go @@ -0,0 +1,22 @@ +package tarsum // import "github.com/docker/docker/pkg/tarsum" + +import ( + "io" +) + +type writeCloseFlusher interface { + io.WriteCloser + Flush() error +} + +type nopCloseFlusher struct { + io.Writer +} + +func (n *nopCloseFlusher) Close() error { + return nil +} + +func (n *nopCloseFlusher) Flush() error { + return nil +} diff --git a/vendor/github.com/docker/docker/pkg/term/ascii.go b/vendor/github.com/docker/docker/pkg/term/ascii.go index f5262bccf..87bca8d4a 100644 --- a/vendor/github.com/docker/docker/pkg/term/ascii.go +++ b/vendor/github.com/docker/docker/pkg/term/ascii.go @@ -1,4 +1,4 @@ -package term +package term // import "github.com/docker/docker/pkg/term" import ( "fmt" @@ -59,7 +59,7 @@ next: return nil, fmt.Errorf("Unknown character: '%s'", key) } } else { - codes = append(codes, byte(key[0])) + codes = append(codes, key[0]) } } return codes, nil diff --git a/vendor/github.com/docker/docker/pkg/term/proxy.go b/vendor/github.com/docker/docker/pkg/term/proxy.go new file mode 100644 index 000000000..da733e584 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/term/proxy.go @@ -0,0 +1,78 @@ +package term // import "github.com/docker/docker/pkg/term" + +import ( + "io" +) + +// EscapeError is special error which returned by a TTY proxy reader's Read() +// method in case its detach escape sequence is read. +type EscapeError struct{} + +func (EscapeError) Error() string { + return "read escape sequence" +} + +// escapeProxy is used only for attaches with a TTY. It is used to proxy +// stdin keypresses from the underlying reader and look for the passed in +// escape key sequence to signal a detach. +type escapeProxy struct { + escapeKeys []byte + escapeKeyPos int + r io.Reader +} + +// NewEscapeProxy returns a new TTY proxy reader which wraps the given reader +// and detects when the specified escape keys are read, in which case the Read +// method will return an error of type EscapeError. +func NewEscapeProxy(r io.Reader, escapeKeys []byte) io.Reader { + return &escapeProxy{ + escapeKeys: escapeKeys, + r: r, + } +} + +func (r *escapeProxy) Read(buf []byte) (int, error) { + nr, err := r.r.Read(buf) + + if len(r.escapeKeys) == 0 { + return nr, err + } + + preserve := func() { + // this preserves the original key presses in the passed in buffer + nr += r.escapeKeyPos + preserve := make([]byte, 0, r.escapeKeyPos+len(buf)) + preserve = append(preserve, r.escapeKeys[:r.escapeKeyPos]...) + preserve = append(preserve, buf...) + r.escapeKeyPos = 0 + copy(buf[0:nr], preserve) + } + + if nr != 1 || err != nil { + if r.escapeKeyPos > 0 { + preserve() + } + return nr, err + } + + if buf[0] != r.escapeKeys[r.escapeKeyPos] { + if r.escapeKeyPos > 0 { + preserve() + } + return nr, nil + } + + if r.escapeKeyPos == len(r.escapeKeys)-1 { + return 0, EscapeError{} + } + + // Looks like we've got an escape key, but we need to match again on the next + // read. + // Store the current escape key we found so we can look for the next one on + // the next read. + // Since this is an escape key, make sure we don't let the caller read it + // If later on we find that this is not the escape sequence, we'll add the + // keys back + r.escapeKeyPos++ + return nr - r.escapeKeyPos, nil +} diff --git a/vendor/github.com/docker/docker/pkg/term/tc.go b/vendor/github.com/docker/docker/pkg/term/tc.go new file mode 100644 index 000000000..01bcaa8ab --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/term/tc.go @@ -0,0 +1,20 @@ +// +build !windows + +package term // import "github.com/docker/docker/pkg/term" + +import ( + "syscall" + "unsafe" + + "golang.org/x/sys/unix" +) + +func tcget(fd uintptr, p *Termios) syscall.Errno { + _, _, err := unix.Syscall(unix.SYS_IOCTL, fd, uintptr(getTermios), uintptr(unsafe.Pointer(p))) + return err +} + +func tcset(fd uintptr, p *Termios) syscall.Errno { + _, _, err := unix.Syscall(unix.SYS_IOCTL, fd, setTermios, uintptr(unsafe.Pointer(p))) + return err +} diff --git a/vendor/github.com/docker/docker/pkg/term/tc_linux_cgo.go b/vendor/github.com/docker/docker/pkg/term/tc_linux_cgo.go deleted file mode 100644 index 59dac5ba8..000000000 --- a/vendor/github.com/docker/docker/pkg/term/tc_linux_cgo.go +++ /dev/null @@ -1,50 +0,0 @@ -// +build linux,cgo - -package term - -import ( - "syscall" - "unsafe" -) - -// #include -import "C" - -// Termios is the Unix API for terminal I/O. -// It is passthrough for syscall.Termios in order to make it portable with -// other platforms where it is not available or handled differently. -type Termios syscall.Termios - -// MakeRaw put the terminal connected to the given file descriptor into raw -// mode and returns the previous state of the terminal so that it can be -// restored. -func MakeRaw(fd uintptr) (*State, error) { - var oldState State - if err := tcget(fd, &oldState.termios); err != 0 { - return nil, err - } - - newState := oldState.termios - - C.cfmakeraw((*C.struct_termios)(unsafe.Pointer(&newState))) - if err := tcset(fd, &newState); err != 0 { - return nil, err - } - return &oldState, nil -} - -func tcget(fd uintptr, p *Termios) syscall.Errno { - ret, err := C.tcgetattr(C.int(fd), (*C.struct_termios)(unsafe.Pointer(p))) - if ret != 0 { - return err.(syscall.Errno) - } - return 0 -} - -func tcset(fd uintptr, p *Termios) syscall.Errno { - ret, err := C.tcsetattr(C.int(fd), C.TCSANOW, (*C.struct_termios)(unsafe.Pointer(p))) - if ret != 0 { - return err.(syscall.Errno) - } - return 0 -} diff --git a/vendor/github.com/docker/docker/pkg/term/tc_other.go b/vendor/github.com/docker/docker/pkg/term/tc_other.go deleted file mode 100644 index 750d7c3f6..000000000 --- a/vendor/github.com/docker/docker/pkg/term/tc_other.go +++ /dev/null @@ -1,20 +0,0 @@ -// +build !windows -// +build !linux !cgo -// +build !solaris !cgo - -package term - -import ( - "syscall" - "unsafe" -) - -func tcget(fd uintptr, p *Termios) syscall.Errno { - _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, uintptr(getTermios), uintptr(unsafe.Pointer(p))) - return err -} - -func tcset(fd uintptr, p *Termios) syscall.Errno { - _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, setTermios, uintptr(unsafe.Pointer(p))) - return err -} diff --git a/vendor/github.com/docker/docker/pkg/term/tc_solaris_cgo.go b/vendor/github.com/docker/docker/pkg/term/tc_solaris_cgo.go deleted file mode 100644 index c9139d0ca..000000000 --- a/vendor/github.com/docker/docker/pkg/term/tc_solaris_cgo.go +++ /dev/null @@ -1,63 +0,0 @@ -// +build solaris,cgo - -package term - -import ( - "syscall" - "unsafe" -) - -// #include -import "C" - -// Termios is the Unix API for terminal I/O. -// It is passthrough for syscall.Termios in order to make it portable with -// other platforms where it is not available or handled differently. -type Termios syscall.Termios - -// MakeRaw put the terminal connected to the given file descriptor into raw -// mode and returns the previous state of the terminal so that it can be -// restored. -func MakeRaw(fd uintptr) (*State, error) { - var oldState State - if err := tcget(fd, &oldState.termios); err != 0 { - return nil, err - } - - newState := oldState.termios - - newState.Iflag &^= (syscall.IGNBRK | syscall.BRKINT | syscall.PARMRK | syscall.ISTRIP | syscall.INLCR | syscall.IGNCR | syscall.ICRNL | syscall.IXON | syscall.IXANY) - newState.Oflag &^= syscall.OPOST - newState.Lflag &^= (syscall.ECHO | syscall.ECHONL | syscall.ICANON | syscall.ISIG | syscall.IEXTEN) - newState.Cflag &^= (syscall.CSIZE | syscall.PARENB) - newState.Cflag |= syscall.CS8 - - /* - VMIN is the minimum number of characters that needs to be read in non-canonical mode for it to be returned - Since VMIN is overloaded with another element in canonical mode when we switch modes it defaults to 4. It - needs to be explicitly set to 1. - */ - newState.Cc[C.VMIN] = 1 - newState.Cc[C.VTIME] = 0 - - if err := tcset(fd, &newState); err != 0 { - return nil, err - } - return &oldState, nil -} - -func tcget(fd uintptr, p *Termios) syscall.Errno { - ret, err := C.tcgetattr(C.int(fd), (*C.struct_termios)(unsafe.Pointer(p))) - if ret != 0 { - return err.(syscall.Errno) - } - return 0 -} - -func tcset(fd uintptr, p *Termios) syscall.Errno { - ret, err := C.tcsetattr(C.int(fd), C.TCSANOW, (*C.struct_termios)(unsafe.Pointer(p))) - if ret != 0 { - return err.(syscall.Errno) - } - return 0 -} diff --git a/vendor/github.com/docker/docker/pkg/term/term.go b/vendor/github.com/docker/docker/pkg/term/term.go index 816f8d752..0589a9551 100644 --- a/vendor/github.com/docker/docker/pkg/term/term.go +++ b/vendor/github.com/docker/docker/pkg/term/term.go @@ -2,7 +2,7 @@ // Package term provides structures and helper functions to work with // terminal (state, sizes). -package term +package term // import "github.com/docker/docker/pkg/term" import ( "errors" @@ -10,7 +10,8 @@ import ( "io" "os" "os/signal" - "syscall" + + "golang.org/x/sys/unix" ) var ( @@ -79,7 +80,7 @@ func SaveState(fd uintptr) (*State, error) { // descriptor, with echo disabled. func DisableEcho(fd uintptr, state *State) error { newState := state.termios - newState.Lflag &^= syscall.ECHO + newState.Lflag &^= unix.ECHO if err := tcset(fd, &newState); err != 0 { return err diff --git a/vendor/github.com/docker/docker/pkg/term/term_solaris.go b/vendor/github.com/docker/docker/pkg/term/term_solaris.go deleted file mode 100644 index 112debbec..000000000 --- a/vendor/github.com/docker/docker/pkg/term/term_solaris.go +++ /dev/null @@ -1,41 +0,0 @@ -// +build solaris - -package term - -import ( - "syscall" - "unsafe" -) - -/* -#include -#include -#include - -// Small wrapper to get rid of variadic args of ioctl() -int my_ioctl(int fd, int cmd, struct winsize *ws) { - return ioctl(fd, cmd, ws); -} -*/ -import "C" - -// GetWinsize returns the window size based on the specified file descriptor. -func GetWinsize(fd uintptr) (*Winsize, error) { - ws := &Winsize{} - ret, err := C.my_ioctl(C.int(fd), C.int(syscall.TIOCGWINSZ), (*C.struct_winsize)(unsafe.Pointer(ws))) - // Skip retval = 0 - if ret == 0 { - return ws, nil - } - return ws, err -} - -// SetWinsize tries to set the specified window size for the specified file descriptor. -func SetWinsize(fd uintptr, ws *Winsize) error { - ret, err := C.my_ioctl(C.int(fd), C.int(syscall.TIOCSWINSZ), (*C.struct_winsize)(unsafe.Pointer(ws))) - // Skip retval = 0 - if ret == 0 { - return nil - } - return err -} diff --git a/vendor/github.com/docker/docker/pkg/term/term_unix.go b/vendor/github.com/docker/docker/pkg/term/term_unix.go deleted file mode 100644 index ddf87a0e5..000000000 --- a/vendor/github.com/docker/docker/pkg/term/term_unix.go +++ /dev/null @@ -1,29 +0,0 @@ -// +build !solaris,!windows - -package term - -import ( - "syscall" - "unsafe" -) - -// GetWinsize returns the window size based on the specified file descriptor. -func GetWinsize(fd uintptr) (*Winsize, error) { - ws := &Winsize{} - _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, uintptr(syscall.TIOCGWINSZ), uintptr(unsafe.Pointer(ws))) - // Skipp errno = 0 - if err == 0 { - return ws, nil - } - return ws, err -} - -// SetWinsize tries to set the specified window size for the specified file descriptor. -func SetWinsize(fd uintptr, ws *Winsize) error { - _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, uintptr(syscall.TIOCSWINSZ), uintptr(unsafe.Pointer(ws))) - // Skipp errno = 0 - if err == 0 { - return nil - } - return err -} diff --git a/vendor/github.com/docker/docker/pkg/term/term_windows.go b/vendor/github.com/docker/docker/pkg/term/term_windows.go index 62e9a9e69..64ead3c53 100644 --- a/vendor/github.com/docker/docker/pkg/term/term_windows.go +++ b/vendor/github.com/docker/docker/pkg/term/term_windows.go @@ -1,12 +1,10 @@ -// +build windows - -package term +package term // import "github.com/docker/docker/pkg/term" import ( "io" "os" "os/signal" - "syscall" + "syscall" // used for STD_INPUT_HANDLE, STD_OUTPUT_HANDLE and STD_ERROR_HANDLE "github.com/Azure/go-ansiterm/winterm" "github.com/docker/docker/pkg/term/windows" @@ -23,14 +21,7 @@ type Winsize struct { Width uint16 } -const ( - // https://msdn.microsoft.com/en-us/library/windows/desktop/ms683167(v=vs.85).aspx - enableVirtualTerminalInput = 0x0200 - enableVirtualTerminalProcessing = 0x0004 - disableNewlineAutoReturn = 0x0008 -) - -// vtInputSupported is true if enableVirtualTerminalInput is supported by the console +// vtInputSupported is true if winterm.ENABLE_VIRTUAL_TERMINAL_INPUT is supported by the console var vtInputSupported bool // StdStreams returns the standard streams (stdin, stdout, stderr). @@ -40,8 +31,8 @@ func StdStreams() (stdIn io.ReadCloser, stdOut, stdErr io.Writer) { var emulateStdin, emulateStdout, emulateStderr bool fd := os.Stdin.Fd() if mode, err := winterm.GetConsoleMode(fd); err == nil { - // Validate that enableVirtualTerminalInput is supported, but do not set it. - if err = winterm.SetConsoleMode(fd, mode|enableVirtualTerminalInput); err != nil { + // Validate that winterm.ENABLE_VIRTUAL_TERMINAL_INPUT is supported, but do not set it. + if err = winterm.SetConsoleMode(fd, mode|winterm.ENABLE_VIRTUAL_TERMINAL_INPUT); err != nil { emulateStdin = true } else { vtInputSupported = true @@ -53,21 +44,21 @@ func StdStreams() (stdIn io.ReadCloser, stdOut, stdErr io.Writer) { fd = os.Stdout.Fd() if mode, err := winterm.GetConsoleMode(fd); err == nil { - // Validate disableNewlineAutoReturn is supported, but do not set it. - if err = winterm.SetConsoleMode(fd, mode|enableVirtualTerminalProcessing|disableNewlineAutoReturn); err != nil { + // Validate winterm.DISABLE_NEWLINE_AUTO_RETURN is supported, but do not set it. + if err = winterm.SetConsoleMode(fd, mode|winterm.ENABLE_VIRTUAL_TERMINAL_PROCESSING|winterm.DISABLE_NEWLINE_AUTO_RETURN); err != nil { emulateStdout = true } else { - winterm.SetConsoleMode(fd, mode|enableVirtualTerminalProcessing) + winterm.SetConsoleMode(fd, mode|winterm.ENABLE_VIRTUAL_TERMINAL_PROCESSING) } } fd = os.Stderr.Fd() if mode, err := winterm.GetConsoleMode(fd); err == nil { - // Validate disableNewlineAutoReturn is supported, but do not set it. - if err = winterm.SetConsoleMode(fd, mode|enableVirtualTerminalProcessing|disableNewlineAutoReturn); err != nil { + // Validate winterm.DISABLE_NEWLINE_AUTO_RETURN is supported, but do not set it. + if err = winterm.SetConsoleMode(fd, mode|winterm.ENABLE_VIRTUAL_TERMINAL_PROCESSING|winterm.DISABLE_NEWLINE_AUTO_RETURN); err != nil { emulateStderr = true } else { - winterm.SetConsoleMode(fd, mode|enableVirtualTerminalProcessing) + winterm.SetConsoleMode(fd, mode|winterm.ENABLE_VIRTUAL_TERMINAL_PROCESSING) } } @@ -78,20 +69,24 @@ func StdStreams() (stdIn io.ReadCloser, stdOut, stdErr io.Writer) { emulateStderr = false } + // Temporarily use STD_INPUT_HANDLE, STD_OUTPUT_HANDLE and + // STD_ERROR_HANDLE from syscall rather than x/sys/windows as long as + // go-ansiterm hasn't switch to x/sys/windows. + // TODO: switch back to x/sys/windows once go-ansiterm has switched if emulateStdin { - stdIn = windows.NewAnsiReader(syscall.STD_INPUT_HANDLE) + stdIn = windowsconsole.NewAnsiReader(syscall.STD_INPUT_HANDLE) } else { stdIn = os.Stdin } if emulateStdout { - stdOut = windows.NewAnsiWriter(syscall.STD_OUTPUT_HANDLE) + stdOut = windowsconsole.NewAnsiWriter(syscall.STD_OUTPUT_HANDLE) } else { stdOut = os.Stdout } if emulateStderr { - stdErr = windows.NewAnsiWriter(syscall.STD_ERROR_HANDLE) + stdErr = windowsconsole.NewAnsiWriter(syscall.STD_ERROR_HANDLE) } else { stdErr = os.Stderr } @@ -101,7 +96,7 @@ func StdStreams() (stdIn io.ReadCloser, stdOut, stdErr io.Writer) { // GetFdInfo returns the file descriptor for an os.File and indicates whether the file represents a terminal. func GetFdInfo(in interface{}) (uintptr, bool) { - return windows.GetHandleInfo(in) + return windowsconsole.GetHandleInfo(in) } // GetWinsize returns the window size based on the specified file descriptor. @@ -121,7 +116,7 @@ func GetWinsize(fd uintptr) (*Winsize, error) { // IsTerminal returns true if the given file descriptor is a terminal. func IsTerminal(fd uintptr) bool { - return windows.IsConsole(fd) + return windowsconsole.IsConsole(fd) } // RestoreTerminal restores the terminal connected to the given file descriptor @@ -179,9 +174,9 @@ func SetRawTerminalOutput(fd uintptr) (*State, error) { return nil, err } - // Ignore failures, since disableNewlineAutoReturn might not be supported on this + // Ignore failures, since winterm.DISABLE_NEWLINE_AUTO_RETURN might not be supported on this // version of Windows. - winterm.SetConsoleMode(fd, state.mode|disableNewlineAutoReturn) + winterm.SetConsoleMode(fd, state.mode|winterm.DISABLE_NEWLINE_AUTO_RETURN) return state, err } @@ -211,7 +206,7 @@ func MakeRaw(fd uintptr) (*State, error) { mode |= winterm.ENABLE_INSERT_MODE mode |= winterm.ENABLE_QUICK_EDIT_MODE if vtInputSupported { - mode |= enableVirtualTerminalInput + mode |= winterm.ENABLE_VIRTUAL_TERMINAL_INPUT } err = winterm.SetConsoleMode(fd, mode) diff --git a/vendor/github.com/docker/docker/pkg/term/termios_bsd.go b/vendor/github.com/docker/docker/pkg/term/termios_bsd.go new file mode 100644 index 000000000..48b16f520 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/term/termios_bsd.go @@ -0,0 +1,42 @@ +// +build darwin freebsd openbsd netbsd + +package term // import "github.com/docker/docker/pkg/term" + +import ( + "unsafe" + + "golang.org/x/sys/unix" +) + +const ( + getTermios = unix.TIOCGETA + setTermios = unix.TIOCSETA +) + +// Termios is the Unix API for terminal I/O. +type Termios unix.Termios + +// MakeRaw put the terminal connected to the given file descriptor into raw +// mode and returns the previous state of the terminal so that it can be +// restored. +func MakeRaw(fd uintptr) (*State, error) { + var oldState State + if _, _, err := unix.Syscall(unix.SYS_IOCTL, fd, getTermios, uintptr(unsafe.Pointer(&oldState.termios))); err != 0 { + return nil, err + } + + newState := oldState.termios + newState.Iflag &^= (unix.IGNBRK | unix.BRKINT | unix.PARMRK | unix.ISTRIP | unix.INLCR | unix.IGNCR | unix.ICRNL | unix.IXON) + newState.Oflag &^= unix.OPOST + newState.Lflag &^= (unix.ECHO | unix.ECHONL | unix.ICANON | unix.ISIG | unix.IEXTEN) + newState.Cflag &^= (unix.CSIZE | unix.PARENB) + newState.Cflag |= unix.CS8 + newState.Cc[unix.VMIN] = 1 + newState.Cc[unix.VTIME] = 0 + + if _, _, err := unix.Syscall(unix.SYS_IOCTL, fd, setTermios, uintptr(unsafe.Pointer(&newState))); err != 0 { + return nil, err + } + + return &oldState, nil +} diff --git a/vendor/github.com/docker/docker/pkg/term/termios_darwin.go b/vendor/github.com/docker/docker/pkg/term/termios_darwin.go deleted file mode 100644 index 480db900a..000000000 --- a/vendor/github.com/docker/docker/pkg/term/termios_darwin.go +++ /dev/null @@ -1,69 +0,0 @@ -package term - -import ( - "syscall" - "unsafe" -) - -const ( - getTermios = syscall.TIOCGETA - setTermios = syscall.TIOCSETA -) - -// Termios magic numbers, passthrough to the ones defined in syscall. -const ( - IGNBRK = syscall.IGNBRK - PARMRK = syscall.PARMRK - INLCR = syscall.INLCR - IGNCR = syscall.IGNCR - ECHONL = syscall.ECHONL - CSIZE = syscall.CSIZE - ICRNL = syscall.ICRNL - ISTRIP = syscall.ISTRIP - PARENB = syscall.PARENB - ECHO = syscall.ECHO - ICANON = syscall.ICANON - ISIG = syscall.ISIG - IXON = syscall.IXON - BRKINT = syscall.BRKINT - INPCK = syscall.INPCK - OPOST = syscall.OPOST - CS8 = syscall.CS8 - IEXTEN = syscall.IEXTEN -) - -// Termios is the Unix API for terminal I/O. -type Termios struct { - Iflag uint64 - Oflag uint64 - Cflag uint64 - Lflag uint64 - Cc [20]byte - Ispeed uint64 - Ospeed uint64 -} - -// MakeRaw put the terminal connected to the given file descriptor into raw -// mode and returns the previous state of the terminal so that it can be -// restored. -func MakeRaw(fd uintptr) (*State, error) { - var oldState State - if _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, uintptr(getTermios), uintptr(unsafe.Pointer(&oldState.termios))); err != 0 { - return nil, err - } - - newState := oldState.termios - newState.Iflag &^= (IGNBRK | BRKINT | PARMRK | ISTRIP | INLCR | IGNCR | ICRNL | IXON) - newState.Oflag &^= OPOST - newState.Lflag &^= (ECHO | ECHONL | ICANON | ISIG | IEXTEN) - newState.Cflag &^= (CSIZE | PARENB) - newState.Cflag |= CS8 - newState.Cc[syscall.VMIN] = 1 - newState.Cc[syscall.VTIME] = 0 - - if _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, uintptr(setTermios), uintptr(unsafe.Pointer(&newState))); err != 0 { - return nil, err - } - - return &oldState, nil -} diff --git a/vendor/github.com/docker/docker/pkg/term/termios_freebsd.go b/vendor/github.com/docker/docker/pkg/term/termios_freebsd.go deleted file mode 100644 index ed843ad69..000000000 --- a/vendor/github.com/docker/docker/pkg/term/termios_freebsd.go +++ /dev/null @@ -1,69 +0,0 @@ -package term - -import ( - "syscall" - "unsafe" -) - -const ( - getTermios = syscall.TIOCGETA - setTermios = syscall.TIOCSETA -) - -// Termios magic numbers, passthrough to the ones defined in syscall. -const ( - IGNBRK = syscall.IGNBRK - PARMRK = syscall.PARMRK - INLCR = syscall.INLCR - IGNCR = syscall.IGNCR - ECHONL = syscall.ECHONL - CSIZE = syscall.CSIZE - ICRNL = syscall.ICRNL - ISTRIP = syscall.ISTRIP - PARENB = syscall.PARENB - ECHO = syscall.ECHO - ICANON = syscall.ICANON - ISIG = syscall.ISIG - IXON = syscall.IXON - BRKINT = syscall.BRKINT - INPCK = syscall.INPCK - OPOST = syscall.OPOST - CS8 = syscall.CS8 - IEXTEN = syscall.IEXTEN -) - -// Termios is the Unix API for terminal I/O. -type Termios struct { - Iflag uint32 - Oflag uint32 - Cflag uint32 - Lflag uint32 - Cc [20]byte - Ispeed uint32 - Ospeed uint32 -} - -// MakeRaw put the terminal connected to the given file descriptor into raw -// mode and returns the previous state of the terminal so that it can be -// restored. -func MakeRaw(fd uintptr) (*State, error) { - var oldState State - if _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, uintptr(getTermios), uintptr(unsafe.Pointer(&oldState.termios))); err != 0 { - return nil, err - } - - newState := oldState.termios - newState.Iflag &^= (IGNBRK | BRKINT | PARMRK | ISTRIP | INLCR | IGNCR | ICRNL | IXON) - newState.Oflag &^= OPOST - newState.Lflag &^= (ECHO | ECHONL | ICANON | ISIG | IEXTEN) - newState.Cflag &^= (CSIZE | PARENB) - newState.Cflag |= CS8 - newState.Cc[syscall.VMIN] = 1 - newState.Cc[syscall.VTIME] = 0 - - if _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, uintptr(setTermios), uintptr(unsafe.Pointer(&newState))); err != 0 { - return nil, err - } - - return &oldState, nil -} diff --git a/vendor/github.com/docker/docker/pkg/term/termios_linux.go b/vendor/github.com/docker/docker/pkg/term/termios_linux.go index 22921b6ae..6d4c63fdb 100644 --- a/vendor/github.com/docker/docker/pkg/term/termios_linux.go +++ b/vendor/github.com/docker/docker/pkg/term/termios_linux.go @@ -1,46 +1,38 @@ -// +build !cgo - -package term +package term // import "github.com/docker/docker/pkg/term" import ( - "syscall" - "unsafe" + "golang.org/x/sys/unix" ) const ( - getTermios = syscall.TCGETS - setTermios = syscall.TCSETS + getTermios = unix.TCGETS + setTermios = unix.TCSETS ) // Termios is the Unix API for terminal I/O. -type Termios struct { - Iflag uint32 - Oflag uint32 - Cflag uint32 - Lflag uint32 - Cc [20]byte - Ispeed uint32 - Ospeed uint32 -} +type Termios unix.Termios // MakeRaw put the terminal connected to the given file descriptor into raw // mode and returns the previous state of the terminal so that it can be // restored. func MakeRaw(fd uintptr) (*State, error) { - var oldState State - if _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, getTermios, uintptr(unsafe.Pointer(&oldState.termios))); err != 0 { + termios, err := unix.IoctlGetTermios(int(fd), getTermios) + if err != nil { return nil, err } - newState := oldState.termios + var oldState State + oldState.termios = Termios(*termios) - newState.Iflag &^= (syscall.IGNBRK | syscall.BRKINT | syscall.PARMRK | syscall.ISTRIP | syscall.INLCR | syscall.IGNCR | syscall.ICRNL | syscall.IXON) - newState.Oflag &^= syscall.OPOST - newState.Lflag &^= (syscall.ECHO | syscall.ECHONL | syscall.ICANON | syscall.ISIG | syscall.IEXTEN) - newState.Cflag &^= (syscall.CSIZE | syscall.PARENB) - newState.Cflag |= syscall.CS8 + termios.Iflag &^= (unix.IGNBRK | unix.BRKINT | unix.PARMRK | unix.ISTRIP | unix.INLCR | unix.IGNCR | unix.ICRNL | unix.IXON) + termios.Oflag &^= unix.OPOST + termios.Lflag &^= (unix.ECHO | unix.ECHONL | unix.ICANON | unix.ISIG | unix.IEXTEN) + termios.Cflag &^= (unix.CSIZE | unix.PARENB) + termios.Cflag |= unix.CS8 + termios.Cc[unix.VMIN] = 1 + termios.Cc[unix.VTIME] = 0 - if _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, setTermios, uintptr(unsafe.Pointer(&newState))); err != 0 { + if err := unix.IoctlSetTermios(int(fd), setTermios, termios); err != nil { return nil, err } return &oldState, nil diff --git a/vendor/github.com/docker/docker/pkg/term/termios_openbsd.go b/vendor/github.com/docker/docker/pkg/term/termios_openbsd.go deleted file mode 100644 index ed843ad69..000000000 --- a/vendor/github.com/docker/docker/pkg/term/termios_openbsd.go +++ /dev/null @@ -1,69 +0,0 @@ -package term - -import ( - "syscall" - "unsafe" -) - -const ( - getTermios = syscall.TIOCGETA - setTermios = syscall.TIOCSETA -) - -// Termios magic numbers, passthrough to the ones defined in syscall. -const ( - IGNBRK = syscall.IGNBRK - PARMRK = syscall.PARMRK - INLCR = syscall.INLCR - IGNCR = syscall.IGNCR - ECHONL = syscall.ECHONL - CSIZE = syscall.CSIZE - ICRNL = syscall.ICRNL - ISTRIP = syscall.ISTRIP - PARENB = syscall.PARENB - ECHO = syscall.ECHO - ICANON = syscall.ICANON - ISIG = syscall.ISIG - IXON = syscall.IXON - BRKINT = syscall.BRKINT - INPCK = syscall.INPCK - OPOST = syscall.OPOST - CS8 = syscall.CS8 - IEXTEN = syscall.IEXTEN -) - -// Termios is the Unix API for terminal I/O. -type Termios struct { - Iflag uint32 - Oflag uint32 - Cflag uint32 - Lflag uint32 - Cc [20]byte - Ispeed uint32 - Ospeed uint32 -} - -// MakeRaw put the terminal connected to the given file descriptor into raw -// mode and returns the previous state of the terminal so that it can be -// restored. -func MakeRaw(fd uintptr) (*State, error) { - var oldState State - if _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, uintptr(getTermios), uintptr(unsafe.Pointer(&oldState.termios))); err != 0 { - return nil, err - } - - newState := oldState.termios - newState.Iflag &^= (IGNBRK | BRKINT | PARMRK | ISTRIP | INLCR | IGNCR | ICRNL | IXON) - newState.Oflag &^= OPOST - newState.Lflag &^= (ECHO | ECHONL | ICANON | ISIG | IEXTEN) - newState.Cflag &^= (CSIZE | PARENB) - newState.Cflag |= CS8 - newState.Cc[syscall.VMIN] = 1 - newState.Cc[syscall.VTIME] = 0 - - if _, _, err := syscall.Syscall(syscall.SYS_IOCTL, fd, uintptr(setTermios), uintptr(unsafe.Pointer(&newState))); err != 0 { - return nil, err - } - - return &oldState, nil -} diff --git a/vendor/github.com/docker/docker/pkg/term/windows/ansi_reader.go b/vendor/github.com/docker/docker/pkg/term/windows/ansi_reader.go index cb0b88356..1d7c452cc 100644 --- a/vendor/github.com/docker/docker/pkg/term/windows/ansi_reader.go +++ b/vendor/github.com/docker/docker/pkg/term/windows/ansi_reader.go @@ -1,6 +1,6 @@ // +build windows -package windows +package windowsconsole // import "github.com/docker/docker/pkg/term/windows" import ( "bytes" diff --git a/vendor/github.com/docker/docker/pkg/term/windows/ansi_writer.go b/vendor/github.com/docker/docker/pkg/term/windows/ansi_writer.go index a3ce5697d..7799a03fc 100644 --- a/vendor/github.com/docker/docker/pkg/term/windows/ansi_writer.go +++ b/vendor/github.com/docker/docker/pkg/term/windows/ansi_writer.go @@ -1,6 +1,6 @@ // +build windows -package windows +package windowsconsole // import "github.com/docker/docker/pkg/term/windows" import ( "io" diff --git a/vendor/github.com/docker/docker/pkg/term/windows/console.go b/vendor/github.com/docker/docker/pkg/term/windows/console.go index ca5c3b2e5..527401975 100644 --- a/vendor/github.com/docker/docker/pkg/term/windows/console.go +++ b/vendor/github.com/docker/docker/pkg/term/windows/console.go @@ -1,6 +1,6 @@ // +build windows -package windows +package windowsconsole // import "github.com/docker/docker/pkg/term/windows" import ( "os" diff --git a/vendor/github.com/docker/docker/pkg/term/windows/windows.go b/vendor/github.com/docker/docker/pkg/term/windows/windows.go index ce4cb5990..3e5593ca6 100644 --- a/vendor/github.com/docker/docker/pkg/term/windows/windows.go +++ b/vendor/github.com/docker/docker/pkg/term/windows/windows.go @@ -2,15 +2,15 @@ // When asked for the set of standard streams (e.g., stdin, stdout, stderr), the code will create // and return pseudo-streams that convert ANSI sequences to / from Windows Console API calls. -package windows +package windowsconsole // import "github.com/docker/docker/pkg/term/windows" import ( "io/ioutil" "os" "sync" - ansiterm "github.com/Azure/go-ansiterm" - "github.com/Sirupsen/logrus" + "github.com/Azure/go-ansiterm" + "github.com/sirupsen/logrus" ) var logger *logrus.Logger diff --git a/vendor/github.com/docker/docker/pkg/term/winsize.go b/vendor/github.com/docker/docker/pkg/term/winsize.go new file mode 100644 index 000000000..a19663ad8 --- /dev/null +++ b/vendor/github.com/docker/docker/pkg/term/winsize.go @@ -0,0 +1,20 @@ +// +build !windows + +package term // import "github.com/docker/docker/pkg/term" + +import ( + "golang.org/x/sys/unix" +) + +// GetWinsize returns the window size based on the specified file descriptor. +func GetWinsize(fd uintptr) (*Winsize, error) { + uws, err := unix.IoctlGetWinsize(int(fd), unix.TIOCGWINSZ) + ws := &Winsize{Height: uws.Row, Width: uws.Col, x: uws.Xpixel, y: uws.Ypixel} + return ws, err +} + +// SetWinsize tries to set the specified window size for the specified file descriptor. +func SetWinsize(fd uintptr, ws *Winsize) error { + uws := &unix.Winsize{Row: ws.Height, Col: ws.Width, Xpixel: ws.x, Ypixel: ws.y} + return unix.IoctlSetWinsize(int(fd), unix.TIOCSWINSZ, uws) +} diff --git a/vendor/github.com/docker/docker/reference/reference.go b/vendor/github.com/docker/docker/reference/reference.go deleted file mode 100644 index d05d86e70..000000000 --- a/vendor/github.com/docker/docker/reference/reference.go +++ /dev/null @@ -1,216 +0,0 @@ -package reference - -import ( - "errors" - "fmt" - "strings" - - distreference "github.com/docker/distribution/reference" - "github.com/docker/docker/pkg/stringid" - digest "github.com/opencontainers/go-digest" -) - -const ( - // DefaultTag defines the default tag used when performing images related actions and no tag or digest is specified - DefaultTag = "latest" - // DefaultHostname is the default built-in hostname - DefaultHostname = "docker.io" - // LegacyDefaultHostname is automatically converted to DefaultHostname - LegacyDefaultHostname = "index.docker.io" - // DefaultRepoPrefix is the prefix used for default repositories in default host - DefaultRepoPrefix = "library/" -) - -// Named is an object with a full name -type Named interface { - // Name returns normalized repository name, like "ubuntu". - Name() string - // String returns full reference, like "ubuntu@sha256:abcdef..." - String() string - // FullName returns full repository name with hostname, like "docker.io/library/ubuntu" - FullName() string - // Hostname returns hostname for the reference, like "docker.io" - Hostname() string - // RemoteName returns the repository component of the full name, like "library/ubuntu" - RemoteName() string -} - -// NamedTagged is an object including a name and tag. -type NamedTagged interface { - Named - Tag() string -} - -// Canonical reference is an object with a fully unique -// name including a name with hostname and digest -type Canonical interface { - Named - Digest() digest.Digest -} - -// ParseNamed parses s and returns a syntactically valid reference implementing -// the Named interface. The reference must have a name, otherwise an error is -// returned. -// If an error was encountered it is returned, along with a nil Reference. -func ParseNamed(s string) (Named, error) { - named, err := distreference.ParseNamed(s) - if err != nil { - return nil, fmt.Errorf("Error parsing reference: %q is not a valid repository/tag: %s", s, err) - } - r, err := WithName(named.Name()) - if err != nil { - return nil, err - } - if canonical, isCanonical := named.(distreference.Canonical); isCanonical { - return WithDigest(r, canonical.Digest()) - } - if tagged, isTagged := named.(distreference.NamedTagged); isTagged { - return WithTag(r, tagged.Tag()) - } - return r, nil -} - -// TrimNamed removes any tag or digest from the named reference -func TrimNamed(ref Named) Named { - return &namedRef{distreference.TrimNamed(ref)} -} - -// WithName returns a named object representing the given string. If the input -// is invalid ErrReferenceInvalidFormat will be returned. -func WithName(name string) (Named, error) { - name, err := normalize(name) - if err != nil { - return nil, err - } - if err := validateName(name); err != nil { - return nil, err - } - r, err := distreference.WithName(name) - if err != nil { - return nil, err - } - return &namedRef{r}, nil -} - -// WithTag combines the name from "name" and the tag from "tag" to form a -// reference incorporating both the name and the tag. -func WithTag(name Named, tag string) (NamedTagged, error) { - r, err := distreference.WithTag(name, tag) - if err != nil { - return nil, err - } - return &taggedRef{namedRef{r}}, nil -} - -// WithDigest combines the name from "name" and the digest from "digest" to form -// a reference incorporating both the name and the digest. -func WithDigest(name Named, digest digest.Digest) (Canonical, error) { - r, err := distreference.WithDigest(name, digest) - if err != nil { - return nil, err - } - return &canonicalRef{namedRef{r}}, nil -} - -type namedRef struct { - distreference.Named -} -type taggedRef struct { - namedRef -} -type canonicalRef struct { - namedRef -} - -func (r *namedRef) FullName() string { - hostname, remoteName := splitHostname(r.Name()) - return hostname + "/" + remoteName -} -func (r *namedRef) Hostname() string { - hostname, _ := splitHostname(r.Name()) - return hostname -} -func (r *namedRef) RemoteName() string { - _, remoteName := splitHostname(r.Name()) - return remoteName -} -func (r *taggedRef) Tag() string { - return r.namedRef.Named.(distreference.NamedTagged).Tag() -} -func (r *canonicalRef) Digest() digest.Digest { - return r.namedRef.Named.(distreference.Canonical).Digest() -} - -// WithDefaultTag adds a default tag to a reference if it only has a repo name. -func WithDefaultTag(ref Named) Named { - if IsNameOnly(ref) { - ref, _ = WithTag(ref, DefaultTag) - } - return ref -} - -// IsNameOnly returns true if reference only contains a repo name. -func IsNameOnly(ref Named) bool { - if _, ok := ref.(NamedTagged); ok { - return false - } - if _, ok := ref.(Canonical); ok { - return false - } - return true -} - -// ParseIDOrReference parses string for an image ID or a reference. ID can be -// without a default prefix. -func ParseIDOrReference(idOrRef string) (digest.Digest, Named, error) { - if err := stringid.ValidateID(idOrRef); err == nil { - idOrRef = "sha256:" + idOrRef - } - if dgst, err := digest.Parse(idOrRef); err == nil { - return dgst, nil, nil - } - ref, err := ParseNamed(idOrRef) - return "", ref, err -} - -// splitHostname splits a repository name to hostname and remotename string. -// If no valid hostname is found, the default hostname is used. Repository name -// needs to be already validated before. -func splitHostname(name string) (hostname, remoteName string) { - i := strings.IndexRune(name, '/') - if i == -1 || (!strings.ContainsAny(name[:i], ".:") && name[:i] != "localhost") { - hostname, remoteName = DefaultHostname, name - } else { - hostname, remoteName = name[:i], name[i+1:] - } - if hostname == LegacyDefaultHostname { - hostname = DefaultHostname - } - if hostname == DefaultHostname && !strings.ContainsRune(remoteName, '/') { - remoteName = DefaultRepoPrefix + remoteName - } - return -} - -// normalize returns a repository name in its normalized form, meaning it -// will not contain default hostname nor library/ prefix for official images. -func normalize(name string) (string, error) { - host, remoteName := splitHostname(name) - if strings.ToLower(remoteName) != remoteName { - return "", errors.New("invalid reference format: repository name must be lowercase") - } - if host == DefaultHostname { - if strings.HasPrefix(remoteName, DefaultRepoPrefix) { - return strings.TrimPrefix(remoteName, DefaultRepoPrefix), nil - } - return remoteName, nil - } - return name, nil -} - -func validateName(name string) error { - if err := stringid.ValidateID(name); err == nil { - return fmt.Errorf("Invalid repository name (%s), cannot specify 64-byte hexadecimal strings", name) - } - return nil -} diff --git a/vendor/github.com/docker/docker/registry/auth.go b/vendor/github.com/docker/docker/registry/auth.go index e24330634..1f2043a0d 100644 --- a/vendor/github.com/docker/docker/registry/auth.go +++ b/vendor/github.com/docker/docker/registry/auth.go @@ -1,10 +1,20 @@ -package registry +package registry // import "github.com/docker/docker/registry" import ( + "io/ioutil" + "net/http" + "net/url" "strings" + "time" + "github.com/docker/distribution/registry/client/auth" + "github.com/docker/distribution/registry/client/auth/challenge" + "github.com/docker/distribution/registry/client/transport" "github.com/docker/docker/api/types" registrytypes "github.com/docker/docker/api/types/registry" + "github.com/docker/docker/errdefs" + "github.com/pkg/errors" + "github.com/sirupsen/logrus" ) const ( @@ -12,6 +22,183 @@ const ( AuthClientID = "docker" ) +// loginV1 tries to register/login to the v1 registry server. +func loginV1(authConfig *types.AuthConfig, apiEndpoint APIEndpoint, userAgent string) (string, string, error) { + registryEndpoint := apiEndpoint.ToV1Endpoint(userAgent, nil) + serverAddress := registryEndpoint.String() + + logrus.Debugf("attempting v1 login to registry endpoint %s", serverAddress) + + if serverAddress == "" { + return "", "", errdefs.System(errors.New("server Error: Server Address not set")) + } + + req, err := http.NewRequest("GET", serverAddress+"users/", nil) + if err != nil { + return "", "", err + } + req.SetBasicAuth(authConfig.Username, authConfig.Password) + resp, err := registryEndpoint.client.Do(req) + if err != nil { + // fallback when request could not be completed + return "", "", fallbackError{ + err: err, + } + } + defer resp.Body.Close() + body, err := ioutil.ReadAll(resp.Body) + if err != nil { + return "", "", errdefs.System(err) + } + + switch resp.StatusCode { + case http.StatusOK: + return "Login Succeeded", "", nil + case http.StatusUnauthorized: + return "", "", errdefs.Unauthorized(errors.New("Wrong login/password, please try again")) + case http.StatusForbidden: + // *TODO: Use registry configuration to determine what this says, if anything? + return "", "", errdefs.Forbidden(errors.Errorf("Login: Account is not active. Please see the documentation of the registry %s for instructions how to activate it.", serverAddress)) + case http.StatusInternalServerError: + logrus.Errorf("%s returned status code %d. Response Body :\n%s", req.URL.String(), resp.StatusCode, body) + return "", "", errdefs.System(errors.New("Internal Server Error")) + } + return "", "", errdefs.System(errors.Errorf("Login: %s (Code: %d; Headers: %s)", body, + resp.StatusCode, resp.Header)) +} + +type loginCredentialStore struct { + authConfig *types.AuthConfig +} + +func (lcs loginCredentialStore) Basic(*url.URL) (string, string) { + return lcs.authConfig.Username, lcs.authConfig.Password +} + +func (lcs loginCredentialStore) RefreshToken(*url.URL, string) string { + return lcs.authConfig.IdentityToken +} + +func (lcs loginCredentialStore) SetRefreshToken(u *url.URL, service, token string) { + lcs.authConfig.IdentityToken = token +} + +type staticCredentialStore struct { + auth *types.AuthConfig +} + +// NewStaticCredentialStore returns a credential store +// which always returns the same credential values. +func NewStaticCredentialStore(auth *types.AuthConfig) auth.CredentialStore { + return staticCredentialStore{ + auth: auth, + } +} + +func (scs staticCredentialStore) Basic(*url.URL) (string, string) { + if scs.auth == nil { + return "", "" + } + return scs.auth.Username, scs.auth.Password +} + +func (scs staticCredentialStore) RefreshToken(*url.URL, string) string { + if scs.auth == nil { + return "" + } + return scs.auth.IdentityToken +} + +func (scs staticCredentialStore) SetRefreshToken(*url.URL, string, string) { +} + +type fallbackError struct { + err error +} + +func (err fallbackError) Error() string { + return err.err.Error() +} + +// loginV2 tries to login to the v2 registry server. The given registry +// endpoint will be pinged to get authorization challenges. These challenges +// will be used to authenticate against the registry to validate credentials. +func loginV2(authConfig *types.AuthConfig, endpoint APIEndpoint, userAgent string) (string, string, error) { + logrus.Debugf("attempting v2 login to registry endpoint %s", strings.TrimRight(endpoint.URL.String(), "/")+"/v2/") + + modifiers := Headers(userAgent, nil) + authTransport := transport.NewTransport(NewTransport(endpoint.TLSConfig), modifiers...) + + credentialAuthConfig := *authConfig + creds := loginCredentialStore{ + authConfig: &credentialAuthConfig, + } + + loginClient, foundV2, err := v2AuthHTTPClient(endpoint.URL, authTransport, modifiers, creds, nil) + if err != nil { + return "", "", err + } + + endpointStr := strings.TrimRight(endpoint.URL.String(), "/") + "/v2/" + req, err := http.NewRequest("GET", endpointStr, nil) + if err != nil { + if !foundV2 { + err = fallbackError{err: err} + } + return "", "", err + } + + resp, err := loginClient.Do(req) + if err != nil { + err = translateV2AuthError(err) + if !foundV2 { + err = fallbackError{err: err} + } + + return "", "", err + } + defer resp.Body.Close() + + if resp.StatusCode == http.StatusOK { + return "Login Succeeded", credentialAuthConfig.IdentityToken, nil + } + + // TODO(dmcgowan): Attempt to further interpret result, status code and error code string + err = errors.Errorf("login attempt to %s failed with status: %d %s", endpointStr, resp.StatusCode, http.StatusText(resp.StatusCode)) + if !foundV2 { + err = fallbackError{err: err} + } + return "", "", err +} + +func v2AuthHTTPClient(endpoint *url.URL, authTransport http.RoundTripper, modifiers []transport.RequestModifier, creds auth.CredentialStore, scopes []auth.Scope) (*http.Client, bool, error) { + challengeManager, foundV2, err := PingV2Registry(endpoint, authTransport) + if err != nil { + if !foundV2 { + err = fallbackError{err: err} + } + return nil, foundV2, err + } + + tokenHandlerOptions := auth.TokenHandlerOptions{ + Transport: authTransport, + Credentials: creds, + OfflineAccess: true, + ClientID: AuthClientID, + Scopes: scopes, + } + tokenHandler := auth.NewTokenHandlerWithOptions(tokenHandlerOptions) + basicHandler := auth.NewBasicHandler(creds) + modifiers = append(modifiers, auth.NewAuthorizer(challengeManager, tokenHandler, basicHandler)) + tr := transport.NewTransport(authTransport, modifiers...) + + return &http.Client{ + Transport: tr, + Timeout: 15 * time.Second, + }, foundV2, nil + +} + // ConvertToHostname converts a registry url which has http|https prepended // to just an hostname. func ConvertToHostname(url string) string { @@ -46,3 +233,64 @@ func ResolveAuthConfig(authConfigs map[string]types.AuthConfig, index *registryt // When all else fails, return an empty auth config return types.AuthConfig{} } + +// PingResponseError is used when the response from a ping +// was received but invalid. +type PingResponseError struct { + Err error +} + +func (err PingResponseError) Error() string { + return err.Err.Error() +} + +// PingV2Registry attempts to ping a v2 registry and on success return a +// challenge manager for the supported authentication types and +// whether v2 was confirmed by the response. If a response is received but +// cannot be interpreted a PingResponseError will be returned. +// nolint: interfacer +func PingV2Registry(endpoint *url.URL, transport http.RoundTripper) (challenge.Manager, bool, error) { + var ( + foundV2 = false + v2Version = auth.APIVersion{ + Type: "registry", + Version: "2.0", + } + ) + + pingClient := &http.Client{ + Transport: transport, + Timeout: 15 * time.Second, + } + endpointStr := strings.TrimRight(endpoint.String(), "/") + "/v2/" + req, err := http.NewRequest("GET", endpointStr, nil) + if err != nil { + return nil, false, err + } + resp, err := pingClient.Do(req) + if err != nil { + return nil, false, err + } + defer resp.Body.Close() + + versions := auth.APIVersions(resp, DefaultRegistryVersionHeader) + for _, pingVersion := range versions { + if pingVersion == v2Version { + // The version header indicates we're definitely + // talking to a v2 registry. So don't allow future + // fallbacks to the v1 protocol. + + foundV2 = true + break + } + } + + challengeManager := challenge.NewSimpleManager() + if err := challengeManager.AddResponse(resp); err != nil { + return nil, foundV2, PingResponseError{ + Err: err, + } + } + + return challengeManager, foundV2, nil +} diff --git a/vendor/github.com/docker/docker/registry/config.go b/vendor/github.com/docker/docker/registry/config.go index c27769772..de5a526b6 100644 --- a/vendor/github.com/docker/docker/registry/config.go +++ b/vendor/github.com/docker/docker/registry/config.go @@ -1,20 +1,24 @@ -package registry +package registry // import "github.com/docker/docker/registry" import ( - "errors" "fmt" "net" "net/url" + "regexp" + "strconv" "strings" + "github.com/docker/distribution/reference" registrytypes "github.com/docker/docker/api/types/registry" - "github.com/docker/docker/reference" + "github.com/pkg/errors" + "github.com/sirupsen/logrus" ) // ServiceOptions holds command line options. type ServiceOptions struct { - Mirrors []string `json:"registry-mirrors,omitempty"` - InsecureRegistries []string `json:"insecure-registries,omitempty"` + AllowNondistributableArtifacts []string `json:"allow-nondistributable-artifacts,omitempty"` + Mirrors []string `json:"registry-mirrors,omitempty"` + InsecureRegistries []string `json:"insecure-registries,omitempty"` // V2Only controls access to legacy registries. If it is set to true via the // command line flag the daemon will not attempt to contact v1 legacy registries @@ -41,9 +45,6 @@ var ( // IndexName is the name of the index IndexName = "docker.io" - // NotaryServer is the endpoint serving the Notary trust server - NotaryServer = "https://notary.docker.io" - // DefaultV2Registry is the URI of the default v2 registry DefaultV2Registry = &url.URL{ Scheme: "https", @@ -56,28 +57,75 @@ var ( // not have the correct form ErrInvalidRepositoryName = errors.New("Invalid repository name (ex: \"registry.domain.tld/myrepos\")") - emptyServiceConfig = newServiceConfig(ServiceOptions{}) + emptyServiceConfig, _ = newServiceConfig(ServiceOptions{}) +) + +var ( + validHostPortRegex = regexp.MustCompile(`^` + reference.DomainRegexp.String() + `$`) ) // for mocking in unit tests var lookupIP = net.LookupIP // newServiceConfig returns a new instance of ServiceConfig -func newServiceConfig(options ServiceOptions) *serviceConfig { +func newServiceConfig(options ServiceOptions) (*serviceConfig, error) { config := &serviceConfig{ ServiceConfig: registrytypes.ServiceConfig{ InsecureRegistryCIDRs: make([]*registrytypes.NetIPNet, 0), - IndexConfigs: make(map[string]*registrytypes.IndexInfo, 0), + IndexConfigs: make(map[string]*registrytypes.IndexInfo), // Hack: Bypass setting the mirrors to IndexConfigs since they are going away // and Mirrors are only for the official registry anyways. }, V2Only: options.V2Only, } + if err := config.LoadAllowNondistributableArtifacts(options.AllowNondistributableArtifacts); err != nil { + return nil, err + } + if err := config.LoadMirrors(options.Mirrors); err != nil { + return nil, err + } + if err := config.LoadInsecureRegistries(options.InsecureRegistries); err != nil { + return nil, err + } - config.LoadMirrors(options.Mirrors) - config.LoadInsecureRegistries(options.InsecureRegistries) + return config, nil +} - return config +// LoadAllowNondistributableArtifacts loads allow-nondistributable-artifacts registries into config. +func (config *serviceConfig) LoadAllowNondistributableArtifacts(registries []string) error { + cidrs := map[string]*registrytypes.NetIPNet{} + hostnames := map[string]bool{} + + for _, r := range registries { + if _, err := ValidateIndexName(r); err != nil { + return err + } + if validateNoScheme(r) != nil { + return fmt.Errorf("allow-nondistributable-artifacts registry %s should not contain '://'", r) + } + + if _, ipnet, err := net.ParseCIDR(r); err == nil { + // Valid CIDR. + cidrs[ipnet.String()] = (*registrytypes.NetIPNet)(ipnet) + } else if err := validateHostPort(r); err == nil { + // Must be `host:port` if not CIDR. + hostnames[r] = true + } else { + return fmt.Errorf("allow-nondistributable-artifacts registry %s is not valid: %v", r, err) + } + } + + config.AllowNondistributableArtifactsCIDRs = make([]*(registrytypes.NetIPNet), 0) + for _, c := range cidrs { + config.AllowNondistributableArtifactsCIDRs = append(config.AllowNondistributableArtifactsCIDRs, c) + } + + config.AllowNondistributableArtifactsHostnames = make([]string, 0) + for h := range hostnames { + config.AllowNondistributableArtifactsHostnames = append(config.AllowNondistributableArtifactsHostnames, h) + } + + return nil } // LoadMirrors loads mirrors to config, after removing duplicates. @@ -125,7 +173,7 @@ func (config *serviceConfig) LoadInsecureRegistries(registries []string) error { originalIndexInfos := config.ServiceConfig.IndexConfigs config.ServiceConfig.InsecureRegistryCIDRs = make([]*registrytypes.NetIPNet, 0) - config.ServiceConfig.IndexConfigs = make(map[string]*registrytypes.IndexInfo, 0) + config.ServiceConfig.IndexConfigs = make(map[string]*registrytypes.IndexInfo) skip: for _, r := range registries { @@ -136,6 +184,19 @@ skip: config.ServiceConfig.IndexConfigs = originalIndexInfos return err } + if strings.HasPrefix(strings.ToLower(r), "http://") { + logrus.Warnf("insecure registry %s should not contain 'http://' and 'http://' has been removed from the insecure registry config", r) + r = r[7:] + } else if strings.HasPrefix(strings.ToLower(r), "https://") { + logrus.Warnf("insecure registry %s should not contain 'https://' and 'https://' has been removed from the insecure registry config", r) + r = r[8:] + } else if validateNoScheme(r) != nil { + // Insecure registry should not contain '://' + // before returning err, roll back to original data + config.ServiceConfig.InsecureRegistryCIDRs = originalCIDRs + config.ServiceConfig.IndexConfigs = originalIndexInfos + return fmt.Errorf("insecure registry %s should not contain '://'", r) + } // Check if CIDR was passed to --insecure-registry _, ipnet, err := net.ParseCIDR(r) if err == nil { @@ -150,6 +211,12 @@ skip: config.InsecureRegistryCIDRs = append(config.InsecureRegistryCIDRs, data) } else { + if err := validateHostPort(r); err != nil { + config.ServiceConfig.InsecureRegistryCIDRs = originalCIDRs + config.ServiceConfig.IndexConfigs = originalIndexInfos + return fmt.Errorf("insecure registry %s is not valid: %v", r, err) + + } // Assume `host:port` if not CIDR. config.IndexConfigs[r] = ®istrytypes.IndexInfo{ Name: r, @@ -171,6 +238,25 @@ skip: return nil } +// allowNondistributableArtifacts returns true if the provided hostname is part of the list of registries +// that allow push of nondistributable artifacts. +// +// The list can contain elements with CIDR notation to specify a whole subnet. If the subnet contains an IP +// of the registry specified by hostname, true is returned. +// +// hostname should be a URL.Host (`host:port` or `host`) where the `host` part can be either a domain name +// or an IP address. If it is a domain name, then it will be resolved to IP addresses for matching. If +// resolution fails, CIDR matching is not performed. +func allowNondistributableArtifacts(config *serviceConfig, hostname string) bool { + for _, h := range config.AllowNondistributableArtifactsHostnames { + if h == hostname { + return true + } + } + + return isCIDRMatch(config.AllowNondistributableArtifactsCIDRs, hostname) +} + // isSecureIndex returns false if the provided indexName is part of the list of insecure registries // Insecure registries accept HTTP and/or accept HTTPS with certificates from unknown CAs. // @@ -189,10 +275,17 @@ func isSecureIndex(config *serviceConfig, indexName string) bool { return index.Secure } - host, _, err := net.SplitHostPort(indexName) + return !isCIDRMatch(config.InsecureRegistryCIDRs, indexName) +} + +// isCIDRMatch returns true if URLHost matches an element of cidrs. URLHost is a URL.Host (`host:port` or `host`) +// where the `host` part can be either a domain name or an IP address. If it is a domain name, then it will be +// resolved to IP addresses for matching. If resolution fails, false is returned. +func isCIDRMatch(cidrs []*registrytypes.NetIPNet, URLHost string) bool { + host, _, err := net.SplitHostPort(URLHost) if err != nil { - // assume indexName is of the form `host` without the port and go on. - host = indexName + // Assume URLHost is of the form `host` without the port and go on. + host = URLHost } addrs, err := lookupIP(host) @@ -209,15 +302,15 @@ func isSecureIndex(config *serviceConfig, indexName string) bool { // Try CIDR notation only if addrs has any elements, i.e. if `host`'s IP could be determined. for _, addr := range addrs { - for _, ipnet := range config.InsecureRegistryCIDRs { + for _, ipnet := range cidrs { // check if the addr falls in the subnet if (*net.IPNet)(ipnet).Contains(addr) { - return false + return true } } } - return true + return false } // ValidateMirror validates an HTTP(S) registry mirror @@ -242,11 +335,12 @@ func ValidateMirror(val string) (string, error) { // ValidateIndexName validates an index name. func ValidateIndexName(val string) (string, error) { - if val == reference.LegacyDefaultHostname { - val = reference.DefaultHostname + // TODO: upstream this to check to reference package + if val == "index.docker.io" { + val = "docker.io" } if strings.HasPrefix(val, "-") || strings.HasSuffix(val, "-") { - return "", fmt.Errorf("Invalid index name (%s). Cannot begin or end with a hyphen.", val) + return "", fmt.Errorf("invalid index name (%s). Cannot begin or end with a hyphen", val) } return val, nil } @@ -259,6 +353,30 @@ func validateNoScheme(reposName string) error { return nil } +func validateHostPort(s string) error { + // Split host and port, and in case s can not be splitted, assume host only + host, port, err := net.SplitHostPort(s) + if err != nil { + host = s + port = "" + } + // If match against the `host:port` pattern fails, + // it might be `IPv6:port`, which will be captured by net.ParseIP(host) + if !validHostPortRegex.MatchString(s) && net.ParseIP(host) == nil { + return fmt.Errorf("invalid host %q", host) + } + if port != "" { + v, err := strconv.Atoi(port) + if err != nil { + return err + } + if v < 0 || v > 65535 { + return fmt.Errorf("invalid port %q", port) + } + } + return nil +} + // newIndexInfo returns IndexInfo configuration from indexName func newIndexInfo(config *serviceConfig, indexName string) (*registrytypes.IndexInfo, error) { var err error @@ -293,13 +411,14 @@ func GetAuthConfigKey(index *registrytypes.IndexInfo) string { // newRepositoryInfo validates and breaks down a repository name into a RepositoryInfo func newRepositoryInfo(config *serviceConfig, name reference.Named) (*RepositoryInfo, error) { - index, err := newIndexInfo(config, name.Hostname()) + index, err := newIndexInfo(config, reference.Domain(name)) if err != nil { return nil, err } - official := !strings.ContainsRune(name.Name(), '/') + official := !strings.ContainsRune(reference.FamiliarName(name), '/') + return &RepositoryInfo{ - Named: name, + Name: reference.TrimNamed(name), Index: index, Official: official, }, nil diff --git a/vendor/github.com/docker/docker/registry/config_unix.go b/vendor/github.com/docker/docker/registry/config_unix.go new file mode 100644 index 000000000..20fb47bca --- /dev/null +++ b/vendor/github.com/docker/docker/registry/config_unix.go @@ -0,0 +1,16 @@ +// +build !windows + +package registry // import "github.com/docker/docker/registry" + +var ( + // CertsDir is the directory where certificates are stored + CertsDir = "/etc/docker/certs.d" +) + +// cleanPath is used to ensure that a directory name is valid on the target +// platform. It will be passed in something *similar* to a URL such as +// https:/index.docker.io/v1. Not all platforms support directory names +// which contain those characters (such as : on Windows) +func cleanPath(s string) string { + return s +} diff --git a/vendor/github.com/docker/docker/registry/config_windows.go b/vendor/github.com/docker/docker/registry/config_windows.go new file mode 100644 index 000000000..6de0508f8 --- /dev/null +++ b/vendor/github.com/docker/docker/registry/config_windows.go @@ -0,0 +1,18 @@ +package registry // import "github.com/docker/docker/registry" + +import ( + "os" + "path/filepath" + "strings" +) + +// CertsDir is the directory where certificates are stored +var CertsDir = os.Getenv("programdata") + `\docker\certs.d` + +// cleanPath is used to ensure that a directory name is valid on the target +// platform. It will be passed in something *similar* to a URL such as +// https:\index.docker.io\v1. Not all platforms support directory names +// which contain those characters (such as : on Windows) +func cleanPath(s string) string { + return filepath.FromSlash(strings.Replace(s, ":", "", -1)) +} diff --git a/vendor/github.com/docker/docker/registry/endpoint_v1.go b/vendor/github.com/docker/docker/registry/endpoint_v1.go new file mode 100644 index 000000000..832fdb95a --- /dev/null +++ b/vendor/github.com/docker/docker/registry/endpoint_v1.go @@ -0,0 +1,198 @@ +package registry // import "github.com/docker/docker/registry" + +import ( + "crypto/tls" + "encoding/json" + "fmt" + "io/ioutil" + "net/http" + "net/url" + "strings" + + "github.com/docker/distribution/registry/client/transport" + registrytypes "github.com/docker/docker/api/types/registry" + "github.com/sirupsen/logrus" +) + +// V1Endpoint stores basic information about a V1 registry endpoint. +type V1Endpoint struct { + client *http.Client + URL *url.URL + IsSecure bool +} + +// NewV1Endpoint parses the given address to return a registry endpoint. +func NewV1Endpoint(index *registrytypes.IndexInfo, userAgent string, metaHeaders http.Header) (*V1Endpoint, error) { + tlsConfig, err := newTLSConfig(index.Name, index.Secure) + if err != nil { + return nil, err + } + + endpoint, err := newV1EndpointFromStr(GetAuthConfigKey(index), tlsConfig, userAgent, metaHeaders) + if err != nil { + return nil, err + } + + if err := validateEndpoint(endpoint); err != nil { + return nil, err + } + + return endpoint, nil +} + +func validateEndpoint(endpoint *V1Endpoint) error { + logrus.Debugf("pinging registry endpoint %s", endpoint) + + // Try HTTPS ping to registry + endpoint.URL.Scheme = "https" + if _, err := endpoint.Ping(); err != nil { + if endpoint.IsSecure { + // If registry is secure and HTTPS failed, show user the error and tell them about `--insecure-registry` + // in case that's what they need. DO NOT accept unknown CA certificates, and DO NOT fallback to HTTP. + return fmt.Errorf("invalid registry endpoint %s: %v. If this private registry supports only HTTP or HTTPS with an unknown CA certificate, please add `--insecure-registry %s` to the daemon's arguments. In the case of HTTPS, if you have access to the registry's CA certificate, no need for the flag; simply place the CA certificate at /etc/docker/certs.d/%s/ca.crt", endpoint, err, endpoint.URL.Host, endpoint.URL.Host) + } + + // If registry is insecure and HTTPS failed, fallback to HTTP. + logrus.Debugf("Error from registry %q marked as insecure: %v. Insecurely falling back to HTTP", endpoint, err) + endpoint.URL.Scheme = "http" + + var err2 error + if _, err2 = endpoint.Ping(); err2 == nil { + return nil + } + + return fmt.Errorf("invalid registry endpoint %q. HTTPS attempt: %v. HTTP attempt: %v", endpoint, err, err2) + } + + return nil +} + +func newV1Endpoint(address url.URL, tlsConfig *tls.Config, userAgent string, metaHeaders http.Header) *V1Endpoint { + endpoint := &V1Endpoint{ + IsSecure: tlsConfig == nil || !tlsConfig.InsecureSkipVerify, + URL: new(url.URL), + } + + *endpoint.URL = address + + // TODO(tiborvass): make sure a ConnectTimeout transport is used + tr := NewTransport(tlsConfig) + endpoint.client = HTTPClient(transport.NewTransport(tr, Headers(userAgent, metaHeaders)...)) + return endpoint +} + +// trimV1Address trims the version off the address and returns the +// trimmed address or an error if there is a non-V1 version. +func trimV1Address(address string) (string, error) { + var ( + chunks []string + apiVersionStr string + ) + + if strings.HasSuffix(address, "/") { + address = address[:len(address)-1] + } + + chunks = strings.Split(address, "/") + apiVersionStr = chunks[len(chunks)-1] + if apiVersionStr == "v1" { + return strings.Join(chunks[:len(chunks)-1], "/"), nil + } + + for k, v := range apiVersions { + if k != APIVersion1 && apiVersionStr == v { + return "", fmt.Errorf("unsupported V1 version path %s", apiVersionStr) + } + } + + return address, nil +} + +func newV1EndpointFromStr(address string, tlsConfig *tls.Config, userAgent string, metaHeaders http.Header) (*V1Endpoint, error) { + if !strings.HasPrefix(address, "http://") && !strings.HasPrefix(address, "https://") { + address = "https://" + address + } + + address, err := trimV1Address(address) + if err != nil { + return nil, err + } + + uri, err := url.Parse(address) + if err != nil { + return nil, err + } + + endpoint := newV1Endpoint(*uri, tlsConfig, userAgent, metaHeaders) + if err != nil { + return nil, err + } + + return endpoint, nil +} + +// Get the formatted URL for the root of this registry Endpoint +func (e *V1Endpoint) String() string { + return e.URL.String() + "/v1/" +} + +// Path returns a formatted string for the URL +// of this endpoint with the given path appended. +func (e *V1Endpoint) Path(path string) string { + return e.URL.String() + "/v1/" + path +} + +// Ping returns a PingResult which indicates whether the registry is standalone or not. +func (e *V1Endpoint) Ping() (PingResult, error) { + logrus.Debugf("attempting v1 ping for registry endpoint %s", e) + + if e.String() == IndexServer { + // Skip the check, we know this one is valid + // (and we never want to fallback to http in case of error) + return PingResult{Standalone: false}, nil + } + + req, err := http.NewRequest("GET", e.Path("_ping"), nil) + if err != nil { + return PingResult{Standalone: false}, err + } + + resp, err := e.client.Do(req) + if err != nil { + return PingResult{Standalone: false}, err + } + + defer resp.Body.Close() + + jsonString, err := ioutil.ReadAll(resp.Body) + if err != nil { + return PingResult{Standalone: false}, fmt.Errorf("error while reading the http response: %s", err) + } + + // If the header is absent, we assume true for compatibility with earlier + // versions of the registry. default to true + info := PingResult{ + Standalone: true, + } + if err := json.Unmarshal(jsonString, &info); err != nil { + logrus.Debugf("Error unmarshaling the _ping PingResult: %s", err) + // don't stop here. Just assume sane defaults + } + if hdr := resp.Header.Get("X-Docker-Registry-Version"); hdr != "" { + logrus.Debugf("Registry version header: '%s'", hdr) + info.Version = hdr + } + logrus.Debugf("PingResult.Version: %q", info.Version) + + standalone := resp.Header.Get("X-Docker-Registry-Standalone") + logrus.Debugf("Registry standalone header: '%s'", standalone) + // Accepted values are "true" (case-insensitive) and "1". + if strings.EqualFold(standalone, "true") || standalone == "1" { + info.Standalone = true + } else if len(standalone) > 0 { + // there is a header set, and it is not "true" or "1", so assume fails + info.Standalone = false + } + logrus.Debugf("PingResult.Standalone: %t", info.Standalone) + return info, nil +} diff --git a/vendor/github.com/docker/docker/registry/errors.go b/vendor/github.com/docker/docker/registry/errors.go new file mode 100644 index 000000000..5bab02e5e --- /dev/null +++ b/vendor/github.com/docker/docker/registry/errors.go @@ -0,0 +1,31 @@ +package registry // import "github.com/docker/docker/registry" + +import ( + "net/url" + + "github.com/docker/distribution/registry/api/errcode" + "github.com/docker/docker/errdefs" +) + +type notFoundError string + +func (e notFoundError) Error() string { + return string(e) +} + +func (notFoundError) NotFound() {} + +func translateV2AuthError(err error) error { + switch e := err.(type) { + case *url.Error: + switch e2 := e.Err.(type) { + case errcode.Error: + switch e2.Code { + case errcode.ErrorCodeUnauthorized: + return errdefs.Unauthorized(err) + } + } + } + + return err +} diff --git a/vendor/github.com/docker/docker/registry/registry.go b/vendor/github.com/docker/docker/registry/registry.go new file mode 100644 index 000000000..7a84bbfb7 --- /dev/null +++ b/vendor/github.com/docker/docker/registry/registry.go @@ -0,0 +1,191 @@ +// Package registry contains client primitives to interact with a remote Docker registry. +package registry // import "github.com/docker/docker/registry" + +import ( + "crypto/tls" + "errors" + "fmt" + "io/ioutil" + "net" + "net/http" + "os" + "path/filepath" + "strings" + "time" + + "github.com/docker/distribution/registry/client/transport" + "github.com/docker/go-connections/sockets" + "github.com/docker/go-connections/tlsconfig" + "github.com/sirupsen/logrus" +) + +var ( + // ErrAlreadyExists is an error returned if an image being pushed + // already exists on the remote side + ErrAlreadyExists = errors.New("Image already exists") +) + +func newTLSConfig(hostname string, isSecure bool) (*tls.Config, error) { + // PreferredServerCipherSuites should have no effect + tlsConfig := tlsconfig.ServerDefault() + + tlsConfig.InsecureSkipVerify = !isSecure + + if isSecure && CertsDir != "" { + hostDir := filepath.Join(CertsDir, cleanPath(hostname)) + logrus.Debugf("hostDir: %s", hostDir) + if err := ReadCertsDirectory(tlsConfig, hostDir); err != nil { + return nil, err + } + } + + return tlsConfig, nil +} + +func hasFile(files []os.FileInfo, name string) bool { + for _, f := range files { + if f.Name() == name { + return true + } + } + return false +} + +// ReadCertsDirectory reads the directory for TLS certificates +// including roots and certificate pairs and updates the +// provided TLS configuration. +func ReadCertsDirectory(tlsConfig *tls.Config, directory string) error { + fs, err := ioutil.ReadDir(directory) + if err != nil && !os.IsNotExist(err) { + return err + } + + for _, f := range fs { + if strings.HasSuffix(f.Name(), ".crt") { + if tlsConfig.RootCAs == nil { + systemPool, err := tlsconfig.SystemCertPool() + if err != nil { + return fmt.Errorf("unable to get system cert pool: %v", err) + } + tlsConfig.RootCAs = systemPool + } + logrus.Debugf("crt: %s", filepath.Join(directory, f.Name())) + data, err := ioutil.ReadFile(filepath.Join(directory, f.Name())) + if err != nil { + return err + } + tlsConfig.RootCAs.AppendCertsFromPEM(data) + } + if strings.HasSuffix(f.Name(), ".cert") { + certName := f.Name() + keyName := certName[:len(certName)-5] + ".key" + logrus.Debugf("cert: %s", filepath.Join(directory, f.Name())) + if !hasFile(fs, keyName) { + return fmt.Errorf("missing key %s for client certificate %s. Note that CA certificates should use the extension .crt", keyName, certName) + } + cert, err := tls.LoadX509KeyPair(filepath.Join(directory, certName), filepath.Join(directory, keyName)) + if err != nil { + return err + } + tlsConfig.Certificates = append(tlsConfig.Certificates, cert) + } + if strings.HasSuffix(f.Name(), ".key") { + keyName := f.Name() + certName := keyName[:len(keyName)-4] + ".cert" + logrus.Debugf("key: %s", filepath.Join(directory, f.Name())) + if !hasFile(fs, certName) { + return fmt.Errorf("Missing client certificate %s for key %s", certName, keyName) + } + } + } + + return nil +} + +// Headers returns request modifiers with a User-Agent and metaHeaders +func Headers(userAgent string, metaHeaders http.Header) []transport.RequestModifier { + modifiers := []transport.RequestModifier{} + if userAgent != "" { + modifiers = append(modifiers, transport.NewHeaderRequestModifier(http.Header{ + "User-Agent": []string{userAgent}, + })) + } + if metaHeaders != nil { + modifiers = append(modifiers, transport.NewHeaderRequestModifier(metaHeaders)) + } + return modifiers +} + +// HTTPClient returns an HTTP client structure which uses the given transport +// and contains the necessary headers for redirected requests +func HTTPClient(transport http.RoundTripper) *http.Client { + return &http.Client{ + Transport: transport, + CheckRedirect: addRequiredHeadersToRedirectedRequests, + } +} + +func trustedLocation(req *http.Request) bool { + var ( + trusteds = []string{"docker.com", "docker.io"} + hostname = strings.SplitN(req.Host, ":", 2)[0] + ) + if req.URL.Scheme != "https" { + return false + } + + for _, trusted := range trusteds { + if hostname == trusted || strings.HasSuffix(hostname, "."+trusted) { + return true + } + } + return false +} + +// addRequiredHeadersToRedirectedRequests adds the necessary redirection headers +// for redirected requests +func addRequiredHeadersToRedirectedRequests(req *http.Request, via []*http.Request) error { + if via != nil && via[0] != nil { + if trustedLocation(req) && trustedLocation(via[0]) { + req.Header = via[0].Header + return nil + } + for k, v := range via[0].Header { + if k != "Authorization" { + for _, vv := range v { + req.Header.Add(k, vv) + } + } + } + } + return nil +} + +// NewTransport returns a new HTTP transport. If tlsConfig is nil, it uses the +// default TLS configuration. +func NewTransport(tlsConfig *tls.Config) *http.Transport { + if tlsConfig == nil { + tlsConfig = tlsconfig.ServerDefault() + } + + direct := &net.Dialer{ + Timeout: 30 * time.Second, + KeepAlive: 30 * time.Second, + DualStack: true, + } + + base := &http.Transport{ + Proxy: http.ProxyFromEnvironment, + Dial: direct.Dial, + TLSHandshakeTimeout: 10 * time.Second, + TLSClientConfig: tlsConfig, + // TODO(dmcgowan): Call close idle connections when complete and use keep alive + DisableKeepAlives: true, + } + + proxyDialer, err := sockets.DialerFromEnvironment(direct) + if err == nil { + base.Dial = proxyDialer.Dial + } + return base +} diff --git a/vendor/github.com/docker/docker/registry/resumable/resumablerequestreader.go b/vendor/github.com/docker/docker/registry/resumable/resumablerequestreader.go new file mode 100644 index 000000000..8e97a1a4d --- /dev/null +++ b/vendor/github.com/docker/docker/registry/resumable/resumablerequestreader.go @@ -0,0 +1,96 @@ +package resumable // import "github.com/docker/docker/registry/resumable" + +import ( + "fmt" + "io" + "net/http" + "time" + + "github.com/sirupsen/logrus" +) + +type requestReader struct { + client *http.Client + request *http.Request + lastRange int64 + totalSize int64 + currentResponse *http.Response + failures uint32 + maxFailures uint32 + waitDuration time.Duration +} + +// NewRequestReader makes it possible to resume reading a request's body transparently +// maxfail is the number of times we retry to make requests again (not resumes) +// totalsize is the total length of the body; auto detect if not provided +func NewRequestReader(c *http.Client, r *http.Request, maxfail uint32, totalsize int64) io.ReadCloser { + return &requestReader{client: c, request: r, maxFailures: maxfail, totalSize: totalsize, waitDuration: 5 * time.Second} +} + +// NewRequestReaderWithInitialResponse makes it possible to resume +// reading the body of an already initiated request. +func NewRequestReaderWithInitialResponse(c *http.Client, r *http.Request, maxfail uint32, totalsize int64, initialResponse *http.Response) io.ReadCloser { + return &requestReader{client: c, request: r, maxFailures: maxfail, totalSize: totalsize, currentResponse: initialResponse, waitDuration: 5 * time.Second} +} + +func (r *requestReader) Read(p []byte) (n int, err error) { + if r.client == nil || r.request == nil { + return 0, fmt.Errorf("client and request can't be nil") + } + isFreshRequest := false + if r.lastRange != 0 && r.currentResponse == nil { + readRange := fmt.Sprintf("bytes=%d-%d", r.lastRange, r.totalSize) + r.request.Header.Set("Range", readRange) + time.Sleep(r.waitDuration) + } + if r.currentResponse == nil { + r.currentResponse, err = r.client.Do(r.request) + isFreshRequest = true + } + if err != nil && r.failures+1 != r.maxFailures { + r.cleanUpResponse() + r.failures++ + time.Sleep(time.Duration(r.failures) * r.waitDuration) + return 0, nil + } else if err != nil { + r.cleanUpResponse() + return 0, err + } + if r.currentResponse.StatusCode == 416 && r.lastRange == r.totalSize && r.currentResponse.ContentLength == 0 { + r.cleanUpResponse() + return 0, io.EOF + } else if r.currentResponse.StatusCode != 206 && r.lastRange != 0 && isFreshRequest { + r.cleanUpResponse() + return 0, fmt.Errorf("the server doesn't support byte ranges") + } + if r.totalSize == 0 { + r.totalSize = r.currentResponse.ContentLength + } else if r.totalSize <= 0 { + r.cleanUpResponse() + return 0, fmt.Errorf("failed to auto detect content length") + } + n, err = r.currentResponse.Body.Read(p) + r.lastRange += int64(n) + if err != nil { + r.cleanUpResponse() + } + if err != nil && err != io.EOF { + logrus.Infof("encountered error during pull and clearing it before resume: %s", err) + err = nil + } + return n, err +} + +func (r *requestReader) Close() error { + r.cleanUpResponse() + r.client = nil + r.request = nil + return nil +} + +func (r *requestReader) cleanUpResponse() { + if r.currentResponse != nil { + r.currentResponse.Body.Close() + r.currentResponse = nil + } +} diff --git a/vendor/github.com/docker/docker/registry/service.go b/vendor/github.com/docker/docker/registry/service.go index b59e29065..b441970ff 100644 --- a/vendor/github.com/docker/docker/registry/service.go +++ b/vendor/github.com/docker/docker/registry/service.go @@ -1,6 +1,151 @@ -package registry +package registry // import "github.com/docker/docker/registry" -import "strings" +import ( + "context" + "crypto/tls" + "net/http" + "net/url" + "strings" + "sync" + + "github.com/docker/distribution/reference" + "github.com/docker/distribution/registry/client/auth" + "github.com/docker/docker/api/types" + registrytypes "github.com/docker/docker/api/types/registry" + "github.com/docker/docker/errdefs" + "github.com/pkg/errors" + "github.com/sirupsen/logrus" +) + +const ( + // DefaultSearchLimit is the default value for maximum number of returned search results. + DefaultSearchLimit = 25 +) + +// Service is the interface defining what a registry service should implement. +type Service interface { + Auth(ctx context.Context, authConfig *types.AuthConfig, userAgent string) (status, token string, err error) + LookupPullEndpoints(hostname string) (endpoints []APIEndpoint, err error) + LookupPushEndpoints(hostname string) (endpoints []APIEndpoint, err error) + ResolveRepository(name reference.Named) (*RepositoryInfo, error) + Search(ctx context.Context, term string, limit int, authConfig *types.AuthConfig, userAgent string, headers map[string][]string) (*registrytypes.SearchResults, error) + ServiceConfig() *registrytypes.ServiceConfig + TLSConfig(hostname string) (*tls.Config, error) + LoadAllowNondistributableArtifacts([]string) error + LoadMirrors([]string) error + LoadInsecureRegistries([]string) error +} + +// DefaultService is a registry service. It tracks configuration data such as a list +// of mirrors. +type DefaultService struct { + config *serviceConfig + mu sync.Mutex +} + +// NewService returns a new instance of DefaultService ready to be +// installed into an engine. +func NewService(options ServiceOptions) (*DefaultService, error) { + config, err := newServiceConfig(options) + + return &DefaultService{config: config}, err +} + +// ServiceConfig returns the public registry service configuration. +func (s *DefaultService) ServiceConfig() *registrytypes.ServiceConfig { + s.mu.Lock() + defer s.mu.Unlock() + + servConfig := registrytypes.ServiceConfig{ + AllowNondistributableArtifactsCIDRs: make([]*(registrytypes.NetIPNet), 0), + AllowNondistributableArtifactsHostnames: make([]string, 0), + InsecureRegistryCIDRs: make([]*(registrytypes.NetIPNet), 0), + IndexConfigs: make(map[string]*(registrytypes.IndexInfo)), + Mirrors: make([]string, 0), + } + + // construct a new ServiceConfig which will not retrieve s.Config directly, + // and look up items in s.config with mu locked + servConfig.AllowNondistributableArtifactsCIDRs = append(servConfig.AllowNondistributableArtifactsCIDRs, s.config.ServiceConfig.AllowNondistributableArtifactsCIDRs...) + servConfig.AllowNondistributableArtifactsHostnames = append(servConfig.AllowNondistributableArtifactsHostnames, s.config.ServiceConfig.AllowNondistributableArtifactsHostnames...) + servConfig.InsecureRegistryCIDRs = append(servConfig.InsecureRegistryCIDRs, s.config.ServiceConfig.InsecureRegistryCIDRs...) + + for key, value := range s.config.ServiceConfig.IndexConfigs { + servConfig.IndexConfigs[key] = value + } + + servConfig.Mirrors = append(servConfig.Mirrors, s.config.ServiceConfig.Mirrors...) + + return &servConfig +} + +// LoadAllowNondistributableArtifacts loads allow-nondistributable-artifacts registries for Service. +func (s *DefaultService) LoadAllowNondistributableArtifacts(registries []string) error { + s.mu.Lock() + defer s.mu.Unlock() + + return s.config.LoadAllowNondistributableArtifacts(registries) +} + +// LoadMirrors loads registry mirrors for Service +func (s *DefaultService) LoadMirrors(mirrors []string) error { + s.mu.Lock() + defer s.mu.Unlock() + + return s.config.LoadMirrors(mirrors) +} + +// LoadInsecureRegistries loads insecure registries for Service +func (s *DefaultService) LoadInsecureRegistries(registries []string) error { + s.mu.Lock() + defer s.mu.Unlock() + + return s.config.LoadInsecureRegistries(registries) +} + +// Auth contacts the public registry with the provided credentials, +// and returns OK if authentication was successful. +// It can be used to verify the validity of a client's credentials. +func (s *DefaultService) Auth(ctx context.Context, authConfig *types.AuthConfig, userAgent string) (status, token string, err error) { + // TODO Use ctx when searching for repositories + serverAddress := authConfig.ServerAddress + if serverAddress == "" { + serverAddress = IndexServer + } + if !strings.HasPrefix(serverAddress, "https://") && !strings.HasPrefix(serverAddress, "http://") { + serverAddress = "https://" + serverAddress + } + u, err := url.Parse(serverAddress) + if err != nil { + return "", "", errdefs.InvalidParameter(errors.Errorf("unable to parse server address: %v", err)) + } + + endpoints, err := s.LookupPushEndpoints(u.Host) + if err != nil { + return "", "", errdefs.InvalidParameter(err) + } + + for _, endpoint := range endpoints { + login := loginV2 + if endpoint.Version == APIVersion1 { + login = loginV1 + } + + status, token, err = login(authConfig, endpoint, userAgent) + if err == nil { + return + } + if fErr, ok := err.(fallbackError); ok { + err = fErr.err + logrus.Infof("Error logging in to %s endpoint, trying next endpoint: %v", endpoint.Version, err) + continue + } + + return "", "", err + } + + return "", "", err +} // splitReposSearchTerm breaks a search term into an index name and remote name func splitReposSearchTerm(reposName string) (string, string) { @@ -18,3 +163,166 @@ func splitReposSearchTerm(reposName string) (string, string) { } return indexName, remoteName } + +// Search queries the public registry for images matching the specified +// search terms, and returns the results. +func (s *DefaultService) Search(ctx context.Context, term string, limit int, authConfig *types.AuthConfig, userAgent string, headers map[string][]string) (*registrytypes.SearchResults, error) { + // TODO Use ctx when searching for repositories + if err := validateNoScheme(term); err != nil { + return nil, err + } + + indexName, remoteName := splitReposSearchTerm(term) + + // Search is a long-running operation, just lock s.config to avoid block others. + s.mu.Lock() + index, err := newIndexInfo(s.config, indexName) + s.mu.Unlock() + + if err != nil { + return nil, err + } + + // *TODO: Search multiple indexes. + endpoint, err := NewV1Endpoint(index, userAgent, http.Header(headers)) + if err != nil { + return nil, err + } + + var client *http.Client + if authConfig != nil && authConfig.IdentityToken != "" && authConfig.Username != "" { + creds := NewStaticCredentialStore(authConfig) + scopes := []auth.Scope{ + auth.RegistryScope{ + Name: "catalog", + Actions: []string{"search"}, + }, + } + + modifiers := Headers(userAgent, nil) + v2Client, foundV2, err := v2AuthHTTPClient(endpoint.URL, endpoint.client.Transport, modifiers, creds, scopes) + if err != nil { + if fErr, ok := err.(fallbackError); ok { + logrus.Errorf("Cannot use identity token for search, v2 auth not supported: %v", fErr.err) + } else { + return nil, err + } + } else if foundV2 { + // Copy non transport http client features + v2Client.Timeout = endpoint.client.Timeout + v2Client.CheckRedirect = endpoint.client.CheckRedirect + v2Client.Jar = endpoint.client.Jar + + logrus.Debugf("using v2 client for search to %s", endpoint.URL) + client = v2Client + } + } + + if client == nil { + client = endpoint.client + if err := authorizeClient(client, authConfig, endpoint); err != nil { + return nil, err + } + } + + r := newSession(client, authConfig, endpoint) + + if index.Official { + localName := remoteName + if strings.HasPrefix(localName, "library/") { + // If pull "library/foo", it's stored locally under "foo" + localName = strings.SplitN(localName, "/", 2)[1] + } + + return r.SearchRepositories(localName, limit) + } + return r.SearchRepositories(remoteName, limit) +} + +// ResolveRepository splits a repository name into its components +// and configuration of the associated registry. +func (s *DefaultService) ResolveRepository(name reference.Named) (*RepositoryInfo, error) { + s.mu.Lock() + defer s.mu.Unlock() + return newRepositoryInfo(s.config, name) +} + +// APIEndpoint represents a remote API endpoint +type APIEndpoint struct { + Mirror bool + URL *url.URL + Version APIVersion + AllowNondistributableArtifacts bool + Official bool + TrimHostname bool + TLSConfig *tls.Config +} + +// ToV1Endpoint returns a V1 API endpoint based on the APIEndpoint +func (e APIEndpoint) ToV1Endpoint(userAgent string, metaHeaders http.Header) *V1Endpoint { + return newV1Endpoint(*e.URL, e.TLSConfig, userAgent, metaHeaders) +} + +// TLSConfig constructs a client TLS configuration based on server defaults +func (s *DefaultService) TLSConfig(hostname string) (*tls.Config, error) { + s.mu.Lock() + defer s.mu.Unlock() + + return newTLSConfig(hostname, isSecureIndex(s.config, hostname)) +} + +// tlsConfig constructs a client TLS configuration based on server defaults +func (s *DefaultService) tlsConfig(hostname string) (*tls.Config, error) { + return newTLSConfig(hostname, isSecureIndex(s.config, hostname)) +} + +func (s *DefaultService) tlsConfigForMirror(mirrorURL *url.URL) (*tls.Config, error) { + return s.tlsConfig(mirrorURL.Host) +} + +// LookupPullEndpoints creates a list of endpoints to try to pull from, in order of preference. +// It gives preference to v2 endpoints over v1, mirrors over the actual +// registry, and HTTPS over plain HTTP. +func (s *DefaultService) LookupPullEndpoints(hostname string) (endpoints []APIEndpoint, err error) { + s.mu.Lock() + defer s.mu.Unlock() + + return s.lookupEndpoints(hostname) +} + +// LookupPushEndpoints creates a list of endpoints to try to push to, in order of preference. +// It gives preference to v2 endpoints over v1, and HTTPS over plain HTTP. +// Mirrors are not included. +func (s *DefaultService) LookupPushEndpoints(hostname string) (endpoints []APIEndpoint, err error) { + s.mu.Lock() + defer s.mu.Unlock() + + allEndpoints, err := s.lookupEndpoints(hostname) + if err == nil { + for _, endpoint := range allEndpoints { + if !endpoint.Mirror { + endpoints = append(endpoints, endpoint) + } + } + } + return endpoints, err +} + +func (s *DefaultService) lookupEndpoints(hostname string) (endpoints []APIEndpoint, err error) { + endpoints, err = s.lookupV2Endpoints(hostname) + if err != nil { + return nil, err + } + + if s.config.V2Only { + return endpoints, nil + } + + legacyEndpoints, err := s.lookupV1Endpoints(hostname) + if err != nil { + return nil, err + } + endpoints = append(endpoints, legacyEndpoints...) + + return endpoints, nil +} diff --git a/vendor/github.com/docker/docker/registry/service_v1.go b/vendor/github.com/docker/docker/registry/service_v1.go new file mode 100644 index 000000000..d955ec51f --- /dev/null +++ b/vendor/github.com/docker/docker/registry/service_v1.go @@ -0,0 +1,40 @@ +package registry // import "github.com/docker/docker/registry" + +import "net/url" + +func (s *DefaultService) lookupV1Endpoints(hostname string) (endpoints []APIEndpoint, err error) { + if hostname == DefaultNamespace || hostname == DefaultV2Registry.Host || hostname == IndexHostname { + return []APIEndpoint{}, nil + } + + tlsConfig, err := s.tlsConfig(hostname) + if err != nil { + return nil, err + } + + endpoints = []APIEndpoint{ + { + URL: &url.URL{ + Scheme: "https", + Host: hostname, + }, + Version: APIVersion1, + TrimHostname: true, + TLSConfig: tlsConfig, + }, + } + + if tlsConfig.InsecureSkipVerify { + endpoints = append(endpoints, APIEndpoint{ // or this + URL: &url.URL{ + Scheme: "http", + Host: hostname, + }, + Version: APIVersion1, + TrimHostname: true, + // used to check if supposed to be secure via InsecureSkipVerify + TLSConfig: tlsConfig, + }) + } + return endpoints, nil +} diff --git a/vendor/github.com/docker/docker/registry/service_v2.go b/vendor/github.com/docker/docker/registry/service_v2.go new file mode 100644 index 000000000..3a56dc911 --- /dev/null +++ b/vendor/github.com/docker/docker/registry/service_v2.go @@ -0,0 +1,82 @@ +package registry // import "github.com/docker/docker/registry" + +import ( + "net/url" + "strings" + + "github.com/docker/go-connections/tlsconfig" +) + +func (s *DefaultService) lookupV2Endpoints(hostname string) (endpoints []APIEndpoint, err error) { + tlsConfig := tlsconfig.ServerDefault() + if hostname == DefaultNamespace || hostname == IndexHostname { + // v2 mirrors + for _, mirror := range s.config.Mirrors { + if !strings.HasPrefix(mirror, "http://") && !strings.HasPrefix(mirror, "https://") { + mirror = "https://" + mirror + } + mirrorURL, err := url.Parse(mirror) + if err != nil { + return nil, err + } + mirrorTLSConfig, err := s.tlsConfigForMirror(mirrorURL) + if err != nil { + return nil, err + } + endpoints = append(endpoints, APIEndpoint{ + URL: mirrorURL, + // guess mirrors are v2 + Version: APIVersion2, + Mirror: true, + TrimHostname: true, + TLSConfig: mirrorTLSConfig, + }) + } + // v2 registry + endpoints = append(endpoints, APIEndpoint{ + URL: DefaultV2Registry, + Version: APIVersion2, + Official: true, + TrimHostname: true, + TLSConfig: tlsConfig, + }) + + return endpoints, nil + } + + ana := allowNondistributableArtifacts(s.config, hostname) + + tlsConfig, err = s.tlsConfig(hostname) + if err != nil { + return nil, err + } + + endpoints = []APIEndpoint{ + { + URL: &url.URL{ + Scheme: "https", + Host: hostname, + }, + Version: APIVersion2, + AllowNondistributableArtifacts: ana, + TrimHostname: true, + TLSConfig: tlsConfig, + }, + } + + if tlsConfig.InsecureSkipVerify { + endpoints = append(endpoints, APIEndpoint{ + URL: &url.URL{ + Scheme: "http", + Host: hostname, + }, + Version: APIVersion2, + AllowNondistributableArtifacts: ana, + TrimHostname: true, + // used to check if supposed to be secure via InsecureSkipVerify + TLSConfig: tlsConfig, + }) + } + + return endpoints, nil +} diff --git a/vendor/github.com/docker/docker/registry/session.go b/vendor/github.com/docker/docker/registry/session.go new file mode 100644 index 000000000..ef1429959 --- /dev/null +++ b/vendor/github.com/docker/docker/registry/session.go @@ -0,0 +1,779 @@ +package registry // import "github.com/docker/docker/registry" + +import ( + "bytes" + "crypto/sha256" + // this is required for some certificates + _ "crypto/sha512" + "encoding/hex" + "encoding/json" + "fmt" + "io" + "io/ioutil" + "net/http" + "net/http/cookiejar" + "net/url" + "strconv" + "strings" + "sync" + + "github.com/docker/distribution/reference" + "github.com/docker/distribution/registry/api/errcode" + "github.com/docker/docker/api/types" + registrytypes "github.com/docker/docker/api/types/registry" + "github.com/docker/docker/errdefs" + "github.com/docker/docker/pkg/ioutils" + "github.com/docker/docker/pkg/jsonmessage" + "github.com/docker/docker/pkg/stringid" + "github.com/docker/docker/pkg/tarsum" + "github.com/docker/docker/registry/resumable" + "github.com/pkg/errors" + "github.com/sirupsen/logrus" +) + +var ( + // ErrRepoNotFound is returned if the repository didn't exist on the + // remote side + ErrRepoNotFound notFoundError = "Repository not found" +) + +// A Session is used to communicate with a V1 registry +type Session struct { + indexEndpoint *V1Endpoint + client *http.Client + // TODO(tiborvass): remove authConfig + authConfig *types.AuthConfig + id string +} + +type authTransport struct { + http.RoundTripper + *types.AuthConfig + + alwaysSetBasicAuth bool + token []string + + mu sync.Mutex // guards modReq + modReq map[*http.Request]*http.Request // original -> modified +} + +// AuthTransport handles the auth layer when communicating with a v1 registry (private or official) +// +// For private v1 registries, set alwaysSetBasicAuth to true. +// +// For the official v1 registry, if there isn't already an Authorization header in the request, +// but there is an X-Docker-Token header set to true, then Basic Auth will be used to set the Authorization header. +// After sending the request with the provided base http.RoundTripper, if an X-Docker-Token header, representing +// a token, is present in the response, then it gets cached and sent in the Authorization header of all subsequent +// requests. +// +// If the server sends a token without the client having requested it, it is ignored. +// +// This RoundTripper also has a CancelRequest method important for correct timeout handling. +func AuthTransport(base http.RoundTripper, authConfig *types.AuthConfig, alwaysSetBasicAuth bool) http.RoundTripper { + if base == nil { + base = http.DefaultTransport + } + return &authTransport{ + RoundTripper: base, + AuthConfig: authConfig, + alwaysSetBasicAuth: alwaysSetBasicAuth, + modReq: make(map[*http.Request]*http.Request), + } +} + +// cloneRequest returns a clone of the provided *http.Request. +// The clone is a shallow copy of the struct and its Header map. +func cloneRequest(r *http.Request) *http.Request { + // shallow copy of the struct + r2 := new(http.Request) + *r2 = *r + // deep copy of the Header + r2.Header = make(http.Header, len(r.Header)) + for k, s := range r.Header { + r2.Header[k] = append([]string(nil), s...) + } + + return r2 +} + +// RoundTrip changes an HTTP request's headers to add the necessary +// authentication-related headers +func (tr *authTransport) RoundTrip(orig *http.Request) (*http.Response, error) { + // Authorization should not be set on 302 redirect for untrusted locations. + // This logic mirrors the behavior in addRequiredHeadersToRedirectedRequests. + // As the authorization logic is currently implemented in RoundTrip, + // a 302 redirect is detected by looking at the Referrer header as go http package adds said header. + // This is safe as Docker doesn't set Referrer in other scenarios. + if orig.Header.Get("Referer") != "" && !trustedLocation(orig) { + return tr.RoundTripper.RoundTrip(orig) + } + + req := cloneRequest(orig) + tr.mu.Lock() + tr.modReq[orig] = req + tr.mu.Unlock() + + if tr.alwaysSetBasicAuth { + if tr.AuthConfig == nil { + return nil, errors.New("unexpected error: empty auth config") + } + req.SetBasicAuth(tr.Username, tr.Password) + return tr.RoundTripper.RoundTrip(req) + } + + // Don't override + if req.Header.Get("Authorization") == "" { + if req.Header.Get("X-Docker-Token") == "true" && tr.AuthConfig != nil && len(tr.Username) > 0 { + req.SetBasicAuth(tr.Username, tr.Password) + } else if len(tr.token) > 0 { + req.Header.Set("Authorization", "Token "+strings.Join(tr.token, ",")) + } + } + resp, err := tr.RoundTripper.RoundTrip(req) + if err != nil { + delete(tr.modReq, orig) + return nil, err + } + if len(resp.Header["X-Docker-Token"]) > 0 { + tr.token = resp.Header["X-Docker-Token"] + } + resp.Body = &ioutils.OnEOFReader{ + Rc: resp.Body, + Fn: func() { + tr.mu.Lock() + delete(tr.modReq, orig) + tr.mu.Unlock() + }, + } + return resp, nil +} + +// CancelRequest cancels an in-flight request by closing its connection. +func (tr *authTransport) CancelRequest(req *http.Request) { + type canceler interface { + CancelRequest(*http.Request) + } + if cr, ok := tr.RoundTripper.(canceler); ok { + tr.mu.Lock() + modReq := tr.modReq[req] + delete(tr.modReq, req) + tr.mu.Unlock() + cr.CancelRequest(modReq) + } +} + +func authorizeClient(client *http.Client, authConfig *types.AuthConfig, endpoint *V1Endpoint) error { + var alwaysSetBasicAuth bool + + // If we're working with a standalone private registry over HTTPS, send Basic Auth headers + // alongside all our requests. + if endpoint.String() != IndexServer && endpoint.URL.Scheme == "https" { + info, err := endpoint.Ping() + if err != nil { + return err + } + if info.Standalone && authConfig != nil { + logrus.Debugf("Endpoint %s is eligible for private registry. Enabling decorator.", endpoint.String()) + alwaysSetBasicAuth = true + } + } + + // Annotate the transport unconditionally so that v2 can + // properly fallback on v1 when an image is not found. + client.Transport = AuthTransport(client.Transport, authConfig, alwaysSetBasicAuth) + + jar, err := cookiejar.New(nil) + if err != nil { + return errors.New("cookiejar.New is not supposed to return an error") + } + client.Jar = jar + + return nil +} + +func newSession(client *http.Client, authConfig *types.AuthConfig, endpoint *V1Endpoint) *Session { + return &Session{ + authConfig: authConfig, + client: client, + indexEndpoint: endpoint, + id: stringid.GenerateRandomID(), + } +} + +// NewSession creates a new session +// TODO(tiborvass): remove authConfig param once registry client v2 is vendored +func NewSession(client *http.Client, authConfig *types.AuthConfig, endpoint *V1Endpoint) (*Session, error) { + if err := authorizeClient(client, authConfig, endpoint); err != nil { + return nil, err + } + + return newSession(client, authConfig, endpoint), nil +} + +// ID returns this registry session's ID. +func (r *Session) ID() string { + return r.id +} + +// GetRemoteHistory retrieves the history of a given image from the registry. +// It returns a list of the parent's JSON files (including the requested image). +func (r *Session) GetRemoteHistory(imgID, registry string) ([]string, error) { + res, err := r.client.Get(registry + "images/" + imgID + "/ancestry") + if err != nil { + return nil, err + } + defer res.Body.Close() + if res.StatusCode != 200 { + if res.StatusCode == 401 { + return nil, errcode.ErrorCodeUnauthorized.WithArgs() + } + return nil, newJSONError(fmt.Sprintf("Server error: %d trying to fetch remote history for %s", res.StatusCode, imgID), res) + } + + var history []string + if err := json.NewDecoder(res.Body).Decode(&history); err != nil { + return nil, fmt.Errorf("Error while reading the http response: %v", err) + } + + logrus.Debugf("Ancestry: %v", history) + return history, nil +} + +// LookupRemoteImage checks if an image exists in the registry +func (r *Session) LookupRemoteImage(imgID, registry string) error { + res, err := r.client.Get(registry + "images/" + imgID + "/json") + if err != nil { + return err + } + res.Body.Close() + if res.StatusCode != 200 { + return newJSONError(fmt.Sprintf("HTTP code %d", res.StatusCode), res) + } + return nil +} + +// GetRemoteImageJSON retrieves an image's JSON metadata from the registry. +func (r *Session) GetRemoteImageJSON(imgID, registry string) ([]byte, int64, error) { + res, err := r.client.Get(registry + "images/" + imgID + "/json") + if err != nil { + return nil, -1, fmt.Errorf("Failed to download json: %s", err) + } + defer res.Body.Close() + if res.StatusCode != 200 { + return nil, -1, newJSONError(fmt.Sprintf("HTTP code %d", res.StatusCode), res) + } + // if the size header is not present, then set it to '-1' + imageSize := int64(-1) + if hdr := res.Header.Get("X-Docker-Size"); hdr != "" { + imageSize, err = strconv.ParseInt(hdr, 10, 64) + if err != nil { + return nil, -1, err + } + } + + jsonString, err := ioutil.ReadAll(res.Body) + if err != nil { + return nil, -1, fmt.Errorf("Failed to parse downloaded json: %v (%s)", err, jsonString) + } + return jsonString, imageSize, nil +} + +// GetRemoteImageLayer retrieves an image layer from the registry +func (r *Session) GetRemoteImageLayer(imgID, registry string, imgSize int64) (io.ReadCloser, error) { + var ( + statusCode = 0 + res *http.Response + err error + imageURL = fmt.Sprintf("%simages/%s/layer", registry, imgID) + ) + + req, err := http.NewRequest("GET", imageURL, nil) + if err != nil { + return nil, fmt.Errorf("Error while getting from the server: %v", err) + } + + res, err = r.client.Do(req) + if err != nil { + logrus.Debugf("Error contacting registry %s: %v", registry, err) + // the only case err != nil && res != nil is https://golang.org/src/net/http/client.go#L515 + if res != nil { + if res.Body != nil { + res.Body.Close() + } + statusCode = res.StatusCode + } + return nil, fmt.Errorf("Server error: Status %d while fetching image layer (%s)", + statusCode, imgID) + } + + if res.StatusCode != 200 { + res.Body.Close() + return nil, fmt.Errorf("Server error: Status %d while fetching image layer (%s)", + res.StatusCode, imgID) + } + + if res.Header.Get("Accept-Ranges") == "bytes" && imgSize > 0 { + logrus.Debug("server supports resume") + return resumable.NewRequestReaderWithInitialResponse(r.client, req, 5, imgSize, res), nil + } + logrus.Debug("server doesn't support resume") + return res.Body, nil +} + +// GetRemoteTag retrieves the tag named in the askedTag argument from the given +// repository. It queries each of the registries supplied in the registries +// argument, and returns data from the first one that answers the query +// successfully. +func (r *Session) GetRemoteTag(registries []string, repositoryRef reference.Named, askedTag string) (string, error) { + repository := reference.Path(repositoryRef) + + if strings.Count(repository, "/") == 0 { + // This will be removed once the registry supports auto-resolution on + // the "library" namespace + repository = "library/" + repository + } + for _, host := range registries { + endpoint := fmt.Sprintf("%srepositories/%s/tags/%s", host, repository, askedTag) + res, err := r.client.Get(endpoint) + if err != nil { + return "", err + } + + logrus.Debugf("Got status code %d from %s", res.StatusCode, endpoint) + defer res.Body.Close() + + if res.StatusCode == 404 { + return "", ErrRepoNotFound + } + if res.StatusCode != 200 { + continue + } + + var tagID string + if err := json.NewDecoder(res.Body).Decode(&tagID); err != nil { + return "", err + } + return tagID, nil + } + return "", fmt.Errorf("Could not reach any registry endpoint") +} + +// GetRemoteTags retrieves all tags from the given repository. It queries each +// of the registries supplied in the registries argument, and returns data from +// the first one that answers the query successfully. It returns a map with +// tag names as the keys and image IDs as the values. +func (r *Session) GetRemoteTags(registries []string, repositoryRef reference.Named) (map[string]string, error) { + repository := reference.Path(repositoryRef) + + if strings.Count(repository, "/") == 0 { + // This will be removed once the registry supports auto-resolution on + // the "library" namespace + repository = "library/" + repository + } + for _, host := range registries { + endpoint := fmt.Sprintf("%srepositories/%s/tags", host, repository) + res, err := r.client.Get(endpoint) + if err != nil { + return nil, err + } + + logrus.Debugf("Got status code %d from %s", res.StatusCode, endpoint) + defer res.Body.Close() + + if res.StatusCode == 404 { + return nil, ErrRepoNotFound + } + if res.StatusCode != 200 { + continue + } + + result := make(map[string]string) + if err := json.NewDecoder(res.Body).Decode(&result); err != nil { + return nil, err + } + return result, nil + } + return nil, fmt.Errorf("Could not reach any registry endpoint") +} + +func buildEndpointsList(headers []string, indexEp string) ([]string, error) { + var endpoints []string + parsedURL, err := url.Parse(indexEp) + if err != nil { + return nil, err + } + var urlScheme = parsedURL.Scheme + // The registry's URL scheme has to match the Index' + for _, ep := range headers { + epList := strings.Split(ep, ",") + for _, epListElement := range epList { + endpoints = append( + endpoints, + fmt.Sprintf("%s://%s/v1/", urlScheme, strings.TrimSpace(epListElement))) + } + } + return endpoints, nil +} + +// GetRepositoryData returns lists of images and endpoints for the repository +func (r *Session) GetRepositoryData(name reference.Named) (*RepositoryData, error) { + repositoryTarget := fmt.Sprintf("%srepositories/%s/images", r.indexEndpoint.String(), reference.Path(name)) + + logrus.Debugf("[registry] Calling GET %s", repositoryTarget) + + req, err := http.NewRequest("GET", repositoryTarget, nil) + if err != nil { + return nil, err + } + // this will set basic auth in r.client.Transport and send cached X-Docker-Token headers for all subsequent requests + req.Header.Set("X-Docker-Token", "true") + res, err := r.client.Do(req) + if err != nil { + // check if the error is because of i/o timeout + // and return a non-obtuse error message for users + // "Get https://index.docker.io/v1/repositories/library/busybox/images: i/o timeout" + // was a top search on the docker user forum + if isTimeout(err) { + return nil, fmt.Errorf("network timed out while trying to connect to %s. You may want to check your internet connection or if you are behind a proxy", repositoryTarget) + } + return nil, fmt.Errorf("Error while pulling image: %v", err) + } + defer res.Body.Close() + if res.StatusCode == 401 { + return nil, errcode.ErrorCodeUnauthorized.WithArgs() + } + // TODO: Right now we're ignoring checksums in the response body. + // In the future, we need to use them to check image validity. + if res.StatusCode == 404 { + return nil, newJSONError(fmt.Sprintf("HTTP code: %d", res.StatusCode), res) + } else if res.StatusCode != 200 { + errBody, err := ioutil.ReadAll(res.Body) + if err != nil { + logrus.Debugf("Error reading response body: %s", err) + } + return nil, newJSONError(fmt.Sprintf("Error: Status %d trying to pull repository %s: %q", res.StatusCode, reference.Path(name), errBody), res) + } + + var endpoints []string + if res.Header.Get("X-Docker-Endpoints") != "" { + endpoints, err = buildEndpointsList(res.Header["X-Docker-Endpoints"], r.indexEndpoint.String()) + if err != nil { + return nil, err + } + } else { + // Assume the endpoint is on the same host + endpoints = append(endpoints, fmt.Sprintf("%s://%s/v1/", r.indexEndpoint.URL.Scheme, req.URL.Host)) + } + + remoteChecksums := []*ImgData{} + if err := json.NewDecoder(res.Body).Decode(&remoteChecksums); err != nil { + return nil, err + } + + // Forge a better object from the retrieved data + imgsData := make(map[string]*ImgData, len(remoteChecksums)) + for _, elem := range remoteChecksums { + imgsData[elem.ID] = elem + } + + return &RepositoryData{ + ImgList: imgsData, + Endpoints: endpoints, + }, nil +} + +// PushImageChecksumRegistry uploads checksums for an image +func (r *Session) PushImageChecksumRegistry(imgData *ImgData, registry string) error { + u := registry + "images/" + imgData.ID + "/checksum" + + logrus.Debugf("[registry] Calling PUT %s", u) + + req, err := http.NewRequest("PUT", u, nil) + if err != nil { + return err + } + req.Header.Set("X-Docker-Checksum", imgData.Checksum) + req.Header.Set("X-Docker-Checksum-Payload", imgData.ChecksumPayload) + + res, err := r.client.Do(req) + if err != nil { + return fmt.Errorf("Failed to upload metadata: %v", err) + } + defer res.Body.Close() + if len(res.Cookies()) > 0 { + r.client.Jar.SetCookies(req.URL, res.Cookies()) + } + if res.StatusCode != 200 { + errBody, err := ioutil.ReadAll(res.Body) + if err != nil { + return fmt.Errorf("HTTP code %d while uploading metadata and error when trying to parse response body: %s", res.StatusCode, err) + } + var jsonBody map[string]string + if err := json.Unmarshal(errBody, &jsonBody); err != nil { + errBody = []byte(err.Error()) + } else if jsonBody["error"] == "Image already exists" { + return ErrAlreadyExists + } + return fmt.Errorf("HTTP code %d while uploading metadata: %q", res.StatusCode, errBody) + } + return nil +} + +// PushImageJSONRegistry pushes JSON metadata for a local image to the registry +func (r *Session) PushImageJSONRegistry(imgData *ImgData, jsonRaw []byte, registry string) error { + + u := registry + "images/" + imgData.ID + "/json" + + logrus.Debugf("[registry] Calling PUT %s", u) + + req, err := http.NewRequest("PUT", u, bytes.NewReader(jsonRaw)) + if err != nil { + return err + } + req.Header.Add("Content-type", "application/json") + + res, err := r.client.Do(req) + if err != nil { + return fmt.Errorf("Failed to upload metadata: %s", err) + } + defer res.Body.Close() + if res.StatusCode == 401 && strings.HasPrefix(registry, "http://") { + return newJSONError("HTTP code 401, Docker will not send auth headers over HTTP.", res) + } + if res.StatusCode != 200 { + errBody, err := ioutil.ReadAll(res.Body) + if err != nil { + return newJSONError(fmt.Sprintf("HTTP code %d while uploading metadata and error when trying to parse response body: %s", res.StatusCode, err), res) + } + var jsonBody map[string]string + if err := json.Unmarshal(errBody, &jsonBody); err != nil { + errBody = []byte(err.Error()) + } else if jsonBody["error"] == "Image already exists" { + return ErrAlreadyExists + } + return newJSONError(fmt.Sprintf("HTTP code %d while uploading metadata: %q", res.StatusCode, errBody), res) + } + return nil +} + +// PushImageLayerRegistry sends the checksum of an image layer to the registry +func (r *Session) PushImageLayerRegistry(imgID string, layer io.Reader, registry string, jsonRaw []byte) (checksum string, checksumPayload string, err error) { + u := registry + "images/" + imgID + "/layer" + + logrus.Debugf("[registry] Calling PUT %s", u) + + tarsumLayer, err := tarsum.NewTarSum(layer, false, tarsum.Version0) + if err != nil { + return "", "", err + } + h := sha256.New() + h.Write(jsonRaw) + h.Write([]byte{'\n'}) + checksumLayer := io.TeeReader(tarsumLayer, h) + + req, err := http.NewRequest("PUT", u, checksumLayer) + if err != nil { + return "", "", err + } + req.Header.Add("Content-Type", "application/octet-stream") + req.ContentLength = -1 + req.TransferEncoding = []string{"chunked"} + res, err := r.client.Do(req) + if err != nil { + return "", "", fmt.Errorf("Failed to upload layer: %v", err) + } + if rc, ok := layer.(io.Closer); ok { + if err := rc.Close(); err != nil { + return "", "", err + } + } + defer res.Body.Close() + + if res.StatusCode != 200 { + errBody, err := ioutil.ReadAll(res.Body) + if err != nil { + return "", "", newJSONError(fmt.Sprintf("HTTP code %d while uploading metadata and error when trying to parse response body: %s", res.StatusCode, err), res) + } + return "", "", newJSONError(fmt.Sprintf("Received HTTP code %d while uploading layer: %q", res.StatusCode, errBody), res) + } + + checksumPayload = "sha256:" + hex.EncodeToString(h.Sum(nil)) + return tarsumLayer.Sum(jsonRaw), checksumPayload, nil +} + +// PushRegistryTag pushes a tag on the registry. +// Remote has the format '/ +func (r *Session) PushRegistryTag(remote reference.Named, revision, tag, registry string) error { + // "jsonify" the string + revision = "\"" + revision + "\"" + path := fmt.Sprintf("repositories/%s/tags/%s", reference.Path(remote), tag) + + req, err := http.NewRequest("PUT", registry+path, strings.NewReader(revision)) + if err != nil { + return err + } + req.Header.Add("Content-type", "application/json") + req.ContentLength = int64(len(revision)) + res, err := r.client.Do(req) + if err != nil { + return err + } + res.Body.Close() + if res.StatusCode != 200 && res.StatusCode != 201 { + return newJSONError(fmt.Sprintf("Internal server error: %d trying to push tag %s on %s", res.StatusCode, tag, reference.Path(remote)), res) + } + return nil +} + +// PushImageJSONIndex uploads an image list to the repository +func (r *Session) PushImageJSONIndex(remote reference.Named, imgList []*ImgData, validate bool, regs []string) (*RepositoryData, error) { + cleanImgList := []*ImgData{} + if validate { + for _, elem := range imgList { + if elem.Checksum != "" { + cleanImgList = append(cleanImgList, elem) + } + } + } else { + cleanImgList = imgList + } + + imgListJSON, err := json.Marshal(cleanImgList) + if err != nil { + return nil, err + } + var suffix string + if validate { + suffix = "images" + } + u := fmt.Sprintf("%srepositories/%s/%s", r.indexEndpoint.String(), reference.Path(remote), suffix) + logrus.Debugf("[registry] PUT %s", u) + logrus.Debugf("Image list pushed to index:\n%s", imgListJSON) + headers := map[string][]string{ + "Content-type": {"application/json"}, + // this will set basic auth in r.client.Transport and send cached X-Docker-Token headers for all subsequent requests + "X-Docker-Token": {"true"}, + } + if validate { + headers["X-Docker-Endpoints"] = regs + } + + // Redirect if necessary + var res *http.Response + for { + if res, err = r.putImageRequest(u, headers, imgListJSON); err != nil { + return nil, err + } + if !shouldRedirect(res) { + break + } + res.Body.Close() + u = res.Header.Get("Location") + logrus.Debugf("Redirected to %s", u) + } + defer res.Body.Close() + + if res.StatusCode == 401 { + return nil, errcode.ErrorCodeUnauthorized.WithArgs() + } + + var tokens, endpoints []string + if !validate { + if res.StatusCode != 200 && res.StatusCode != 201 { + errBody, err := ioutil.ReadAll(res.Body) + if err != nil { + logrus.Debugf("Error reading response body: %s", err) + } + return nil, newJSONError(fmt.Sprintf("Error: Status %d trying to push repository %s: %q", res.StatusCode, reference.Path(remote), errBody), res) + } + tokens = res.Header["X-Docker-Token"] + logrus.Debugf("Auth token: %v", tokens) + + if res.Header.Get("X-Docker-Endpoints") == "" { + return nil, fmt.Errorf("Index response didn't contain any endpoints") + } + endpoints, err = buildEndpointsList(res.Header["X-Docker-Endpoints"], r.indexEndpoint.String()) + if err != nil { + return nil, err + } + } else { + if res.StatusCode != 204 { + errBody, err := ioutil.ReadAll(res.Body) + if err != nil { + logrus.Debugf("Error reading response body: %s", err) + } + return nil, newJSONError(fmt.Sprintf("Error: Status %d trying to push checksums %s: %q", res.StatusCode, reference.Path(remote), errBody), res) + } + } + + return &RepositoryData{ + Endpoints: endpoints, + }, nil +} + +func (r *Session) putImageRequest(u string, headers map[string][]string, body []byte) (*http.Response, error) { + req, err := http.NewRequest("PUT", u, bytes.NewReader(body)) + if err != nil { + return nil, err + } + req.ContentLength = int64(len(body)) + for k, v := range headers { + req.Header[k] = v + } + response, err := r.client.Do(req) + if err != nil { + return nil, err + } + return response, nil +} + +func shouldRedirect(response *http.Response) bool { + return response.StatusCode >= 300 && response.StatusCode < 400 +} + +// SearchRepositories performs a search against the remote repository +func (r *Session) SearchRepositories(term string, limit int) (*registrytypes.SearchResults, error) { + if limit < 1 || limit > 100 { + return nil, errdefs.InvalidParameter(errors.Errorf("Limit %d is outside the range of [1, 100]", limit)) + } + logrus.Debugf("Index server: %s", r.indexEndpoint) + u := r.indexEndpoint.String() + "search?q=" + url.QueryEscape(term) + "&n=" + url.QueryEscape(fmt.Sprintf("%d", limit)) + + req, err := http.NewRequest("GET", u, nil) + if err != nil { + return nil, errors.Wrap(errdefs.InvalidParameter(err), "Error building request") + } + // Have the AuthTransport send authentication, when logged in. + req.Header.Set("X-Docker-Token", "true") + res, err := r.client.Do(req) + if err != nil { + return nil, errdefs.System(err) + } + defer res.Body.Close() + if res.StatusCode != 200 { + return nil, newJSONError(fmt.Sprintf("Unexpected status code %d", res.StatusCode), res) + } + result := new(registrytypes.SearchResults) + return result, errors.Wrap(json.NewDecoder(res.Body).Decode(result), "error decoding registry search results") +} + +func isTimeout(err error) bool { + type timeout interface { + Timeout() bool + } + e := err + switch urlErr := err.(type) { + case *url.Error: + e = urlErr.Err + } + t, ok := e.(timeout) + return ok && t.Timeout() +} + +func newJSONError(msg string, res *http.Response) error { + return &jsonmessage.JSONError{ + Message: msg, + Code: res.StatusCode, + } +} diff --git a/vendor/github.com/docker/docker/registry/types.go b/vendor/github.com/docker/docker/registry/types.go index 8dbcb961f..28ed2bfa5 100644 --- a/vendor/github.com/docker/docker/registry/types.go +++ b/vendor/github.com/docker/docker/registry/types.go @@ -1,8 +1,8 @@ -package registry +package registry // import "github.com/docker/docker/registry" import ( + "github.com/docker/distribution/reference" registrytypes "github.com/docker/docker/api/types/registry" - "github.com/docker/docker/reference" ) // RepositoryData tracks the image list, list of endpoints for a repository @@ -57,7 +57,7 @@ var apiVersions = map[APIVersion]string{ // RepositoryInfo describes a repository type RepositoryInfo struct { - reference.Named + Name reference.Named // Index points to registry information Index *registrytypes.IndexInfo // Official indicates whether the repository is considered official. diff --git a/vendor/github.com/docker/go-connections/sockets/README.md b/vendor/github.com/docker/go-connections/sockets/README.md new file mode 100644 index 000000000..e69de29bb diff --git a/vendor/github.com/docker/go-connections/sockets/inmem_socket.go b/vendor/github.com/docker/go-connections/sockets/inmem_socket.go new file mode 100644 index 000000000..99846ffdd --- /dev/null +++ b/vendor/github.com/docker/go-connections/sockets/inmem_socket.go @@ -0,0 +1,81 @@ +package sockets + +import ( + "errors" + "net" + "sync" +) + +var errClosed = errors.New("use of closed network connection") + +// InmemSocket implements net.Listener using in-memory only connections. +type InmemSocket struct { + chConn chan net.Conn + chClose chan struct{} + addr string + mu sync.Mutex +} + +// dummyAddr is used to satisfy net.Addr for the in-mem socket +// it is just stored as a string and returns the string for all calls +type dummyAddr string + +// NewInmemSocket creates an in-memory only net.Listener +// The addr argument can be any string, but is used to satisfy the `Addr()` part +// of the net.Listener interface +func NewInmemSocket(addr string, bufSize int) *InmemSocket { + return &InmemSocket{ + chConn: make(chan net.Conn, bufSize), + chClose: make(chan struct{}), + addr: addr, + } +} + +// Addr returns the socket's addr string to satisfy net.Listener +func (s *InmemSocket) Addr() net.Addr { + return dummyAddr(s.addr) +} + +// Accept implements the Accept method in the Listener interface; it waits for the next call and returns a generic Conn. +func (s *InmemSocket) Accept() (net.Conn, error) { + select { + case conn := <-s.chConn: + return conn, nil + case <-s.chClose: + return nil, errClosed + } +} + +// Close closes the listener. It will be unavailable for use once closed. +func (s *InmemSocket) Close() error { + s.mu.Lock() + defer s.mu.Unlock() + select { + case <-s.chClose: + default: + close(s.chClose) + } + return nil +} + +// Dial is used to establish a connection with the in-mem server +func (s *InmemSocket) Dial(network, addr string) (net.Conn, error) { + srvConn, clientConn := net.Pipe() + select { + case s.chConn <- srvConn: + case <-s.chClose: + return nil, errClosed + } + + return clientConn, nil +} + +// Network returns the addr string, satisfies net.Addr +func (a dummyAddr) Network() string { + return string(a) +} + +// String returns the string form +func (a dummyAddr) String() string { + return string(a) +} diff --git a/vendor/github.com/docker/go-connections/sockets/proxy.go b/vendor/github.com/docker/go-connections/sockets/proxy.go new file mode 100644 index 000000000..98e9a1dc6 --- /dev/null +++ b/vendor/github.com/docker/go-connections/sockets/proxy.go @@ -0,0 +1,51 @@ +package sockets + +import ( + "net" + "net/url" + "os" + "strings" + + "golang.org/x/net/proxy" +) + +// GetProxyEnv allows access to the uppercase and the lowercase forms of +// proxy-related variables. See the Go specification for details on these +// variables. https://golang.org/pkg/net/http/ +func GetProxyEnv(key string) string { + proxyValue := os.Getenv(strings.ToUpper(key)) + if proxyValue == "" { + return os.Getenv(strings.ToLower(key)) + } + return proxyValue +} + +// DialerFromEnvironment takes in a "direct" *net.Dialer and returns a +// proxy.Dialer which will route the connections through the proxy using the +// given dialer. +func DialerFromEnvironment(direct *net.Dialer) (proxy.Dialer, error) { + allProxy := GetProxyEnv("all_proxy") + if len(allProxy) == 0 { + return direct, nil + } + + proxyURL, err := url.Parse(allProxy) + if err != nil { + return direct, err + } + + proxyFromURL, err := proxy.FromURL(proxyURL, direct) + if err != nil { + return direct, err + } + + noProxy := GetProxyEnv("no_proxy") + if len(noProxy) == 0 { + return proxyFromURL, nil + } + + perHost := proxy.NewPerHost(proxyFromURL, direct) + perHost.AddFromString(noProxy) + + return perHost, nil +} diff --git a/vendor/github.com/docker/go-connections/sockets/sockets.go b/vendor/github.com/docker/go-connections/sockets/sockets.go new file mode 100644 index 000000000..a1d7beb4d --- /dev/null +++ b/vendor/github.com/docker/go-connections/sockets/sockets.go @@ -0,0 +1,38 @@ +// Package sockets provides helper functions to create and configure Unix or TCP sockets. +package sockets + +import ( + "errors" + "net" + "net/http" + "time" +) + +// Why 32? See https://github.com/docker/docker/pull/8035. +const defaultTimeout = 32 * time.Second + +// ErrProtocolNotAvailable is returned when a given transport protocol is not provided by the operating system. +var ErrProtocolNotAvailable = errors.New("protocol not available") + +// ConfigureTransport configures the specified Transport according to the +// specified proto and addr. +// If the proto is unix (using a unix socket to communicate) or npipe the +// compression is disabled. +func ConfigureTransport(tr *http.Transport, proto, addr string) error { + switch proto { + case "unix": + return configureUnixTransport(tr, proto, addr) + case "npipe": + return configureNpipeTransport(tr, proto, addr) + default: + tr.Proxy = http.ProxyFromEnvironment + dialer, err := DialerFromEnvironment(&net.Dialer{ + Timeout: defaultTimeout, + }) + if err != nil { + return err + } + tr.Dial = dialer.Dial + } + return nil +} diff --git a/vendor/github.com/docker/go-connections/sockets/sockets_unix.go b/vendor/github.com/docker/go-connections/sockets/sockets_unix.go new file mode 100644 index 000000000..386cf0dbb --- /dev/null +++ b/vendor/github.com/docker/go-connections/sockets/sockets_unix.go @@ -0,0 +1,35 @@ +// +build !windows + +package sockets + +import ( + "fmt" + "net" + "net/http" + "syscall" + "time" +) + +const maxUnixSocketPathSize = len(syscall.RawSockaddrUnix{}.Path) + +func configureUnixTransport(tr *http.Transport, proto, addr string) error { + if len(addr) > maxUnixSocketPathSize { + return fmt.Errorf("Unix socket path %q is too long", addr) + } + // No need for compression in local communications. + tr.DisableCompression = true + tr.Dial = func(_, _ string) (net.Conn, error) { + return net.DialTimeout(proto, addr, defaultTimeout) + } + return nil +} + +func configureNpipeTransport(tr *http.Transport, proto, addr string) error { + return ErrProtocolNotAvailable +} + +// DialPipe connects to a Windows named pipe. +// This is not supported on other OSes. +func DialPipe(_ string, _ time.Duration) (net.Conn, error) { + return nil, syscall.EAFNOSUPPORT +} diff --git a/vendor/github.com/docker/go-connections/sockets/sockets_windows.go b/vendor/github.com/docker/go-connections/sockets/sockets_windows.go new file mode 100644 index 000000000..5c21644e1 --- /dev/null +++ b/vendor/github.com/docker/go-connections/sockets/sockets_windows.go @@ -0,0 +1,27 @@ +package sockets + +import ( + "net" + "net/http" + "time" + + "github.com/Microsoft/go-winio" +) + +func configureUnixTransport(tr *http.Transport, proto, addr string) error { + return ErrProtocolNotAvailable +} + +func configureNpipeTransport(tr *http.Transport, proto, addr string) error { + // No need for compression in local communications. + tr.DisableCompression = true + tr.Dial = func(_, _ string) (net.Conn, error) { + return DialPipe(addr, defaultTimeout) + } + return nil +} + +// DialPipe connects to a Windows named pipe. +func DialPipe(addr string, timeout time.Duration) (net.Conn, error) { + return winio.DialPipe(addr, &timeout) +} diff --git a/vendor/github.com/docker/go-connections/sockets/tcp_socket.go b/vendor/github.com/docker/go-connections/sockets/tcp_socket.go new file mode 100644 index 000000000..53cbb6c79 --- /dev/null +++ b/vendor/github.com/docker/go-connections/sockets/tcp_socket.go @@ -0,0 +1,22 @@ +// Package sockets provides helper functions to create and configure Unix or TCP sockets. +package sockets + +import ( + "crypto/tls" + "net" +) + +// NewTCPSocket creates a TCP socket listener with the specified address and +// the specified tls configuration. If TLSConfig is set, will encapsulate the +// TCP listener inside a TLS one. +func NewTCPSocket(addr string, tlsConfig *tls.Config) (net.Listener, error) { + l, err := net.Listen("tcp", addr) + if err != nil { + return nil, err + } + if tlsConfig != nil { + tlsConfig.NextProtos = []string{"http/1.1"} + l = tls.NewListener(l, tlsConfig) + } + return l, nil +} diff --git a/vendor/github.com/docker/go-connections/sockets/unix_socket.go b/vendor/github.com/docker/go-connections/sockets/unix_socket.go new file mode 100644 index 000000000..a8b5dbb6f --- /dev/null +++ b/vendor/github.com/docker/go-connections/sockets/unix_socket.go @@ -0,0 +1,32 @@ +// +build !windows + +package sockets + +import ( + "net" + "os" + "syscall" +) + +// NewUnixSocket creates a unix socket with the specified path and group. +func NewUnixSocket(path string, gid int) (net.Listener, error) { + if err := syscall.Unlink(path); err != nil && !os.IsNotExist(err) { + return nil, err + } + mask := syscall.Umask(0777) + defer syscall.Umask(mask) + + l, err := net.Listen("unix", path) + if err != nil { + return nil, err + } + if err := os.Chown(path, 0, gid); err != nil { + l.Close() + return nil, err + } + if err := os.Chmod(path, 0660); err != nil { + l.Close() + return nil, err + } + return l, nil +} diff --git a/vendor/github.com/docker/go-connections/tlsconfig/certpool_go17.go b/vendor/github.com/docker/go-connections/tlsconfig/certpool_go17.go new file mode 100644 index 000000000..1ca0965e0 --- /dev/null +++ b/vendor/github.com/docker/go-connections/tlsconfig/certpool_go17.go @@ -0,0 +1,18 @@ +// +build go1.7 + +package tlsconfig + +import ( + "crypto/x509" + "runtime" +) + +// SystemCertPool returns a copy of the system cert pool, +// returns an error if failed to load or empty pool on windows. +func SystemCertPool() (*x509.CertPool, error) { + certpool, err := x509.SystemCertPool() + if err != nil && runtime.GOOS == "windows" { + return x509.NewCertPool(), nil + } + return certpool, err +} diff --git a/vendor/github.com/docker/go-connections/tlsconfig/certpool_other.go b/vendor/github.com/docker/go-connections/tlsconfig/certpool_other.go new file mode 100644 index 000000000..1ff81c333 --- /dev/null +++ b/vendor/github.com/docker/go-connections/tlsconfig/certpool_other.go @@ -0,0 +1,13 @@ +// +build !go1.7 + +package tlsconfig + +import ( + "crypto/x509" +) + +// SystemCertPool returns an new empty cert pool, +// accessing system cert pool is supported in go 1.7 +func SystemCertPool() (*x509.CertPool, error) { + return x509.NewCertPool(), nil +} diff --git a/vendor/github.com/docker/go-connections/tlsconfig/config.go b/vendor/github.com/docker/go-connections/tlsconfig/config.go new file mode 100644 index 000000000..f11f166a4 --- /dev/null +++ b/vendor/github.com/docker/go-connections/tlsconfig/config.go @@ -0,0 +1,256 @@ +// Package tlsconfig provides primitives to retrieve secure-enough TLS configurations for both clients and servers. +// +// As a reminder from https://golang.org/pkg/crypto/tls/#Config: +// A Config structure is used to configure a TLS client or server. After one has been passed to a TLS function it must not be modified. +// A Config may be reused; the tls package will also not modify it. +package tlsconfig + +import ( + "crypto/tls" + "crypto/x509" + "encoding/pem" + "fmt" + "io/ioutil" + "os" + + "github.com/pkg/errors" +) + +// Options represents the information needed to create client and server TLS configurations. +type Options struct { + CAFile string + + // If either CertFile or KeyFile is empty, Client() will not load them + // preventing the client from authenticating to the server. + // However, Server() requires them and will error out if they are empty. + CertFile string + KeyFile string + + // client-only option + InsecureSkipVerify bool + // server-only option + ClientAuth tls.ClientAuthType + // If ExclusiveRootPools is set, then if a CA file is provided, the root pool used for TLS + // creds will include exclusively the roots in that CA file. If no CA file is provided, + // the system pool will be used. + ExclusiveRootPools bool + MinVersion uint16 + // If Passphrase is set, it will be used to decrypt a TLS private key + // if the key is encrypted + Passphrase string +} + +// Extra (server-side) accepted CBC cipher suites - will phase out in the future +var acceptedCBCCiphers = []uint16{ + tls.TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA, + tls.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA, + tls.TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA, + tls.TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA, + tls.TLS_RSA_WITH_AES_256_CBC_SHA, + tls.TLS_RSA_WITH_AES_128_CBC_SHA, +} + +// DefaultServerAcceptedCiphers should be uses by code which already has a crypto/tls +// options struct but wants to use a commonly accepted set of TLS cipher suites, with +// known weak algorithms removed. +var DefaultServerAcceptedCiphers = append(clientCipherSuites, acceptedCBCCiphers...) + +// allTLSVersions lists all the TLS versions and is used by the code that validates +// a uint16 value as a TLS version. +var allTLSVersions = map[uint16]struct{}{ + tls.VersionSSL30: {}, + tls.VersionTLS10: {}, + tls.VersionTLS11: {}, + tls.VersionTLS12: {}, +} + +// ServerDefault returns a secure-enough TLS configuration for the server TLS configuration. +func ServerDefault(ops ...func(*tls.Config)) *tls.Config { + tlsconfig := &tls.Config{ + // Avoid fallback by default to SSL protocols < TLS1.0 + MinVersion: tls.VersionTLS10, + PreferServerCipherSuites: true, + CipherSuites: DefaultServerAcceptedCiphers, + } + + for _, op := range ops { + op(tlsconfig) + } + + return tlsconfig +} + +// ClientDefault returns a secure-enough TLS configuration for the client TLS configuration. +func ClientDefault(ops ...func(*tls.Config)) *tls.Config { + tlsconfig := &tls.Config{ + // Prefer TLS1.2 as the client minimum + MinVersion: tls.VersionTLS12, + CipherSuites: clientCipherSuites, + } + + for _, op := range ops { + op(tlsconfig) + } + + return tlsconfig +} + +// certPool returns an X.509 certificate pool from `caFile`, the certificate file. +func certPool(caFile string, exclusivePool bool) (*x509.CertPool, error) { + // If we should verify the server, we need to load a trusted ca + var ( + certPool *x509.CertPool + err error + ) + if exclusivePool { + certPool = x509.NewCertPool() + } else { + certPool, err = SystemCertPool() + if err != nil { + return nil, fmt.Errorf("failed to read system certificates: %v", err) + } + } + pem, err := ioutil.ReadFile(caFile) + if err != nil { + return nil, fmt.Errorf("could not read CA certificate %q: %v", caFile, err) + } + if !certPool.AppendCertsFromPEM(pem) { + return nil, fmt.Errorf("failed to append certificates from PEM file: %q", caFile) + } + return certPool, nil +} + +// isValidMinVersion checks that the input value is a valid tls minimum version +func isValidMinVersion(version uint16) bool { + _, ok := allTLSVersions[version] + return ok +} + +// adjustMinVersion sets the MinVersion on `config`, the input configuration. +// It assumes the current MinVersion on the `config` is the lowest allowed. +func adjustMinVersion(options Options, config *tls.Config) error { + if options.MinVersion > 0 { + if !isValidMinVersion(options.MinVersion) { + return fmt.Errorf("Invalid minimum TLS version: %x", options.MinVersion) + } + if options.MinVersion < config.MinVersion { + return fmt.Errorf("Requested minimum TLS version is too low. Should be at-least: %x", config.MinVersion) + } + config.MinVersion = options.MinVersion + } + + return nil +} + +// IsErrEncryptedKey returns true if the 'err' is an error of incorrect +// password when tryin to decrypt a TLS private key +func IsErrEncryptedKey(err error) bool { + return errors.Cause(err) == x509.IncorrectPasswordError +} + +// getPrivateKey returns the private key in 'keyBytes', in PEM-encoded format. +// If the private key is encrypted, 'passphrase' is used to decrypted the +// private key. +func getPrivateKey(keyBytes []byte, passphrase string) ([]byte, error) { + // this section makes some small changes to code from notary/tuf/utils/x509.go + pemBlock, _ := pem.Decode(keyBytes) + if pemBlock == nil { + return nil, fmt.Errorf("no valid private key found") + } + + var err error + if x509.IsEncryptedPEMBlock(pemBlock) { + keyBytes, err = x509.DecryptPEMBlock(pemBlock, []byte(passphrase)) + if err != nil { + return nil, errors.Wrap(err, "private key is encrypted, but could not decrypt it") + } + keyBytes = pem.EncodeToMemory(&pem.Block{Type: pemBlock.Type, Bytes: keyBytes}) + } + + return keyBytes, nil +} + +// getCert returns a Certificate from the CertFile and KeyFile in 'options', +// if the key is encrypted, the Passphrase in 'options' will be used to +// decrypt it. +func getCert(options Options) ([]tls.Certificate, error) { + if options.CertFile == "" && options.KeyFile == "" { + return nil, nil + } + + errMessage := "Could not load X509 key pair" + + cert, err := ioutil.ReadFile(options.CertFile) + if err != nil { + return nil, errors.Wrap(err, errMessage) + } + + prKeyBytes, err := ioutil.ReadFile(options.KeyFile) + if err != nil { + return nil, errors.Wrap(err, errMessage) + } + + prKeyBytes, err = getPrivateKey(prKeyBytes, options.Passphrase) + if err != nil { + return nil, errors.Wrap(err, errMessage) + } + + tlsCert, err := tls.X509KeyPair(cert, prKeyBytes) + if err != nil { + return nil, errors.Wrap(err, errMessage) + } + + return []tls.Certificate{tlsCert}, nil +} + +// Client returns a TLS configuration meant to be used by a client. +func Client(options Options) (*tls.Config, error) { + tlsConfig := ClientDefault() + tlsConfig.InsecureSkipVerify = options.InsecureSkipVerify + if !options.InsecureSkipVerify && options.CAFile != "" { + CAs, err := certPool(options.CAFile, options.ExclusiveRootPools) + if err != nil { + return nil, err + } + tlsConfig.RootCAs = CAs + } + + tlsCerts, err := getCert(options) + if err != nil { + return nil, err + } + tlsConfig.Certificates = tlsCerts + + if err := adjustMinVersion(options, tlsConfig); err != nil { + return nil, err + } + + return tlsConfig, nil +} + +// Server returns a TLS configuration meant to be used by a server. +func Server(options Options) (*tls.Config, error) { + tlsConfig := ServerDefault() + tlsConfig.ClientAuth = options.ClientAuth + tlsCert, err := tls.LoadX509KeyPair(options.CertFile, options.KeyFile) + if err != nil { + if os.IsNotExist(err) { + return nil, fmt.Errorf("Could not load X509 key pair (cert: %q, key: %q): %v", options.CertFile, options.KeyFile, err) + } + return nil, fmt.Errorf("Error reading X509 key pair (cert: %q, key: %q): %v. Make sure the key is not encrypted.", options.CertFile, options.KeyFile, err) + } + tlsConfig.Certificates = []tls.Certificate{tlsCert} + if options.ClientAuth >= tls.VerifyClientCertIfGiven && options.CAFile != "" { + CAs, err := certPool(options.CAFile, options.ExclusiveRootPools) + if err != nil { + return nil, err + } + tlsConfig.ClientCAs = CAs + } + + if err := adjustMinVersion(options, tlsConfig); err != nil { + return nil, err + } + + return tlsConfig, nil +} diff --git a/vendor/github.com/docker/go-connections/tlsconfig/config_client_ciphers.go b/vendor/github.com/docker/go-connections/tlsconfig/config_client_ciphers.go new file mode 100644 index 000000000..6b4c6a7c0 --- /dev/null +++ b/vendor/github.com/docker/go-connections/tlsconfig/config_client_ciphers.go @@ -0,0 +1,17 @@ +// +build go1.5 + +// Package tlsconfig provides primitives to retrieve secure-enough TLS configurations for both clients and servers. +// +package tlsconfig + +import ( + "crypto/tls" +) + +// Client TLS cipher suites (dropping CBC ciphers for client preferred suite set) +var clientCipherSuites = []uint16{ + tls.TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384, + tls.TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384, + tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, + tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, +} diff --git a/vendor/github.com/docker/go-connections/tlsconfig/config_legacy_client_ciphers.go b/vendor/github.com/docker/go-connections/tlsconfig/config_legacy_client_ciphers.go new file mode 100644 index 000000000..ee22df47c --- /dev/null +++ b/vendor/github.com/docker/go-connections/tlsconfig/config_legacy_client_ciphers.go @@ -0,0 +1,15 @@ +// +build !go1.5 + +// Package tlsconfig provides primitives to retrieve secure-enough TLS configurations for both clients and servers. +// +package tlsconfig + +import ( + "crypto/tls" +) + +// Client TLS cipher suites (dropping CBC ciphers for client preferred suite set) +var clientCipherSuites = []uint16{ + tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, + tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, +} diff --git a/vendor/github.com/docker/go-metrics/CONTRIBUTING.md b/vendor/github.com/docker/go-metrics/CONTRIBUTING.md new file mode 100644 index 000000000..b8a512c36 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/CONTRIBUTING.md @@ -0,0 +1,55 @@ +# Contributing + +## Sign your work + +The sign-off is a simple line at the end of the explanation for the patch. Your +signature certifies that you wrote the patch or otherwise have the right to pass +it on as an open-source patch. The rules are pretty simple: if you can certify +the below (from [developercertificate.org](http://developercertificate.org/)): + +``` +Developer Certificate of Origin +Version 1.1 + +Copyright (C) 2004, 2006 The Linux Foundation and its contributors. +660 York Street, Suite 102, +San Francisco, CA 94110 USA + +Everyone is permitted to copy and distribute verbatim copies of this +license document, but changing it is not allowed. + +Developer's Certificate of Origin 1.1 + +By making a contribution to this project, I certify that: + +(a) The contribution was created in whole or in part by me and I + have the right to submit it under the open source license + indicated in the file; or + +(b) The contribution is based upon previous work that, to the best + of my knowledge, is covered under an appropriate open source + license and I have the right under that license to submit that + work with modifications, whether created in whole or in part + by me, under the same open source license (unless I am + permitted to submit under a different license), as indicated + in the file; or + +(c) The contribution was provided directly to me by some other + person who certified (a), (b) or (c) and I have not modified + it. + +(d) I understand and agree that this project and the contribution + are public and that a record of the contribution (including all + personal information I submit with it, including my sign-off) is + maintained indefinitely and may be redistributed consistent with + this project or the open source license(s) involved. +``` + +Then you just add a line to every git commit message: + + Signed-off-by: Joe Smith + +Use your real name (sorry, no pseudonyms or anonymous contributions.) + +If you set your `user.name` and `user.email` git configs, you can sign your +commit automatically with `git commit -s`. diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/LICENSE b/vendor/github.com/docker/go-metrics/LICENSE.code similarity index 99% rename from vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/LICENSE rename to vendor/github.com/docker/go-metrics/LICENSE.code index c7a3f0cfd..8f3fee627 100644 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/LICENSE +++ b/vendor/github.com/docker/go-metrics/LICENSE.code @@ -176,7 +176,7 @@ END OF TERMS AND CONDITIONS - Copyright 2013-2015 Docker, Inc. + Copyright 2013-2016 Docker, Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. diff --git a/vendor/github.com/docker/go-metrics/LICENSE.docs b/vendor/github.com/docker/go-metrics/LICENSE.docs new file mode 100644 index 000000000..e26cd4fc8 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/LICENSE.docs @@ -0,0 +1,425 @@ +Attribution-ShareAlike 4.0 International + +======================================================================= + +Creative Commons Corporation ("Creative Commons") is not a law firm and +does not provide legal services or legal advice. Distribution of +Creative Commons public licenses does not create a lawyer-client or +other relationship. Creative Commons makes its licenses and related +information available on an "as-is" basis. Creative Commons gives no +warranties regarding its licenses, any material licensed under their +terms and conditions, or any related information. Creative Commons +disclaims all liability for damages resulting from their use to the +fullest extent possible. + +Using Creative Commons Public Licenses + +Creative Commons public licenses provide a standard set of terms and +conditions that creators and other rights holders may use to share +original works of authorship and other material subject to copyright +and certain other rights specified in the public license below. The +following considerations are for informational purposes only, are not +exhaustive, and do not form part of our licenses. + + Considerations for licensors: Our public licenses are + intended for use by those authorized to give the public + permission to use material in ways otherwise restricted by + copyright and certain other rights. Our licenses are + irrevocable. Licensors should read and understand the terms + and conditions of the license they choose before applying it. + Licensors should also secure all rights necessary before + applying our licenses so that the public can reuse the + material as expected. Licensors should clearly mark any + material not subject to the license. This includes other CC- + licensed material, or material used under an exception or + limitation to copyright. More considerations for licensors: + wiki.creativecommons.org/Considerations_for_licensors + + Considerations for the public: By using one of our public + licenses, a licensor grants the public permission to use the + licensed material under specified terms and conditions. If + the licensor's permission is not necessary for any reason--for + example, because of any applicable exception or limitation to + copyright--then that use is not regulated by the license. Our + licenses grant only permissions under copyright and certain + other rights that a licensor has authority to grant. Use of + the licensed material may still be restricted for other + reasons, including because others have copyright or other + rights in the material. A licensor may make special requests, + such as asking that all changes be marked or described. + Although not required by our licenses, you are encouraged to + respect those requests where reasonable. More_considerations + for the public: + wiki.creativecommons.org/Considerations_for_licensees + +======================================================================= + +Creative Commons Attribution-ShareAlike 4.0 International Public +License + +By exercising the Licensed Rights (defined below), You accept and agree +to be bound by the terms and conditions of this Creative Commons +Attribution-ShareAlike 4.0 International Public License ("Public +License"). To the extent this Public License may be interpreted as a +contract, You are granted the Licensed Rights in consideration of Your +acceptance of these terms and conditions, and the Licensor grants You +such rights in consideration of benefits the Licensor receives from +making the Licensed Material available under these terms and +conditions. + + +Section 1 -- Definitions. + + a. Adapted Material means material subject to Copyright and Similar + Rights that is derived from or based upon the Licensed Material + and in which the Licensed Material is translated, altered, + arranged, transformed, or otherwise modified in a manner requiring + permission under the Copyright and Similar Rights held by the + Licensor. For purposes of this Public License, where the Licensed + Material is a musical work, performance, or sound recording, + Adapted Material is always produced where the Licensed Material is + synched in timed relation with a moving image. + + b. Adapter's License means the license You apply to Your Copyright + and Similar Rights in Your contributions to Adapted Material in + accordance with the terms and conditions of this Public License. + + c. BY-SA Compatible License means a license listed at + creativecommons.org/compatiblelicenses, approved by Creative + Commons as essentially the equivalent of this Public License. + + d. Copyright and Similar Rights means copyright and/or similar rights + closely related to copyright including, without limitation, + performance, broadcast, sound recording, and Sui Generis Database + Rights, without regard to how the rights are labeled or + categorized. For purposes of this Public License, the rights + specified in Section 2(b)(1)-(2) are not Copyright and Similar + Rights. + + e. Effective Technological Measures means those measures that, in the + absence of proper authority, may not be circumvented under laws + fulfilling obligations under Article 11 of the WIPO Copyright + Treaty adopted on December 20, 1996, and/or similar international + agreements. + + f. Exceptions and Limitations means fair use, fair dealing, and/or + any other exception or limitation to Copyright and Similar Rights + that applies to Your use of the Licensed Material. + + g. License Elements means the license attributes listed in the name + of a Creative Commons Public License. The License Elements of this + Public License are Attribution and ShareAlike. + + h. Licensed Material means the artistic or literary work, database, + or other material to which the Licensor applied this Public + License. + + i. Licensed Rights means the rights granted to You subject to the + terms and conditions of this Public License, which are limited to + all Copyright and Similar Rights that apply to Your use of the + Licensed Material and that the Licensor has authority to license. + + j. Licensor means the individual(s) or entity(ies) granting rights + under this Public License. + + k. Share means to provide material to the public by any means or + process that requires permission under the Licensed Rights, such + as reproduction, public display, public performance, distribution, + dissemination, communication, or importation, and to make material + available to the public including in ways that members of the + public may access the material from a place and at a time + individually chosen by them. + + l. Sui Generis Database Rights means rights other than copyright + resulting from Directive 96/9/EC of the European Parliament and of + the Council of 11 March 1996 on the legal protection of databases, + as amended and/or succeeded, as well as other essentially + equivalent rights anywhere in the world. + + m. You means the individual or entity exercising the Licensed Rights + under this Public License. Your has a corresponding meaning. + + +Section 2 -- Scope. + + a. License grant. + + 1. Subject to the terms and conditions of this Public License, + the Licensor hereby grants You a worldwide, royalty-free, + non-sublicensable, non-exclusive, irrevocable license to + exercise the Licensed Rights in the Licensed Material to: + + a. reproduce and Share the Licensed Material, in whole or + in part; and + + b. produce, reproduce, and Share Adapted Material. + + 2. Exceptions and Limitations. For the avoidance of doubt, where + Exceptions and Limitations apply to Your use, this Public + License does not apply, and You do not need to comply with + its terms and conditions. + + 3. Term. The term of this Public License is specified in Section + 6(a). + + 4. Media and formats; technical modifications allowed. The + Licensor authorizes You to exercise the Licensed Rights in + all media and formats whether now known or hereafter created, + and to make technical modifications necessary to do so. The + Licensor waives and/or agrees not to assert any right or + authority to forbid You from making technical modifications + necessary to exercise the Licensed Rights, including + technical modifications necessary to circumvent Effective + Technological Measures. For purposes of this Public License, + simply making modifications authorized by this Section 2(a) + (4) never produces Adapted Material. + + 5. Downstream recipients. + + a. Offer from the Licensor -- Licensed Material. Every + recipient of the Licensed Material automatically + receives an offer from the Licensor to exercise the + Licensed Rights under the terms and conditions of this + Public License. + + b. Additional offer from the Licensor -- Adapted Material. + Every recipient of Adapted Material from You + automatically receives an offer from the Licensor to + exercise the Licensed Rights in the Adapted Material + under the conditions of the Adapter's License You apply. + + c. No downstream restrictions. You may not offer or impose + any additional or different terms or conditions on, or + apply any Effective Technological Measures to, the + Licensed Material if doing so restricts exercise of the + Licensed Rights by any recipient of the Licensed + Material. + + 6. No endorsement. Nothing in this Public License constitutes or + may be construed as permission to assert or imply that You + are, or that Your use of the Licensed Material is, connected + with, or sponsored, endorsed, or granted official status by, + the Licensor or others designated to receive attribution as + provided in Section 3(a)(1)(A)(i). + + b. Other rights. + + 1. Moral rights, such as the right of integrity, are not + licensed under this Public License, nor are publicity, + privacy, and/or other similar personality rights; however, to + the extent possible, the Licensor waives and/or agrees not to + assert any such rights held by the Licensor to the limited + extent necessary to allow You to exercise the Licensed + Rights, but not otherwise. + + 2. Patent and trademark rights are not licensed under this + Public License. + + 3. To the extent possible, the Licensor waives any right to + collect royalties from You for the exercise of the Licensed + Rights, whether directly or through a collecting society + under any voluntary or waivable statutory or compulsory + licensing scheme. In all other cases the Licensor expressly + reserves any right to collect such royalties. + + +Section 3 -- License Conditions. + +Your exercise of the Licensed Rights is expressly made subject to the +following conditions. + + a. Attribution. + + 1. If You Share the Licensed Material (including in modified + form), You must: + + a. retain the following if it is supplied by the Licensor + with the Licensed Material: + + i. identification of the creator(s) of the Licensed + Material and any others designated to receive + attribution, in any reasonable manner requested by + the Licensor (including by pseudonym if + designated); + + ii. a copyright notice; + + iii. a notice that refers to this Public License; + + iv. a notice that refers to the disclaimer of + warranties; + + v. a URI or hyperlink to the Licensed Material to the + extent reasonably practicable; + + b. indicate if You modified the Licensed Material and + retain an indication of any previous modifications; and + + c. indicate the Licensed Material is licensed under this + Public License, and include the text of, or the URI or + hyperlink to, this Public License. + + 2. You may satisfy the conditions in Section 3(a)(1) in any + reasonable manner based on the medium, means, and context in + which You Share the Licensed Material. For example, it may be + reasonable to satisfy the conditions by providing a URI or + hyperlink to a resource that includes the required + information. + + 3. If requested by the Licensor, You must remove any of the + information required by Section 3(a)(1)(A) to the extent + reasonably practicable. + + b. ShareAlike. + + In addition to the conditions in Section 3(a), if You Share + Adapted Material You produce, the following conditions also apply. + + 1. The Adapter's License You apply must be a Creative Commons + license with the same License Elements, this version or + later, or a BY-SA Compatible License. + + 2. You must include the text of, or the URI or hyperlink to, the + Adapter's License You apply. You may satisfy this condition + in any reasonable manner based on the medium, means, and + context in which You Share Adapted Material. + + 3. You may not offer or impose any additional or different terms + or conditions on, or apply any Effective Technological + Measures to, Adapted Material that restrict exercise of the + rights granted under the Adapter's License You apply. + + +Section 4 -- Sui Generis Database Rights. + +Where the Licensed Rights include Sui Generis Database Rights that +apply to Your use of the Licensed Material: + + a. for the avoidance of doubt, Section 2(a)(1) grants You the right + to extract, reuse, reproduce, and Share all or a substantial + portion of the contents of the database; + + b. if You include all or a substantial portion of the database + contents in a database in which You have Sui Generis Database + Rights, then the database in which You have Sui Generis Database + Rights (but not its individual contents) is Adapted Material, + + including for purposes of Section 3(b); and + c. You must comply with the conditions in Section 3(a) if You Share + all or a substantial portion of the contents of the database. + +For the avoidance of doubt, this Section 4 supplements and does not +replace Your obligations under this Public License where the Licensed +Rights include other Copyright and Similar Rights. + + +Section 5 -- Disclaimer of Warranties and Limitation of Liability. + + a. UNLESS OTHERWISE SEPARATELY UNDERTAKEN BY THE LICENSOR, TO THE + EXTENT POSSIBLE, THE LICENSOR OFFERS THE LICENSED MATERIAL AS-IS + AND AS-AVAILABLE, AND MAKES NO REPRESENTATIONS OR WARRANTIES OF + ANY KIND CONCERNING THE LICENSED MATERIAL, WHETHER EXPRESS, + IMPLIED, STATUTORY, OR OTHER. THIS INCLUDES, WITHOUT LIMITATION, + WARRANTIES OF TITLE, MERCHANTABILITY, FITNESS FOR A PARTICULAR + PURPOSE, NON-INFRINGEMENT, ABSENCE OF LATENT OR OTHER DEFECTS, + ACCURACY, OR THE PRESENCE OR ABSENCE OF ERRORS, WHETHER OR NOT + KNOWN OR DISCOVERABLE. WHERE DISCLAIMERS OF WARRANTIES ARE NOT + ALLOWED IN FULL OR IN PART, THIS DISCLAIMER MAY NOT APPLY TO YOU. + + b. TO THE EXTENT POSSIBLE, IN NO EVENT WILL THE LICENSOR BE LIABLE + TO YOU ON ANY LEGAL THEORY (INCLUDING, WITHOUT LIMITATION, + NEGLIGENCE) OR OTHERWISE FOR ANY DIRECT, SPECIAL, INDIRECT, + INCIDENTAL, CONSEQUENTIAL, PUNITIVE, EXEMPLARY, OR OTHER LOSSES, + COSTS, EXPENSES, OR DAMAGES ARISING OUT OF THIS PUBLIC LICENSE OR + USE OF THE LICENSED MATERIAL, EVEN IF THE LICENSOR HAS BEEN + ADVISED OF THE POSSIBILITY OF SUCH LOSSES, COSTS, EXPENSES, OR + DAMAGES. WHERE A LIMITATION OF LIABILITY IS NOT ALLOWED IN FULL OR + IN PART, THIS LIMITATION MAY NOT APPLY TO YOU. + + c. The disclaimer of warranties and limitation of liability provided + above shall be interpreted in a manner that, to the extent + possible, most closely approximates an absolute disclaimer and + waiver of all liability. + + +Section 6 -- Term and Termination. + + a. This Public License applies for the term of the Copyright and + Similar Rights licensed here. However, if You fail to comply with + this Public License, then Your rights under this Public License + terminate automatically. + + b. Where Your right to use the Licensed Material has terminated under + Section 6(a), it reinstates: + + 1. automatically as of the date the violation is cured, provided + it is cured within 30 days of Your discovery of the + violation; or + + 2. upon express reinstatement by the Licensor. + + For the avoidance of doubt, this Section 6(b) does not affect any + right the Licensor may have to seek remedies for Your violations + of this Public License. + + c. For the avoidance of doubt, the Licensor may also offer the + Licensed Material under separate terms or conditions or stop + distributing the Licensed Material at any time; however, doing so + will not terminate this Public License. + + d. Sections 1, 5, 6, 7, and 8 survive termination of this Public + License. + + +Section 7 -- Other Terms and Conditions. + + a. The Licensor shall not be bound by any additional or different + terms or conditions communicated by You unless expressly agreed. + + b. Any arrangements, understandings, or agreements regarding the + Licensed Material not stated herein are separate from and + independent of the terms and conditions of this Public License. + + +Section 8 -- Interpretation. + + a. For the avoidance of doubt, this Public License does not, and + shall not be interpreted to, reduce, limit, restrict, or impose + conditions on any use of the Licensed Material that could lawfully + be made without permission under this Public License. + + b. To the extent possible, if any provision of this Public License is + deemed unenforceable, it shall be automatically reformed to the + minimum extent necessary to make it enforceable. If the provision + cannot be reformed, it shall be severed from this Public License + without affecting the enforceability of the remaining terms and + conditions. + + c. No term or condition of this Public License will be waived and no + failure to comply consented to unless expressly agreed to by the + Licensor. + + d. Nothing in this Public License constitutes or may be interpreted + as a limitation upon, or waiver of, any privileges and immunities + that apply to the Licensor or You, including from the legal + processes of any jurisdiction or authority. + + +======================================================================= + +Creative Commons is not a party to its public licenses. +Notwithstanding, Creative Commons may elect to apply one of its public +licenses to material it publishes and in those instances will be +considered the "Licensor." Except for the limited purpose of indicating +that material is shared under a Creative Commons public license or as +otherwise permitted by the Creative Commons policies published at +creativecommons.org/policies, Creative Commons does not authorize the +use of the trademark "Creative Commons" or any other trademark or logo +of Creative Commons without its prior written consent including, +without limitation, in connection with any unauthorized modifications +to any of its public licenses or any other arrangements, +understandings, or agreements concerning use of licensed material. For +the avoidance of doubt, this paragraph does not form part of the public +licenses. + +Creative Commons may be contacted at creativecommons.org. diff --git a/vendor/github.com/docker/go-metrics/NOTICE b/vendor/github.com/docker/go-metrics/NOTICE new file mode 100644 index 000000000..8915f0277 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/NOTICE @@ -0,0 +1,16 @@ +Docker +Copyright 2012-2015 Docker, Inc. + +This product includes software developed at Docker, Inc. (https://www.docker.com). + +The following is courtesy of our legal counsel: + + +Use and transfer of Docker may be subject to certain restrictions by the +United States and other governments. +It is your responsibility to ensure that your use and/or transfer does not +violate applicable laws. + +For more information, please see https://www.bis.doc.gov + +See also https://www.apache.org/dev/crypto.html and/or seek legal counsel. diff --git a/vendor/github.com/docker/go-metrics/README.md b/vendor/github.com/docker/go-metrics/README.md new file mode 100644 index 000000000..a9e947cb5 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/README.md @@ -0,0 +1,91 @@ +# go-metrics [![GoDoc](https://godoc.org/github.com/docker/go-metrics?status.svg)](https://godoc.org/github.com/docker/go-metrics) ![Badge Badge](http://doyouevenbadge.com/github.com/docker/go-metrics) + +This package is small wrapper around the prometheus go client to help enforce convention and best practices for metrics collection in Docker projects. + +## Best Practices + +This packages is meant to be used for collecting metrics in Docker projects. +It is not meant to be used as a replacement for the prometheus client but to help enforce consistent naming across metrics collected. +If you have not already read the prometheus best practices around naming and labels you can read the page [here](https://prometheus.io/docs/practices/naming/). + +The following are a few Docker specific rules that will help you name and work with metrics in your project. + +1. Namespace and Subsystem + +This package provides you with a namespace type that allows you to specify the same namespace and subsystem for your metrics. + +```go +ns := metrics.NewNamespace("engine", "daemon", metrics.Labels{ + "version": dockerversion.Version, + "commit": dockerversion.GitCommit, +}) +``` + +In the example above we are creating metrics for the Docker engine's daemon package. +`engine` would be the namespace in this example where `daemon` is the subsystem or package where we are collecting the metrics. + +A namespace also allows you to attach constant labels to the metrics such as the git commit and version that it is collecting. + +2. Declaring your Metrics + +Try to keep all your metric declarations in one file. +This makes it easy for others to see what constant labels are defined on the namespace and what labels are defined on the metrics when they are created. + +3. Use labels instead of multiple metrics + +Labels allow you to define one metric such as the time it takes to perform a certain action on an object. +If we wanted to collect timings on various container actions such as create, start, and delete then we can define one metric called `container_actions` and use labels to specify the type of action. + + +```go +containerActions = ns.NewLabeledTimer("container_actions", "The number of milliseconds it takes to process each container action", "action") +``` + +The last parameter is the label name or key. +When adding a data point to the metric you will use the `WithValues` function to specify the `action` that you are collecting for. + +```go +containerActions.WithValues("create").UpdateSince(start) +``` + +4. Always use a unit + +The metric name should describe what you are measuring but you also need to provide the unit that it is being measured with. +For a timer, the standard unit is seconds and a counter's standard unit is a total. +For gauges you must provide the unit. +This package provides a standard set of units for use within the Docker projects. + +```go +Nanoseconds Unit = "nanoseconds" +Seconds Unit = "seconds" +Bytes Unit = "bytes" +Total Unit = "total" +``` + +If you need to use a unit but it is not defined in the package please open a PR to add it but first try to see if one of the already created units will work for your metric, i.e. seconds or nanoseconds vs adding milliseconds. + +## Docs + +Package documentation can be found [here](https://godoc.org/github.com/docker/go-metrics). + +## HTTP Metrics + +To instrument a http handler, you can wrap the code like this: + +```go +namespace := metrics.NewNamespace("docker_distribution", "http", metrics.Labels{"handler": "your_http_handler_name"}) +httpMetrics := namespace.NewDefaultHttpMetrics() +metrics.Register(namespace) +instrumentedHandler = metrics.InstrumentHandler(httpMetrics, unInstrumentedHandler) +``` +Note: The `handler` label must be provided when a new namespace is created. + +## Additional Metrics + +Additional metrics are also defined here that are not available in the prometheus client. +If you need a custom metrics and it is generic enough to be used by multiple projects, define it here. + + +## Copyright and license + +Copyright © 2016 Docker, Inc. All rights reserved, except as follows. Code is released under the Apache 2.0 license. The README.md file, and files in the "docs" folder are licensed under the Creative Commons Attribution 4.0 International License under the terms and conditions set forth in the file "LICENSE.docs". You may obtain a duplicate copy of the same license, titled CC-BY-SA-4.0, at http://creativecommons.org/licenses/by/4.0/. diff --git a/vendor/github.com/docker/go-metrics/counter.go b/vendor/github.com/docker/go-metrics/counter.go new file mode 100644 index 000000000..fe36316a4 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/counter.go @@ -0,0 +1,52 @@ +package metrics + +import "github.com/prometheus/client_golang/prometheus" + +// Counter is a metrics that can only increment its current count +type Counter interface { + // Inc adds Sum(vs) to the counter. Sum(vs) must be positive. + // + // If len(vs) == 0, increments the counter by 1. + Inc(vs ...float64) +} + +// LabeledCounter is counter that must have labels populated before use. +type LabeledCounter interface { + WithValues(vs ...string) Counter +} + +type labeledCounter struct { + pc *prometheus.CounterVec +} + +func (lc *labeledCounter) WithValues(vs ...string) Counter { + return &counter{pc: lc.pc.WithLabelValues(vs...)} +} + +func (lc *labeledCounter) Describe(ch chan<- *prometheus.Desc) { + lc.pc.Describe(ch) +} + +func (lc *labeledCounter) Collect(ch chan<- prometheus.Metric) { + lc.pc.Collect(ch) +} + +type counter struct { + pc prometheus.Counter +} + +func (c *counter) Inc(vs ...float64) { + if len(vs) == 0 { + c.pc.Inc() + } + + c.pc.Add(sumFloat64(vs...)) +} + +func (c *counter) Describe(ch chan<- *prometheus.Desc) { + c.pc.Describe(ch) +} + +func (c *counter) Collect(ch chan<- prometheus.Metric) { + c.pc.Collect(ch) +} diff --git a/vendor/github.com/docker/go-metrics/docs.go b/vendor/github.com/docker/go-metrics/docs.go new file mode 100644 index 000000000..8fbdfc697 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/docs.go @@ -0,0 +1,3 @@ +// This package is small wrapper around the prometheus go client to help enforce convention and best practices for metrics collection in Docker projects. + +package metrics diff --git a/vendor/github.com/docker/go-metrics/gauge.go b/vendor/github.com/docker/go-metrics/gauge.go new file mode 100644 index 000000000..74296e877 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/gauge.go @@ -0,0 +1,72 @@ +package metrics + +import "github.com/prometheus/client_golang/prometheus" + +// Gauge is a metric that allows incrementing and decrementing a value +type Gauge interface { + Inc(...float64) + Dec(...float64) + + // Add adds the provided value to the gauge's current value + Add(float64) + + // Set replaces the gauge's current value with the provided value + Set(float64) +} + +// LabeledGauge describes a gauge the must have values populated before use. +type LabeledGauge interface { + WithValues(labels ...string) Gauge +} + +type labeledGauge struct { + pg *prometheus.GaugeVec +} + +func (lg *labeledGauge) WithValues(labels ...string) Gauge { + return &gauge{pg: lg.pg.WithLabelValues(labels...)} +} + +func (lg *labeledGauge) Describe(c chan<- *prometheus.Desc) { + lg.pg.Describe(c) +} + +func (lg *labeledGauge) Collect(c chan<- prometheus.Metric) { + lg.pg.Collect(c) +} + +type gauge struct { + pg prometheus.Gauge +} + +func (g *gauge) Inc(vs ...float64) { + if len(vs) == 0 { + g.pg.Inc() + } + + g.Add(sumFloat64(vs...)) +} + +func (g *gauge) Dec(vs ...float64) { + if len(vs) == 0 { + g.pg.Dec() + } + + g.Add(-sumFloat64(vs...)) +} + +func (g *gauge) Add(v float64) { + g.pg.Add(v) +} + +func (g *gauge) Set(v float64) { + g.pg.Set(v) +} + +func (g *gauge) Describe(c chan<- *prometheus.Desc) { + g.pg.Describe(c) +} + +func (g *gauge) Collect(c chan<- prometheus.Metric) { + g.pg.Collect(c) +} diff --git a/vendor/github.com/docker/go-metrics/handler.go b/vendor/github.com/docker/go-metrics/handler.go new file mode 100644 index 000000000..05601e9ec --- /dev/null +++ b/vendor/github.com/docker/go-metrics/handler.go @@ -0,0 +1,74 @@ +package metrics + +import ( + "net/http" + + "github.com/prometheus/client_golang/prometheus" + "github.com/prometheus/client_golang/prometheus/promhttp" +) + +// HTTPHandlerOpts describes a set of configurable options of http metrics +type HTTPHandlerOpts struct { + DurationBuckets []float64 + RequestSizeBuckets []float64 + ResponseSizeBuckets []float64 +} + +const ( + InstrumentHandlerResponseSize = iota + InstrumentHandlerRequestSize + InstrumentHandlerDuration + InstrumentHandlerCounter + InstrumentHandlerInFlight +) + +type HTTPMetric struct { + prometheus.Collector + handlerType int +} + +var ( + defaultDurationBuckets = []float64{.005, .01, .025, .05, .1, .25, .5, 1, 2.5, 5, 10, 25, 60} + defaultRequestSizeBuckets = prometheus.ExponentialBuckets(1024, 2, 22) //1K to 4G + defaultResponseSizeBuckets = defaultRequestSizeBuckets +) + +// Handler returns the global http.Handler that provides the prometheus +// metrics format on GET requests. This handler is no longer instrumented. +func Handler() http.Handler { + return promhttp.Handler() +} + +func InstrumentHandler(metrics []*HTTPMetric, handler http.Handler) http.HandlerFunc { + return InstrumentHandlerFunc(metrics, handler.ServeHTTP) +} + +func InstrumentHandlerFunc(metrics []*HTTPMetric, handlerFunc http.HandlerFunc) http.HandlerFunc { + var handler http.Handler + handler = http.HandlerFunc(handlerFunc) + for _, metric := range metrics { + switch metric.handlerType { + case InstrumentHandlerResponseSize: + if collector, ok := metric.Collector.(prometheus.ObserverVec); ok { + handler = promhttp.InstrumentHandlerResponseSize(collector, handler) + } + case InstrumentHandlerRequestSize: + if collector, ok := metric.Collector.(prometheus.ObserverVec); ok { + handler = promhttp.InstrumentHandlerRequestSize(collector, handler) + } + case InstrumentHandlerDuration: + if collector, ok := metric.Collector.(prometheus.ObserverVec); ok { + handler = promhttp.InstrumentHandlerDuration(collector, handler) + } + case InstrumentHandlerCounter: + if collector, ok := metric.Collector.(*prometheus.CounterVec); ok { + handler = promhttp.InstrumentHandlerCounter(collector, handler) + } + case InstrumentHandlerInFlight: + if collector, ok := metric.Collector.(prometheus.Gauge); ok { + handler = promhttp.InstrumentHandlerInFlight(collector, handler) + } + } + } + return handler.ServeHTTP +} diff --git a/vendor/github.com/docker/go-metrics/helpers.go b/vendor/github.com/docker/go-metrics/helpers.go new file mode 100644 index 000000000..68b7f51b3 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/helpers.go @@ -0,0 +1,10 @@ +package metrics + +func sumFloat64(vs ...float64) float64 { + var sum float64 + for _, v := range vs { + sum += v + } + + return sum +} diff --git a/vendor/github.com/docker/go-metrics/namespace.go b/vendor/github.com/docker/go-metrics/namespace.go new file mode 100644 index 000000000..798315451 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/namespace.go @@ -0,0 +1,315 @@ +package metrics + +import ( + "fmt" + "sync" + + "github.com/prometheus/client_golang/prometheus" +) + +type Labels map[string]string + +// NewNamespace returns a namespaces that is responsible for managing a collection of +// metrics for a particual namespace and subsystem +// +// labels allows const labels to be added to all metrics created in this namespace +// and are commonly used for data like application version and git commit +func NewNamespace(name, subsystem string, labels Labels) *Namespace { + if labels == nil { + labels = make(map[string]string) + } + return &Namespace{ + name: name, + subsystem: subsystem, + labels: labels, + } +} + +// Namespace describes a set of metrics that share a namespace and subsystem. +type Namespace struct { + name string + subsystem string + labels Labels + mu sync.Mutex + metrics []prometheus.Collector +} + +// WithConstLabels returns a namespace with the provided set of labels merged +// with the existing constant labels on the namespace. +// +// Only metrics created with the returned namespace will get the new constant +// labels. The returned namespace must be registered separately. +func (n *Namespace) WithConstLabels(labels Labels) *Namespace { + n.mu.Lock() + ns := &Namespace{ + name: n.name, + subsystem: n.subsystem, + labels: mergeLabels(n.labels, labels), + } + n.mu.Unlock() + return ns +} + +func (n *Namespace) NewCounter(name, help string) Counter { + c := &counter{pc: prometheus.NewCounter(n.newCounterOpts(name, help))} + n.Add(c) + return c +} + +func (n *Namespace) NewLabeledCounter(name, help string, labels ...string) LabeledCounter { + c := &labeledCounter{pc: prometheus.NewCounterVec(n.newCounterOpts(name, help), labels)} + n.Add(c) + return c +} + +func (n *Namespace) newCounterOpts(name, help string) prometheus.CounterOpts { + return prometheus.CounterOpts{ + Namespace: n.name, + Subsystem: n.subsystem, + Name: makeName(name, Total), + Help: help, + ConstLabels: prometheus.Labels(n.labels), + } +} + +func (n *Namespace) NewTimer(name, help string) Timer { + t := &timer{ + m: prometheus.NewHistogram(n.newTimerOpts(name, help)), + } + n.Add(t) + return t +} + +func (n *Namespace) NewLabeledTimer(name, help string, labels ...string) LabeledTimer { + t := &labeledTimer{ + m: prometheus.NewHistogramVec(n.newTimerOpts(name, help), labels), + } + n.Add(t) + return t +} + +func (n *Namespace) newTimerOpts(name, help string) prometheus.HistogramOpts { + return prometheus.HistogramOpts{ + Namespace: n.name, + Subsystem: n.subsystem, + Name: makeName(name, Seconds), + Help: help, + ConstLabels: prometheus.Labels(n.labels), + } +} + +func (n *Namespace) NewGauge(name, help string, unit Unit) Gauge { + g := &gauge{ + pg: prometheus.NewGauge(n.newGaugeOpts(name, help, unit)), + } + n.Add(g) + return g +} + +func (n *Namespace) NewLabeledGauge(name, help string, unit Unit, labels ...string) LabeledGauge { + g := &labeledGauge{ + pg: prometheus.NewGaugeVec(n.newGaugeOpts(name, help, unit), labels), + } + n.Add(g) + return g +} + +func (n *Namespace) newGaugeOpts(name, help string, unit Unit) prometheus.GaugeOpts { + return prometheus.GaugeOpts{ + Namespace: n.name, + Subsystem: n.subsystem, + Name: makeName(name, unit), + Help: help, + ConstLabels: prometheus.Labels(n.labels), + } +} + +func (n *Namespace) Describe(ch chan<- *prometheus.Desc) { + n.mu.Lock() + defer n.mu.Unlock() + + for _, metric := range n.metrics { + metric.Describe(ch) + } +} + +func (n *Namespace) Collect(ch chan<- prometheus.Metric) { + n.mu.Lock() + defer n.mu.Unlock() + + for _, metric := range n.metrics { + metric.Collect(ch) + } +} + +func (n *Namespace) Add(collector prometheus.Collector) { + n.mu.Lock() + n.metrics = append(n.metrics, collector) + n.mu.Unlock() +} + +func (n *Namespace) NewDesc(name, help string, unit Unit, labels ...string) *prometheus.Desc { + name = makeName(name, unit) + namespace := n.name + if n.subsystem != "" { + namespace = fmt.Sprintf("%s_%s", namespace, n.subsystem) + } + name = fmt.Sprintf("%s_%s", namespace, name) + return prometheus.NewDesc(name, help, labels, prometheus.Labels(n.labels)) +} + +// mergeLabels merges two or more labels objects into a single map, favoring +// the later labels. +func mergeLabels(lbs ...Labels) Labels { + merged := make(Labels) + + for _, target := range lbs { + for k, v := range target { + merged[k] = v + } + } + + return merged +} + +func makeName(name string, unit Unit) string { + if unit == "" { + return name + } + + return fmt.Sprintf("%s_%s", name, unit) +} + +func (n *Namespace) NewDefaultHttpMetrics(handlerName string) []*HTTPMetric { + return n.NewHttpMetricsWithOpts(handlerName, HTTPHandlerOpts{ + DurationBuckets: defaultDurationBuckets, + RequestSizeBuckets: defaultResponseSizeBuckets, + ResponseSizeBuckets: defaultResponseSizeBuckets, + }) +} + +func (n *Namespace) NewHttpMetrics(handlerName string, durationBuckets, requestSizeBuckets, responseSizeBuckets []float64) []*HTTPMetric { + return n.NewHttpMetricsWithOpts(handlerName, HTTPHandlerOpts{ + DurationBuckets: durationBuckets, + RequestSizeBuckets: requestSizeBuckets, + ResponseSizeBuckets: responseSizeBuckets, + }) +} + +func (n *Namespace) NewHttpMetricsWithOpts(handlerName string, opts HTTPHandlerOpts) []*HTTPMetric { + var httpMetrics []*HTTPMetric + inFlightMetric := n.NewInFlightGaugeMetric(handlerName) + requestTotalMetric := n.NewRequestTotalMetric(handlerName) + requestDurationMetric := n.NewRequestDurationMetric(handlerName, opts.DurationBuckets) + requestSizeMetric := n.NewRequestSizeMetric(handlerName, opts.RequestSizeBuckets) + responseSizeMetric := n.NewResponseSizeMetric(handlerName, opts.ResponseSizeBuckets) + httpMetrics = append(httpMetrics, inFlightMetric, requestDurationMetric, requestTotalMetric, requestSizeMetric, responseSizeMetric) + return httpMetrics +} + +func (n *Namespace) NewInFlightGaugeMetric(handlerName string) *HTTPMetric { + labels := prometheus.Labels(n.labels) + labels["handler"] = handlerName + metric := prometheus.NewGauge(prometheus.GaugeOpts{ + Namespace: n.name, + Subsystem: n.subsystem, + Name: "in_flight_requests", + Help: "The in-flight HTTP requests", + ConstLabels: prometheus.Labels(labels), + }) + httpMetric := &HTTPMetric{ + Collector: metric, + handlerType: InstrumentHandlerInFlight, + } + n.Add(httpMetric) + return httpMetric +} + +func (n *Namespace) NewRequestTotalMetric(handlerName string) *HTTPMetric { + labels := prometheus.Labels(n.labels) + labels["handler"] = handlerName + metric := prometheus.NewCounterVec( + prometheus.CounterOpts{ + Namespace: n.name, + Subsystem: n.subsystem, + Name: "requests_total", + Help: "Total number of HTTP requests made.", + ConstLabels: prometheus.Labels(labels), + }, + []string{"code", "method"}, + ) + httpMetric := &HTTPMetric{ + Collector: metric, + handlerType: InstrumentHandlerCounter, + } + n.Add(httpMetric) + return httpMetric +} +func (n *Namespace) NewRequestDurationMetric(handlerName string, buckets []float64) *HTTPMetric { + if len(buckets) == 0 { + panic("DurationBuckets must be provided") + } + labels := prometheus.Labels(n.labels) + labels["handler"] = handlerName + opts := prometheus.HistogramOpts{ + Namespace: n.name, + Subsystem: n.subsystem, + Name: "request_duration_seconds", + Help: "The HTTP request latencies in seconds.", + Buckets: buckets, + ConstLabels: prometheus.Labels(labels), + } + metric := prometheus.NewHistogramVec(opts, []string{"method"}) + httpMetric := &HTTPMetric{ + Collector: metric, + handlerType: InstrumentHandlerDuration, + } + n.Add(httpMetric) + return httpMetric +} + +func (n *Namespace) NewRequestSizeMetric(handlerName string, buckets []float64) *HTTPMetric { + if len(buckets) == 0 { + panic("RequestSizeBuckets must be provided") + } + labels := prometheus.Labels(n.labels) + labels["handler"] = handlerName + opts := prometheus.HistogramOpts{ + Namespace: n.name, + Subsystem: n.subsystem, + Name: "request_size_bytes", + Help: "The HTTP request sizes in bytes.", + Buckets: buckets, + ConstLabels: prometheus.Labels(labels), + } + metric := prometheus.NewHistogramVec(opts, []string{}) + httpMetric := &HTTPMetric{ + Collector: metric, + handlerType: InstrumentHandlerRequestSize, + } + n.Add(httpMetric) + return httpMetric +} + +func (n *Namespace) NewResponseSizeMetric(handlerName string, buckets []float64) *HTTPMetric { + if len(buckets) == 0 { + panic("ResponseSizeBuckets must be provided") + } + labels := prometheus.Labels(n.labels) + labels["handler"] = handlerName + opts := prometheus.HistogramOpts{ + Namespace: n.name, + Subsystem: n.subsystem, + Name: "response_size_bytes", + Help: "The HTTP response sizes in bytes.", + Buckets: buckets, + ConstLabels: prometheus.Labels(labels), + } + metrics := prometheus.NewHistogramVec(opts, []string{}) + httpMetric := &HTTPMetric{ + Collector: metrics, + handlerType: InstrumentHandlerResponseSize, + } + n.Add(httpMetric) + return httpMetric +} diff --git a/vendor/github.com/docker/go-metrics/register.go b/vendor/github.com/docker/go-metrics/register.go new file mode 100644 index 000000000..708358df0 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/register.go @@ -0,0 +1,15 @@ +package metrics + +import "github.com/prometheus/client_golang/prometheus" + +// Register adds all the metrics in the provided namespace to the global +// metrics registry +func Register(n *Namespace) { + prometheus.MustRegister(n) +} + +// Deregister removes all the metrics in the provided namespace from the +// global metrics registry +func Deregister(n *Namespace) { + prometheus.Unregister(n) +} diff --git a/vendor/github.com/docker/go-metrics/timer.go b/vendor/github.com/docker/go-metrics/timer.go new file mode 100644 index 000000000..824c98739 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/timer.go @@ -0,0 +1,85 @@ +package metrics + +import ( + "time" + + "github.com/prometheus/client_golang/prometheus" +) + +// StartTimer begins a timer observation at the callsite. When the target +// operation is completed, the caller should call the return done func(). +func StartTimer(timer Timer) (done func()) { + start := time.Now() + return func() { + timer.Update(time.Since(start)) + } +} + +// Timer is a metric that allows collecting the duration of an action in seconds +type Timer interface { + // Update records an observation, duration, and converts to the target + // units. + Update(duration time.Duration) + + // UpdateSince will add the duration from the provided starting time to the + // timer's summary with the precisions that was used in creation of the timer + UpdateSince(time.Time) +} + +// LabeledTimer is a timer that must have label values populated before use. +type LabeledTimer interface { + WithValues(labels ...string) *labeledTimerObserver +} + +type labeledTimer struct { + m *prometheus.HistogramVec +} + +type labeledTimerObserver struct { + m prometheus.Observer +} + +func (lbo *labeledTimerObserver) Update(duration time.Duration) { + lbo.m.Observe(duration.Seconds()) +} + +func (lbo *labeledTimerObserver) UpdateSince(since time.Time) { + lbo.m.Observe(time.Since(since).Seconds()) +} + +func (lt *labeledTimer) WithValues(labels ...string) *labeledTimerObserver { + return &labeledTimerObserver{m: lt.m.WithLabelValues(labels...)} +} + +func (lt *labeledTimer) Describe(c chan<- *prometheus.Desc) { + lt.m.Describe(c) +} + +func (lt *labeledTimer) Collect(c chan<- prometheus.Metric) { + lt.m.Collect(c) +} + +type timer struct { + m prometheus.Observer +} + +func (t *timer) Update(duration time.Duration) { + t.m.Observe(duration.Seconds()) +} + +func (t *timer) UpdateSince(since time.Time) { + t.m.Observe(time.Since(since).Seconds()) +} + +func (t *timer) Describe(c chan<- *prometheus.Desc) { + c <- t.m.(prometheus.Metric).Desc() +} + +func (t *timer) Collect(c chan<- prometheus.Metric) { + // Are there any observers that don't implement Collector? It is really + // unclear what the point of the upstream change was, but we'll let this + // panic if we get an observer that doesn't implement collector. In this + // case, we should almost always see metricVec objects, so this should + // never panic. + t.m.(prometheus.Collector).Collect(c) +} diff --git a/vendor/github.com/docker/go-metrics/unit.go b/vendor/github.com/docker/go-metrics/unit.go new file mode 100644 index 000000000..c96622f90 --- /dev/null +++ b/vendor/github.com/docker/go-metrics/unit.go @@ -0,0 +1,12 @@ +package metrics + +// Unit represents the type or precision of a metric that is appended to +// the metrics fully qualified name +type Unit string + +const ( + Nanoseconds Unit = "nanoseconds" + Seconds Unit = "seconds" + Bytes Unit = "bytes" + Total Unit = "total" +) diff --git a/vendor/github.com/docker/libnetwork/LICENSE b/vendor/github.com/docker/libnetwork/LICENSE new file mode 100644 index 000000000..5c304d1a4 --- /dev/null +++ b/vendor/github.com/docker/libnetwork/LICENSE @@ -0,0 +1,201 @@ +Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "{}" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright {yyyy} {name of copyright owner} + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/vendor/github.com/docker/libnetwork/ipamutils/utils.go b/vendor/github.com/docker/libnetwork/ipamutils/utils.go new file mode 100644 index 000000000..f8eca58e8 --- /dev/null +++ b/vendor/github.com/docker/libnetwork/ipamutils/utils.go @@ -0,0 +1,96 @@ +// Package ipamutils provides utility functions for ipam management +package ipamutils + +import ( + "fmt" + "net" + "sync" + + "github.com/sirupsen/logrus" +) + +var ( + // PredefinedBroadNetworks contains a list of 31 IPv4 private networks with host size 16 and 12 + // (172.17-31.x.x/16, 192.168.x.x/20) which do not overlap with the networks in `PredefinedGranularNetworks` + PredefinedBroadNetworks []*net.IPNet + // PredefinedGranularNetworks contains a list of 64K IPv4 private networks with host size 8 + // (10.x.x.x/24) which do not overlap with the networks in `PredefinedBroadNetworks` + PredefinedGranularNetworks []*net.IPNet + initNetworksOnce sync.Once + + defaultBroadNetwork = []*NetworkToSplit{{"172.17.0.0/16", 16}, {"172.18.0.0/16", 16}, {"172.19.0.0/16", 16}, + {"172.20.0.0/14", 16}, {"172.24.0.0/14", 16}, {"172.28.0.0/14", 16}, + {"192.168.0.0/16", 20}} + defaultGranularNetwork = []*NetworkToSplit{{"10.0.0.0/8", 24}} +) + +// NetworkToSplit represent a network that has to be split in chunks with mask length Size. +// Each subnet in the set is derived from the Base pool. Base is to be passed +// in CIDR format. +// Example: a Base "10.10.0.0/16 with Size 24 will define the set of 256 +// 10.10.[0-255].0/24 address pools +type NetworkToSplit struct { + Base string `json:"base"` + Size int `json:"size"` +} + +// InitNetworks initializes the broad network pool and the granular network pool +func InitNetworks(defaultAddressPool []*NetworkToSplit) { + initNetworksOnce.Do(func() { + // error ingnored should never fail + PredefinedGranularNetworks, _ = splitNetworks(defaultGranularNetwork) + if defaultAddressPool == nil { + defaultAddressPool = defaultBroadNetwork + } + var err error + if PredefinedBroadNetworks, err = splitNetworks(defaultAddressPool); err != nil { + logrus.WithError(err).Error("InitAddressPools failed to initialize the default address pool") + } + }) +} + +// splitNetworks takes a slice of networks, split them accordingly and returns them +func splitNetworks(list []*NetworkToSplit) ([]*net.IPNet, error) { + localPools := make([]*net.IPNet, 0, len(list)) + + for _, p := range list { + _, b, err := net.ParseCIDR(p.Base) + if err != nil { + return nil, fmt.Errorf("invalid base pool %q: %v", p.Base, err) + } + ones, _ := b.Mask.Size() + if p.Size <= 0 || p.Size < ones { + return nil, fmt.Errorf("invalid pools size: %d", p.Size) + } + localPools = append(localPools, splitNetwork(p.Size, b)...) + } + return localPools, nil +} + +func splitNetwork(size int, base *net.IPNet) []*net.IPNet { + one, bits := base.Mask.Size() + mask := net.CIDRMask(size, bits) + n := 1 << uint(size-one) + s := uint(bits - size) + list := make([]*net.IPNet, 0, n) + + for i := 0; i < n; i++ { + ip := copyIP(base.IP) + addIntToIP(ip, uint(i<= 0; i-- { + array[i] |= (byte)(ordinal & 0xff) + ordinal >>= 8 + } +} diff --git a/vendor/github.com/dustin/go-humanize/.gitignore b/vendor/github.com/dustin/go-humanize/.gitignore deleted file mode 100644 index 05b40514a..000000000 --- a/vendor/github.com/dustin/go-humanize/.gitignore +++ /dev/null @@ -1,6 +0,0 @@ -#* -*.[568] -*.a -*~ -[568].out -_* diff --git a/vendor/github.com/go-ini/ini/.gitignore b/vendor/github.com/go-ini/ini/.gitignore deleted file mode 100644 index 7adca9439..000000000 --- a/vendor/github.com/go-ini/ini/.gitignore +++ /dev/null @@ -1,4 +0,0 @@ -testdata/conf_out.ini -ini.sublime-project -ini.sublime-workspace -testdata/conf_reflect.ini diff --git a/vendor/github.com/go-ole/go-ole/oleutil/connection_windows.go b/vendor/github.com/go-ole/go-ole/oleutil/connection_windows.go index ab9c0d8dc..6b5c05999 100644 --- a/vendor/github.com/go-ole/go-ole/oleutil/connection_windows.go +++ b/vendor/github.com/go-ole/go-ole/oleutil/connection_windows.go @@ -49,7 +49,6 @@ func ConnectObject(disp *ole.IDispatch, iid *ole.GUID, idisp interface{}) (cooki point.Release() return } - return } container.Release() diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/godbus/dbus/LICENSE b/vendor/github.com/godbus/dbus/LICENSE similarity index 100% rename from vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/godbus/dbus/LICENSE rename to vendor/github.com/godbus/dbus/LICENSE diff --git a/vendor/github.com/godbus/dbus/README.markdown b/vendor/github.com/godbus/dbus/README.markdown new file mode 100644 index 000000000..bde2a39c3 --- /dev/null +++ b/vendor/github.com/godbus/dbus/README.markdown @@ -0,0 +1,41 @@ +dbus +---- + +dbus is a simple library that implements native Go client bindings for the +D-Bus message bus system. + +### Features + +* Complete native implementation of the D-Bus message protocol +* Go-like API (channels for signals / asynchronous method calls, Goroutine-safe connections) +* Subpackages that help with the introspection / property interfaces + +### Installation + +This packages requires Go 1.1. If you installed it and set up your GOPATH, just run: + +``` +go get github.com/godbus/dbus +``` + +If you want to use the subpackages, you can install them the same way. + +### Usage + +The complete package documentation and some simple examples are available at +[godoc.org](http://godoc.org/github.com/godbus/dbus). Also, the +[_examples](https://github.com/godbus/dbus/tree/master/_examples) directory +gives a short overview over the basic usage. + +#### Projects using godbus +- [notify](https://github.com/esiqveland/notify) provides desktop notifications over dbus into a library. + +Please note that the API is considered unstable for now and may change without +further notice. + +### License + +go.dbus is available under the Simplified BSD License; see LICENSE for the full +text. + +Nearly all of the credit for this library goes to github.com/guelfey/go.dbus. diff --git a/vendor/github.com/godbus/dbus/auth.go b/vendor/github.com/godbus/dbus/auth.go new file mode 100644 index 000000000..98017b693 --- /dev/null +++ b/vendor/github.com/godbus/dbus/auth.go @@ -0,0 +1,253 @@ +package dbus + +import ( + "bufio" + "bytes" + "errors" + "io" + "os" + "strconv" +) + +// AuthStatus represents the Status of an authentication mechanism. +type AuthStatus byte + +const ( + // AuthOk signals that authentication is finished; the next command + // from the server should be an OK. + AuthOk AuthStatus = iota + + // AuthContinue signals that additional data is needed; the next command + // from the server should be a DATA. + AuthContinue + + // AuthError signals an error; the server sent invalid data or some + // other unexpected thing happened and the current authentication + // process should be aborted. + AuthError +) + +type authState byte + +const ( + waitingForData authState = iota + waitingForOk + waitingForReject +) + +// Auth defines the behaviour of an authentication mechanism. +type Auth interface { + // Return the name of the mechnism, the argument to the first AUTH command + // and the next status. + FirstData() (name, resp []byte, status AuthStatus) + + // Process the given DATA command, and return the argument to the DATA + // command and the next status. If len(resp) == 0, no DATA command is sent. + HandleData(data []byte) (resp []byte, status AuthStatus) +} + +// Auth authenticates the connection, trying the given list of authentication +// mechanisms (in that order). If nil is passed, the EXTERNAL and +// DBUS_COOKIE_SHA1 mechanisms are tried for the current user. For private +// connections, this method must be called before sending any messages to the +// bus. Auth must not be called on shared connections. +func (conn *Conn) Auth(methods []Auth) error { + if methods == nil { + uid := strconv.Itoa(os.Getuid()) + methods = []Auth{AuthExternal(uid), AuthCookieSha1(uid, getHomeDir())} + } + in := bufio.NewReader(conn.transport) + err := conn.transport.SendNullByte() + if err != nil { + return err + } + err = authWriteLine(conn.transport, []byte("AUTH")) + if err != nil { + return err + } + s, err := authReadLine(in) + if err != nil { + return err + } + if len(s) < 2 || !bytes.Equal(s[0], []byte("REJECTED")) { + return errors.New("dbus: authentication protocol error") + } + s = s[1:] + for _, v := range s { + for _, m := range methods { + if name, data, status := m.FirstData(); bytes.Equal(v, name) { + var ok bool + err = authWriteLine(conn.transport, []byte("AUTH"), []byte(v), data) + if err != nil { + return err + } + switch status { + case AuthOk: + err, ok = conn.tryAuth(m, waitingForOk, in) + case AuthContinue: + err, ok = conn.tryAuth(m, waitingForData, in) + default: + panic("dbus: invalid authentication status") + } + if err != nil { + return err + } + if ok { + if conn.transport.SupportsUnixFDs() { + err = authWriteLine(conn, []byte("NEGOTIATE_UNIX_FD")) + if err != nil { + return err + } + line, err := authReadLine(in) + if err != nil { + return err + } + switch { + case bytes.Equal(line[0], []byte("AGREE_UNIX_FD")): + conn.EnableUnixFDs() + conn.unixFD = true + case bytes.Equal(line[0], []byte("ERROR")): + default: + return errors.New("dbus: authentication protocol error") + } + } + err = authWriteLine(conn.transport, []byte("BEGIN")) + if err != nil { + return err + } + go conn.inWorker() + go conn.outWorker() + return nil + } + } + } + } + return errors.New("dbus: authentication failed") +} + +// tryAuth tries to authenticate with m as the mechanism, using state as the +// initial authState and in for reading input. It returns (nil, true) on +// success, (nil, false) on a REJECTED and (someErr, false) if some other +// error occured. +func (conn *Conn) tryAuth(m Auth, state authState, in *bufio.Reader) (error, bool) { + for { + s, err := authReadLine(in) + if err != nil { + return err, false + } + switch { + case state == waitingForData && string(s[0]) == "DATA": + if len(s) != 2 { + err = authWriteLine(conn.transport, []byte("ERROR")) + if err != nil { + return err, false + } + continue + } + data, status := m.HandleData(s[1]) + switch status { + case AuthOk, AuthContinue: + if len(data) != 0 { + err = authWriteLine(conn.transport, []byte("DATA"), data) + if err != nil { + return err, false + } + } + if status == AuthOk { + state = waitingForOk + } + case AuthError: + err = authWriteLine(conn.transport, []byte("ERROR")) + if err != nil { + return err, false + } + } + case state == waitingForData && string(s[0]) == "REJECTED": + return nil, false + case state == waitingForData && string(s[0]) == "ERROR": + err = authWriteLine(conn.transport, []byte("CANCEL")) + if err != nil { + return err, false + } + state = waitingForReject + case state == waitingForData && string(s[0]) == "OK": + if len(s) != 2 { + err = authWriteLine(conn.transport, []byte("CANCEL")) + if err != nil { + return err, false + } + state = waitingForReject + } + conn.uuid = string(s[1]) + return nil, true + case state == waitingForData: + err = authWriteLine(conn.transport, []byte("ERROR")) + if err != nil { + return err, false + } + case state == waitingForOk && string(s[0]) == "OK": + if len(s) != 2 { + err = authWriteLine(conn.transport, []byte("CANCEL")) + if err != nil { + return err, false + } + state = waitingForReject + } + conn.uuid = string(s[1]) + return nil, true + case state == waitingForOk && string(s[0]) == "REJECTED": + return nil, false + case state == waitingForOk && (string(s[0]) == "DATA" || + string(s[0]) == "ERROR"): + + err = authWriteLine(conn.transport, []byte("CANCEL")) + if err != nil { + return err, false + } + state = waitingForReject + case state == waitingForOk: + err = authWriteLine(conn.transport, []byte("ERROR")) + if err != nil { + return err, false + } + case state == waitingForReject && string(s[0]) == "REJECTED": + return nil, false + case state == waitingForReject: + return errors.New("dbus: authentication protocol error"), false + default: + panic("dbus: invalid auth state") + } + } +} + +// authReadLine reads a line and separates it into its fields. +func authReadLine(in *bufio.Reader) ([][]byte, error) { + data, err := in.ReadBytes('\n') + if err != nil { + return nil, err + } + data = bytes.TrimSuffix(data, []byte("\r\n")) + return bytes.Split(data, []byte{' '}), nil +} + +// authWriteLine writes the given line in the authentication protocol format +// (elements of data separated by a " " and terminated by "\r\n"). +func authWriteLine(out io.Writer, data ...[]byte) error { + buf := make([]byte, 0) + for i, v := range data { + buf = append(buf, v...) + if i != len(data)-1 { + buf = append(buf, ' ') + } + } + buf = append(buf, '\r') + buf = append(buf, '\n') + n, err := out.Write(buf) + if err != nil { + return err + } + if n != len(buf) { + return io.ErrUnexpectedEOF + } + return nil +} diff --git a/vendor/github.com/godbus/dbus/auth_external.go b/vendor/github.com/godbus/dbus/auth_external.go new file mode 100644 index 000000000..7e376d3ef --- /dev/null +++ b/vendor/github.com/godbus/dbus/auth_external.go @@ -0,0 +1,26 @@ +package dbus + +import ( + "encoding/hex" +) + +// AuthExternal returns an Auth that authenticates as the given user with the +// EXTERNAL mechanism. +func AuthExternal(user string) Auth { + return authExternal{user} +} + +// AuthExternal implements the EXTERNAL authentication mechanism. +type authExternal struct { + user string +} + +func (a authExternal) FirstData() ([]byte, []byte, AuthStatus) { + b := make([]byte, 2*len(a.user)) + hex.Encode(b, []byte(a.user)) + return []byte("EXTERNAL"), b, AuthOk +} + +func (a authExternal) HandleData(b []byte) ([]byte, AuthStatus) { + return nil, AuthError +} diff --git a/vendor/github.com/godbus/dbus/auth_sha1.go b/vendor/github.com/godbus/dbus/auth_sha1.go new file mode 100644 index 000000000..df15b4611 --- /dev/null +++ b/vendor/github.com/godbus/dbus/auth_sha1.go @@ -0,0 +1,102 @@ +package dbus + +import ( + "bufio" + "bytes" + "crypto/rand" + "crypto/sha1" + "encoding/hex" + "os" +) + +// AuthCookieSha1 returns an Auth that authenticates as the given user with the +// DBUS_COOKIE_SHA1 mechanism. The home parameter should specify the home +// directory of the user. +func AuthCookieSha1(user, home string) Auth { + return authCookieSha1{user, home} +} + +type authCookieSha1 struct { + user, home string +} + +func (a authCookieSha1) FirstData() ([]byte, []byte, AuthStatus) { + b := make([]byte, 2*len(a.user)) + hex.Encode(b, []byte(a.user)) + return []byte("DBUS_COOKIE_SHA1"), b, AuthContinue +} + +func (a authCookieSha1) HandleData(data []byte) ([]byte, AuthStatus) { + challenge := make([]byte, len(data)/2) + _, err := hex.Decode(challenge, data) + if err != nil { + return nil, AuthError + } + b := bytes.Split(challenge, []byte{' '}) + if len(b) != 3 { + return nil, AuthError + } + context := b[0] + id := b[1] + svchallenge := b[2] + cookie := a.getCookie(context, id) + if cookie == nil { + return nil, AuthError + } + clchallenge := a.generateChallenge() + if clchallenge == nil { + return nil, AuthError + } + hash := sha1.New() + hash.Write(bytes.Join([][]byte{svchallenge, clchallenge, cookie}, []byte{':'})) + hexhash := make([]byte, 2*hash.Size()) + hex.Encode(hexhash, hash.Sum(nil)) + data = append(clchallenge, ' ') + data = append(data, hexhash...) + resp := make([]byte, 2*len(data)) + hex.Encode(resp, data) + return resp, AuthOk +} + +// getCookie searches for the cookie identified by id in context and returns +// the cookie content or nil. (Since HandleData can't return a specific error, +// but only whether an error occured, this function also doesn't bother to +// return an error.) +func (a authCookieSha1) getCookie(context, id []byte) []byte { + file, err := os.Open(a.home + "/.dbus-keyrings/" + string(context)) + if err != nil { + return nil + } + defer file.Close() + rd := bufio.NewReader(file) + for { + line, err := rd.ReadBytes('\n') + if err != nil { + return nil + } + line = line[:len(line)-1] + b := bytes.Split(line, []byte{' '}) + if len(b) != 3 { + return nil + } + if bytes.Equal(b[0], id) { + return b[2] + } + } +} + +// generateChallenge returns a random, hex-encoded challenge, or nil on error +// (see above). +func (a authCookieSha1) generateChallenge() []byte { + b := make([]byte, 16) + n, err := rand.Read(b) + if err != nil { + return nil + } + if n != 16 { + return nil + } + enc := make([]byte, 32) + hex.Encode(enc, b) + return enc +} diff --git a/vendor/github.com/godbus/dbus/call.go b/vendor/github.com/godbus/dbus/call.go new file mode 100644 index 000000000..ba6e73f60 --- /dev/null +++ b/vendor/github.com/godbus/dbus/call.go @@ -0,0 +1,36 @@ +package dbus + +import ( + "errors" +) + +// Call represents a pending or completed method call. +type Call struct { + Destination string + Path ObjectPath + Method string + Args []interface{} + + // Strobes when the call is complete. + Done chan *Call + + // After completion, the error status. If this is non-nil, it may be an + // error message from the peer (with Error as its type) or some other error. + Err error + + // Holds the response once the call is done. + Body []interface{} +} + +var errSignature = errors.New("dbus: mismatched signature") + +// Store stores the body of the reply into the provided pointers. It returns +// an error if the signatures of the body and retvalues don't match, or if +// the error status is not nil. +func (c *Call) Store(retvalues ...interface{}) error { + if c.Err != nil { + return c.Err + } + + return Store(c.Body, retvalues...) +} diff --git a/vendor/github.com/godbus/dbus/conn.go b/vendor/github.com/godbus/dbus/conn.go new file mode 100644 index 000000000..a4f539401 --- /dev/null +++ b/vendor/github.com/godbus/dbus/conn.go @@ -0,0 +1,625 @@ +package dbus + +import ( + "errors" + "io" + "os" + "reflect" + "strings" + "sync" +) + +const defaultSystemBusAddress = "unix:path=/var/run/dbus/system_bus_socket" + +var ( + systemBus *Conn + systemBusLck sync.Mutex + sessionBus *Conn + sessionBusLck sync.Mutex +) + +// ErrClosed is the error returned by calls on a closed connection. +var ErrClosed = errors.New("dbus: connection closed by user") + +// Conn represents a connection to a message bus (usually, the system or +// session bus). +// +// Connections are either shared or private. Shared connections +// are shared between calls to the functions that return them. As a result, +// the methods Close, Auth and Hello must not be called on them. +// +// Multiple goroutines may invoke methods on a connection simultaneously. +type Conn struct { + transport + + busObj BusObject + unixFD bool + uuid string + + names []string + namesLck sync.RWMutex + + serialLck sync.Mutex + nextSerial uint32 + serialUsed map[uint32]bool + + calls map[uint32]*Call + callsLck sync.RWMutex + + handlers map[ObjectPath]map[string]exportWithMapping + handlersLck sync.RWMutex + + out chan *Message + closed bool + outLck sync.RWMutex + + signals []chan<- *Signal + signalsLck sync.Mutex + + eavesdropped chan<- *Message + eavesdroppedLck sync.Mutex +} + +// SessionBus returns a shared connection to the session bus, connecting to it +// if not already done. +func SessionBus() (conn *Conn, err error) { + sessionBusLck.Lock() + defer sessionBusLck.Unlock() + if sessionBus != nil { + return sessionBus, nil + } + defer func() { + if conn != nil { + sessionBus = conn + } + }() + conn, err = SessionBusPrivate() + if err != nil { + return + } + if err = conn.Auth(nil); err != nil { + conn.Close() + conn = nil + return + } + if err = conn.Hello(); err != nil { + conn.Close() + conn = nil + } + return +} + +// SessionBusPrivate returns a new private connection to the session bus. +func SessionBusPrivate() (*Conn, error) { + address := os.Getenv("DBUS_SESSION_BUS_ADDRESS") + if address != "" && address != "autolaunch:" { + return Dial(address) + } + + return sessionBusPlatform() +} + +// SystemBus returns a shared connection to the system bus, connecting to it if +// not already done. +func SystemBus() (conn *Conn, err error) { + systemBusLck.Lock() + defer systemBusLck.Unlock() + if systemBus != nil { + return systemBus, nil + } + defer func() { + if conn != nil { + systemBus = conn + } + }() + conn, err = SystemBusPrivate() + if err != nil { + return + } + if err = conn.Auth(nil); err != nil { + conn.Close() + conn = nil + return + } + if err = conn.Hello(); err != nil { + conn.Close() + conn = nil + } + return +} + +// SystemBusPrivate returns a new private connection to the system bus. +func SystemBusPrivate() (*Conn, error) { + address := os.Getenv("DBUS_SYSTEM_BUS_ADDRESS") + if address != "" { + return Dial(address) + } + return Dial(defaultSystemBusAddress) +} + +// Dial establishes a new private connection to the message bus specified by address. +func Dial(address string) (*Conn, error) { + tr, err := getTransport(address) + if err != nil { + return nil, err + } + return newConn(tr) +} + +// NewConn creates a new private *Conn from an already established connection. +func NewConn(conn io.ReadWriteCloser) (*Conn, error) { + return newConn(genericTransport{conn}) +} + +// newConn creates a new *Conn from a transport. +func newConn(tr transport) (*Conn, error) { + conn := new(Conn) + conn.transport = tr + conn.calls = make(map[uint32]*Call) + conn.out = make(chan *Message, 10) + conn.handlers = make(map[ObjectPath]map[string]exportWithMapping) + conn.nextSerial = 1 + conn.serialUsed = map[uint32]bool{0: true} + conn.busObj = conn.Object("org.freedesktop.DBus", "/org/freedesktop/DBus") + return conn, nil +} + +// BusObject returns the object owned by the bus daemon which handles +// administrative requests. +func (conn *Conn) BusObject() BusObject { + return conn.busObj +} + +// Close closes the connection. Any blocked operations will return with errors +// and the channels passed to Eavesdrop and Signal are closed. This method must +// not be called on shared connections. +func (conn *Conn) Close() error { + conn.outLck.Lock() + if conn.closed { + // inWorker calls Close on read error, the read error may + // be caused by another caller calling Close to shutdown the + // dbus connection, a double-close scenario we prevent here. + conn.outLck.Unlock() + return nil + } + close(conn.out) + conn.closed = true + conn.outLck.Unlock() + conn.signalsLck.Lock() + for _, ch := range conn.signals { + close(ch) + } + conn.signalsLck.Unlock() + conn.eavesdroppedLck.Lock() + if conn.eavesdropped != nil { + close(conn.eavesdropped) + } + conn.eavesdroppedLck.Unlock() + return conn.transport.Close() +} + +// Eavesdrop causes conn to send all incoming messages to the given channel +// without further processing. Method replies, errors and signals will not be +// sent to the appropiate channels and method calls will not be handled. If nil +// is passed, the normal behaviour is restored. +// +// The caller has to make sure that ch is sufficiently buffered; +// if a message arrives when a write to ch is not possible, the message is +// discarded. +func (conn *Conn) Eavesdrop(ch chan<- *Message) { + conn.eavesdroppedLck.Lock() + conn.eavesdropped = ch + conn.eavesdroppedLck.Unlock() +} + +// getSerial returns an unused serial. +func (conn *Conn) getSerial() uint32 { + conn.serialLck.Lock() + defer conn.serialLck.Unlock() + n := conn.nextSerial + for conn.serialUsed[n] { + n++ + } + conn.serialUsed[n] = true + conn.nextSerial = n + 1 + return n +} + +// Hello sends the initial org.freedesktop.DBus.Hello call. This method must be +// called after authentication, but before sending any other messages to the +// bus. Hello must not be called for shared connections. +func (conn *Conn) Hello() error { + var s string + err := conn.busObj.Call("org.freedesktop.DBus.Hello", 0).Store(&s) + if err != nil { + return err + } + conn.namesLck.Lock() + conn.names = make([]string, 1) + conn.names[0] = s + conn.namesLck.Unlock() + return nil +} + +// inWorker runs in an own goroutine, reading incoming messages from the +// transport and dispatching them appropiately. +func (conn *Conn) inWorker() { + for { + msg, err := conn.ReadMessage() + if err == nil { + conn.eavesdroppedLck.Lock() + if conn.eavesdropped != nil { + select { + case conn.eavesdropped <- msg: + default: + } + conn.eavesdroppedLck.Unlock() + continue + } + conn.eavesdroppedLck.Unlock() + dest, _ := msg.Headers[FieldDestination].value.(string) + found := false + if dest == "" { + found = true + } else { + conn.namesLck.RLock() + if len(conn.names) == 0 { + found = true + } + for _, v := range conn.names { + if dest == v { + found = true + break + } + } + conn.namesLck.RUnlock() + } + if !found { + // Eavesdropped a message, but no channel for it is registered. + // Ignore it. + continue + } + switch msg.Type { + case TypeMethodReply, TypeError: + serial := msg.Headers[FieldReplySerial].value.(uint32) + conn.callsLck.Lock() + if c, ok := conn.calls[serial]; ok { + if msg.Type == TypeError { + name, _ := msg.Headers[FieldErrorName].value.(string) + c.Err = Error{name, msg.Body} + } else { + c.Body = msg.Body + } + c.Done <- c + conn.serialLck.Lock() + delete(conn.serialUsed, serial) + conn.serialLck.Unlock() + delete(conn.calls, serial) + } + conn.callsLck.Unlock() + case TypeSignal: + iface := msg.Headers[FieldInterface].value.(string) + member := msg.Headers[FieldMember].value.(string) + // as per http://dbus.freedesktop.org/doc/dbus-specification.html , + // sender is optional for signals. + sender, _ := msg.Headers[FieldSender].value.(string) + if iface == "org.freedesktop.DBus" && sender == "org.freedesktop.DBus" { + if member == "NameLost" { + // If we lost the name on the bus, remove it from our + // tracking list. + name, ok := msg.Body[0].(string) + if !ok { + panic("Unable to read the lost name") + } + conn.namesLck.Lock() + for i, v := range conn.names { + if v == name { + conn.names = append(conn.names[:i], + conn.names[i+1:]...) + } + } + conn.namesLck.Unlock() + } else if member == "NameAcquired" { + // If we acquired the name on the bus, add it to our + // tracking list. + name, ok := msg.Body[0].(string) + if !ok { + panic("Unable to read the acquired name") + } + conn.namesLck.Lock() + conn.names = append(conn.names, name) + conn.namesLck.Unlock() + } + } + signal := &Signal{ + Sender: sender, + Path: msg.Headers[FieldPath].value.(ObjectPath), + Name: iface + "." + member, + Body: msg.Body, + } + conn.signalsLck.Lock() + for _, ch := range conn.signals { + ch <- signal + } + conn.signalsLck.Unlock() + case TypeMethodCall: + go conn.handleCall(msg) + } + } else if _, ok := err.(InvalidMessageError); !ok { + // Some read error occured (usually EOF); we can't really do + // anything but to shut down all stuff and returns errors to all + // pending replies. + conn.Close() + conn.callsLck.RLock() + for _, v := range conn.calls { + v.Err = err + v.Done <- v + } + conn.callsLck.RUnlock() + return + } + // invalid messages are ignored + } +} + +// Names returns the list of all names that are currently owned by this +// connection. The slice is always at least one element long, the first element +// being the unique name of the connection. +func (conn *Conn) Names() []string { + conn.namesLck.RLock() + // copy the slice so it can't be modified + s := make([]string, len(conn.names)) + copy(s, conn.names) + conn.namesLck.RUnlock() + return s +} + +// Object returns the object identified by the given destination name and path. +func (conn *Conn) Object(dest string, path ObjectPath) BusObject { + return &Object{conn, dest, path} +} + +// outWorker runs in an own goroutine, encoding and sending messages that are +// sent to conn.out. +func (conn *Conn) outWorker() { + for msg := range conn.out { + err := conn.SendMessage(msg) + conn.callsLck.RLock() + if err != nil { + if c := conn.calls[msg.serial]; c != nil { + c.Err = err + c.Done <- c + } + conn.serialLck.Lock() + delete(conn.serialUsed, msg.serial) + conn.serialLck.Unlock() + } else if msg.Type != TypeMethodCall { + conn.serialLck.Lock() + delete(conn.serialUsed, msg.serial) + conn.serialLck.Unlock() + } + conn.callsLck.RUnlock() + } +} + +// Send sends the given message to the message bus. You usually don't need to +// use this; use the higher-level equivalents (Call / Go, Emit and Export) +// instead. If msg is a method call and NoReplyExpected is not set, a non-nil +// call is returned and the same value is sent to ch (which must be buffered) +// once the call is complete. Otherwise, ch is ignored and a Call structure is +// returned of which only the Err member is valid. +func (conn *Conn) Send(msg *Message, ch chan *Call) *Call { + var call *Call + + msg.serial = conn.getSerial() + if msg.Type == TypeMethodCall && msg.Flags&FlagNoReplyExpected == 0 { + if ch == nil { + ch = make(chan *Call, 5) + } else if cap(ch) == 0 { + panic("dbus: unbuffered channel passed to (*Conn).Send") + } + call = new(Call) + call.Destination, _ = msg.Headers[FieldDestination].value.(string) + call.Path, _ = msg.Headers[FieldPath].value.(ObjectPath) + iface, _ := msg.Headers[FieldInterface].value.(string) + member, _ := msg.Headers[FieldMember].value.(string) + call.Method = iface + "." + member + call.Args = msg.Body + call.Done = ch + conn.callsLck.Lock() + conn.calls[msg.serial] = call + conn.callsLck.Unlock() + conn.outLck.RLock() + if conn.closed { + call.Err = ErrClosed + call.Done <- call + } else { + conn.out <- msg + } + conn.outLck.RUnlock() + } else { + conn.outLck.RLock() + if conn.closed { + call = &Call{Err: ErrClosed} + } else { + conn.out <- msg + call = &Call{Err: nil} + } + conn.outLck.RUnlock() + } + return call +} + +// sendError creates an error message corresponding to the parameters and sends +// it to conn.out. +func (conn *Conn) sendError(e Error, dest string, serial uint32) { + msg := new(Message) + msg.Type = TypeError + msg.serial = conn.getSerial() + msg.Headers = make(map[HeaderField]Variant) + if dest != "" { + msg.Headers[FieldDestination] = MakeVariant(dest) + } + msg.Headers[FieldErrorName] = MakeVariant(e.Name) + msg.Headers[FieldReplySerial] = MakeVariant(serial) + msg.Body = e.Body + if len(e.Body) > 0 { + msg.Headers[FieldSignature] = MakeVariant(SignatureOf(e.Body...)) + } + conn.outLck.RLock() + if !conn.closed { + conn.out <- msg + } + conn.outLck.RUnlock() +} + +// sendReply creates a method reply message corresponding to the parameters and +// sends it to conn.out. +func (conn *Conn) sendReply(dest string, serial uint32, values ...interface{}) { + msg := new(Message) + msg.Type = TypeMethodReply + msg.serial = conn.getSerial() + msg.Headers = make(map[HeaderField]Variant) + if dest != "" { + msg.Headers[FieldDestination] = MakeVariant(dest) + } + msg.Headers[FieldReplySerial] = MakeVariant(serial) + msg.Body = values + if len(values) > 0 { + msg.Headers[FieldSignature] = MakeVariant(SignatureOf(values...)) + } + conn.outLck.RLock() + if !conn.closed { + conn.out <- msg + } + conn.outLck.RUnlock() +} + +// Signal registers the given channel to be passed all received signal messages. +// The caller has to make sure that ch is sufficiently buffered; if a message +// arrives when a write to c is not possible, it is discarded. +// +// Multiple of these channels can be registered at the same time. Passing a +// channel that already is registered will remove it from the list of the +// registered channels. +// +// These channels are "overwritten" by Eavesdrop; i.e., if there currently is a +// channel for eavesdropped messages, this channel receives all signals, and +// none of the channels passed to Signal will receive any signals. +func (conn *Conn) Signal(ch chan<- *Signal) { + conn.signalsLck.Lock() + conn.signals = append(conn.signals, ch) + conn.signalsLck.Unlock() +} + +// SupportsUnixFDs returns whether the underlying transport supports passing of +// unix file descriptors. If this is false, method calls containing unix file +// descriptors will return an error and emitted signals containing them will +// not be sent. +func (conn *Conn) SupportsUnixFDs() bool { + return conn.unixFD +} + +// Error represents a D-Bus message of type Error. +type Error struct { + Name string + Body []interface{} +} + +func NewError(name string, body []interface{}) *Error { + return &Error{name, body} +} + +func (e Error) Error() string { + if len(e.Body) >= 1 { + s, ok := e.Body[0].(string) + if ok { + return s + } + } + return e.Name +} + +// Signal represents a D-Bus message of type Signal. The name member is given in +// "interface.member" notation, e.g. org.freedesktop.D-Bus.NameLost. +type Signal struct { + Sender string + Path ObjectPath + Name string + Body []interface{} +} + +// transport is a D-Bus transport. +type transport interface { + // Read and Write raw data (for example, for the authentication protocol). + io.ReadWriteCloser + + // Send the initial null byte used for the EXTERNAL mechanism. + SendNullByte() error + + // Returns whether this transport supports passing Unix FDs. + SupportsUnixFDs() bool + + // Signal the transport that Unix FD passing is enabled for this connection. + EnableUnixFDs() + + // Read / send a message, handling things like Unix FDs. + ReadMessage() (*Message, error) + SendMessage(*Message) error +} + +var ( + transports = make(map[string]func(string) (transport, error)) +) + +func getTransport(address string) (transport, error) { + var err error + var t transport + + addresses := strings.Split(address, ";") + for _, v := range addresses { + i := strings.IndexRune(v, ':') + if i == -1 { + err = errors.New("dbus: invalid bus address (no transport)") + continue + } + f := transports[v[:i]] + if f == nil { + err = errors.New("dbus: invalid bus address (invalid or unsupported transport)") + continue + } + t, err = f(v[i+1:]) + if err == nil { + return t, nil + } + } + return nil, err +} + +// dereferenceAll returns a slice that, assuming that vs is a slice of pointers +// of arbitrary types, containes the values that are obtained from dereferencing +// all elements in vs. +func dereferenceAll(vs []interface{}) []interface{} { + for i := range vs { + v := reflect.ValueOf(vs[i]) + v = v.Elem() + vs[i] = v.Interface() + } + return vs +} + +// getKey gets a key from a the list of keys. Returns "" on error / not found... +func getKey(s, key string) string { + i := strings.Index(s, key) + if i == -1 { + return "" + } + if i+len(key)+1 >= len(s) || s[i+len(key)] != '=' { + return "" + } + j := strings.Index(s, ",") + if j == -1 { + j = len(s) + } + return s[i+len(key)+1 : j] +} diff --git a/vendor/github.com/godbus/dbus/conn_darwin.go b/vendor/github.com/godbus/dbus/conn_darwin.go new file mode 100644 index 000000000..b67bb1b81 --- /dev/null +++ b/vendor/github.com/godbus/dbus/conn_darwin.go @@ -0,0 +1,21 @@ +package dbus + +import ( + "errors" + "os/exec" +) + +func sessionBusPlatform() (*Conn, error) { + cmd := exec.Command("launchctl", "getenv", "DBUS_LAUNCHD_SESSION_BUS_SOCKET") + b, err := cmd.CombinedOutput() + + if err != nil { + return nil, err + } + + if len(b) == 0 { + return nil, errors.New("dbus: couldn't determine address of session bus") + } + + return Dial("unix:path=" + string(b[:len(b)-1])) +} diff --git a/vendor/github.com/godbus/dbus/conn_other.go b/vendor/github.com/godbus/dbus/conn_other.go new file mode 100644 index 000000000..f74b8758d --- /dev/null +++ b/vendor/github.com/godbus/dbus/conn_other.go @@ -0,0 +1,27 @@ +// +build !darwin + +package dbus + +import ( + "bytes" + "errors" + "os/exec" +) + +func sessionBusPlatform() (*Conn, error) { + cmd := exec.Command("dbus-launch") + b, err := cmd.CombinedOutput() + + if err != nil { + return nil, err + } + + i := bytes.IndexByte(b, '=') + j := bytes.IndexByte(b, '\n') + + if i == -1 || j == -1 { + return nil, errors.New("dbus: couldn't determine address of session bus") + } + + return Dial(string(b[i+1 : j])) +} diff --git a/vendor/github.com/godbus/dbus/dbus.go b/vendor/github.com/godbus/dbus/dbus.go new file mode 100644 index 000000000..2ce68735c --- /dev/null +++ b/vendor/github.com/godbus/dbus/dbus.go @@ -0,0 +1,258 @@ +package dbus + +import ( + "errors" + "reflect" + "strings" +) + +var ( + byteType = reflect.TypeOf(byte(0)) + boolType = reflect.TypeOf(false) + uint8Type = reflect.TypeOf(uint8(0)) + int16Type = reflect.TypeOf(int16(0)) + uint16Type = reflect.TypeOf(uint16(0)) + int32Type = reflect.TypeOf(int32(0)) + uint32Type = reflect.TypeOf(uint32(0)) + int64Type = reflect.TypeOf(int64(0)) + uint64Type = reflect.TypeOf(uint64(0)) + float64Type = reflect.TypeOf(float64(0)) + stringType = reflect.TypeOf("") + signatureType = reflect.TypeOf(Signature{""}) + objectPathType = reflect.TypeOf(ObjectPath("")) + variantType = reflect.TypeOf(Variant{Signature{""}, nil}) + interfacesType = reflect.TypeOf([]interface{}{}) + unixFDType = reflect.TypeOf(UnixFD(0)) + unixFDIndexType = reflect.TypeOf(UnixFDIndex(0)) +) + +// An InvalidTypeError signals that a value which cannot be represented in the +// D-Bus wire format was passed to a function. +type InvalidTypeError struct { + Type reflect.Type +} + +func (e InvalidTypeError) Error() string { + return "dbus: invalid type " + e.Type.String() +} + +// Store copies the values contained in src to dest, which must be a slice of +// pointers. It converts slices of interfaces from src to corresponding structs +// in dest. An error is returned if the lengths of src and dest or the types of +// their elements don't match. +func Store(src []interface{}, dest ...interface{}) error { + if len(src) != len(dest) { + return errors.New("dbus.Store: length mismatch") + } + + for i := range src { + if err := store(src[i], dest[i]); err != nil { + return err + } + } + return nil +} + +func store(src, dest interface{}) error { + if reflect.TypeOf(dest).Elem() == reflect.TypeOf(src) { + reflect.ValueOf(dest).Elem().Set(reflect.ValueOf(src)) + return nil + } else if hasStruct(dest) { + rv := reflect.ValueOf(dest).Elem() + switch rv.Kind() { + case reflect.Struct: + vs, ok := src.([]interface{}) + if !ok { + return errors.New("dbus.Store: type mismatch") + } + t := rv.Type() + ndest := make([]interface{}, 0, rv.NumField()) + for i := 0; i < rv.NumField(); i++ { + field := t.Field(i) + if field.PkgPath == "" && field.Tag.Get("dbus") != "-" { + ndest = append(ndest, rv.Field(i).Addr().Interface()) + } + } + if len(vs) != len(ndest) { + return errors.New("dbus.Store: type mismatch") + } + err := Store(vs, ndest...) + if err != nil { + return errors.New("dbus.Store: type mismatch") + } + case reflect.Slice: + sv := reflect.ValueOf(src) + if sv.Kind() != reflect.Slice { + return errors.New("dbus.Store: type mismatch") + } + rv.Set(reflect.MakeSlice(rv.Type(), sv.Len(), sv.Len())) + for i := 0; i < sv.Len(); i++ { + if err := store(sv.Index(i).Interface(), rv.Index(i).Addr().Interface()); err != nil { + return err + } + } + case reflect.Map: + sv := reflect.ValueOf(src) + if sv.Kind() != reflect.Map { + return errors.New("dbus.Store: type mismatch") + } + keys := sv.MapKeys() + rv.Set(reflect.MakeMap(sv.Type())) + for _, key := range keys { + v := reflect.New(sv.Type().Elem()) + if err := store(v, sv.MapIndex(key).Interface()); err != nil { + return err + } + rv.SetMapIndex(key, v.Elem()) + } + default: + return errors.New("dbus.Store: type mismatch") + } + return nil + } else { + return errors.New("dbus.Store: type mismatch") + } +} + +func hasStruct(v interface{}) bool { + t := reflect.TypeOf(v) + for { + switch t.Kind() { + case reflect.Struct: + return true + case reflect.Slice, reflect.Ptr, reflect.Map: + t = t.Elem() + default: + return false + } + } +} + +// An ObjectPath is an object path as defined by the D-Bus spec. +type ObjectPath string + +// IsValid returns whether the object path is valid. +func (o ObjectPath) IsValid() bool { + s := string(o) + if len(s) == 0 { + return false + } + if s[0] != '/' { + return false + } + if s[len(s)-1] == '/' && len(s) != 1 { + return false + } + // probably not used, but technically possible + if s == "/" { + return true + } + split := strings.Split(s[1:], "/") + for _, v := range split { + if len(v) == 0 { + return false + } + for _, c := range v { + if !isMemberChar(c) { + return false + } + } + } + return true +} + +// A UnixFD is a Unix file descriptor sent over the wire. See the package-level +// documentation for more information about Unix file descriptor passsing. +type UnixFD int32 + +// A UnixFDIndex is the representation of a Unix file descriptor in a message. +type UnixFDIndex uint32 + +// alignment returns the alignment of values of type t. +func alignment(t reflect.Type) int { + switch t { + case variantType: + return 1 + case objectPathType: + return 4 + case signatureType: + return 1 + case interfacesType: // sometimes used for structs + return 8 + } + switch t.Kind() { + case reflect.Uint8: + return 1 + case reflect.Uint16, reflect.Int16: + return 2 + case reflect.Uint32, reflect.Int32, reflect.String, reflect.Array, reflect.Slice, reflect.Map: + return 4 + case reflect.Uint64, reflect.Int64, reflect.Float64, reflect.Struct: + return 8 + case reflect.Ptr: + return alignment(t.Elem()) + } + return 1 +} + +// isKeyType returns whether t is a valid type for a D-Bus dict. +func isKeyType(t reflect.Type) bool { + switch t.Kind() { + case reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, + reflect.Int16, reflect.Int32, reflect.Int64, reflect.Float64, + reflect.String: + + return true + } + return false +} + +// isValidInterface returns whether s is a valid name for an interface. +func isValidInterface(s string) bool { + if len(s) == 0 || len(s) > 255 || s[0] == '.' { + return false + } + elem := strings.Split(s, ".") + if len(elem) < 2 { + return false + } + for _, v := range elem { + if len(v) == 0 { + return false + } + if v[0] >= '0' && v[0] <= '9' { + return false + } + for _, c := range v { + if !isMemberChar(c) { + return false + } + } + } + return true +} + +// isValidMember returns whether s is a valid name for a member. +func isValidMember(s string) bool { + if len(s) == 0 || len(s) > 255 { + return false + } + i := strings.Index(s, ".") + if i != -1 { + return false + } + if s[0] >= '0' && s[0] <= '9' { + return false + } + for _, c := range s { + if !isMemberChar(c) { + return false + } + } + return true +} + +func isMemberChar(c rune) bool { + return (c >= '0' && c <= '9') || (c >= 'A' && c <= 'Z') || + (c >= 'a' && c <= 'z') || c == '_' +} diff --git a/vendor/github.com/godbus/dbus/decoder.go b/vendor/github.com/godbus/dbus/decoder.go new file mode 100644 index 000000000..ef50dcab9 --- /dev/null +++ b/vendor/github.com/godbus/dbus/decoder.go @@ -0,0 +1,228 @@ +package dbus + +import ( + "encoding/binary" + "io" + "reflect" +) + +type decoder struct { + in io.Reader + order binary.ByteOrder + pos int +} + +// newDecoder returns a new decoder that reads values from in. The input is +// expected to be in the given byte order. +func newDecoder(in io.Reader, order binary.ByteOrder) *decoder { + dec := new(decoder) + dec.in = in + dec.order = order + return dec +} + +// align aligns the input to the given boundary and panics on error. +func (dec *decoder) align(n int) { + if dec.pos%n != 0 { + newpos := (dec.pos + n - 1) & ^(n - 1) + empty := make([]byte, newpos-dec.pos) + if _, err := io.ReadFull(dec.in, empty); err != nil { + panic(err) + } + dec.pos = newpos + } +} + +// Calls binary.Read(dec.in, dec.order, v) and panics on read errors. +func (dec *decoder) binread(v interface{}) { + if err := binary.Read(dec.in, dec.order, v); err != nil { + panic(err) + } +} + +func (dec *decoder) Decode(sig Signature) (vs []interface{}, err error) { + defer func() { + var ok bool + v := recover() + if err, ok = v.(error); ok { + if err == io.EOF || err == io.ErrUnexpectedEOF { + err = FormatError("unexpected EOF") + } + } + }() + vs = make([]interface{}, 0) + s := sig.str + for s != "" { + err, rem := validSingle(s, 0) + if err != nil { + return nil, err + } + v := dec.decode(s[:len(s)-len(rem)], 0) + vs = append(vs, v) + s = rem + } + return vs, nil +} + +func (dec *decoder) decode(s string, depth int) interface{} { + dec.align(alignment(typeFor(s))) + switch s[0] { + case 'y': + var b [1]byte + if _, err := dec.in.Read(b[:]); err != nil { + panic(err) + } + dec.pos++ + return b[0] + case 'b': + i := dec.decode("u", depth).(uint32) + switch { + case i == 0: + return false + case i == 1: + return true + default: + panic(FormatError("invalid value for boolean")) + } + case 'n': + var i int16 + dec.binread(&i) + dec.pos += 2 + return i + case 'i': + var i int32 + dec.binread(&i) + dec.pos += 4 + return i + case 'x': + var i int64 + dec.binread(&i) + dec.pos += 8 + return i + case 'q': + var i uint16 + dec.binread(&i) + dec.pos += 2 + return i + case 'u': + var i uint32 + dec.binread(&i) + dec.pos += 4 + return i + case 't': + var i uint64 + dec.binread(&i) + dec.pos += 8 + return i + case 'd': + var f float64 + dec.binread(&f) + dec.pos += 8 + return f + case 's': + length := dec.decode("u", depth).(uint32) + b := make([]byte, int(length)+1) + if _, err := io.ReadFull(dec.in, b); err != nil { + panic(err) + } + dec.pos += int(length) + 1 + return string(b[:len(b)-1]) + case 'o': + return ObjectPath(dec.decode("s", depth).(string)) + case 'g': + length := dec.decode("y", depth).(byte) + b := make([]byte, int(length)+1) + if _, err := io.ReadFull(dec.in, b); err != nil { + panic(err) + } + dec.pos += int(length) + 1 + sig, err := ParseSignature(string(b[:len(b)-1])) + if err != nil { + panic(err) + } + return sig + case 'v': + if depth >= 64 { + panic(FormatError("input exceeds container depth limit")) + } + var variant Variant + sig := dec.decode("g", depth).(Signature) + if len(sig.str) == 0 { + panic(FormatError("variant signature is empty")) + } + err, rem := validSingle(sig.str, 0) + if err != nil { + panic(err) + } + if rem != "" { + panic(FormatError("variant signature has multiple types")) + } + variant.sig = sig + variant.value = dec.decode(sig.str, depth+1) + return variant + case 'h': + return UnixFDIndex(dec.decode("u", depth).(uint32)) + case 'a': + if len(s) > 1 && s[1] == '{' { + ksig := s[2:3] + vsig := s[3 : len(s)-1] + v := reflect.MakeMap(reflect.MapOf(typeFor(ksig), typeFor(vsig))) + if depth >= 63 { + panic(FormatError("input exceeds container depth limit")) + } + length := dec.decode("u", depth).(uint32) + // Even for empty maps, the correct padding must be included + dec.align(8) + spos := dec.pos + for dec.pos < spos+int(length) { + dec.align(8) + if !isKeyType(v.Type().Key()) { + panic(InvalidTypeError{v.Type()}) + } + kv := dec.decode(ksig, depth+2) + vv := dec.decode(vsig, depth+2) + v.SetMapIndex(reflect.ValueOf(kv), reflect.ValueOf(vv)) + } + return v.Interface() + } + if depth >= 64 { + panic(FormatError("input exceeds container depth limit")) + } + length := dec.decode("u", depth).(uint32) + v := reflect.MakeSlice(reflect.SliceOf(typeFor(s[1:])), 0, int(length)) + // Even for empty arrays, the correct padding must be included + dec.align(alignment(typeFor(s[1:]))) + spos := dec.pos + for dec.pos < spos+int(length) { + ev := dec.decode(s[1:], depth+1) + v = reflect.Append(v, reflect.ValueOf(ev)) + } + return v.Interface() + case '(': + if depth >= 64 { + panic(FormatError("input exceeds container depth limit")) + } + dec.align(8) + v := make([]interface{}, 0) + s = s[1 : len(s)-1] + for s != "" { + err, rem := validSingle(s, 0) + if err != nil { + panic(err) + } + ev := dec.decode(s[:len(s)-len(rem)], depth+1) + v = append(v, ev) + s = rem + } + return v + default: + panic(SignatureError{Sig: s}) + } +} + +// A FormatError is an error in the wire format. +type FormatError string + +func (e FormatError) Error() string { + return "dbus: wire format error: " + string(e) +} diff --git a/vendor/github.com/godbus/dbus/doc.go b/vendor/github.com/godbus/dbus/doc.go new file mode 100644 index 000000000..deff554a3 --- /dev/null +++ b/vendor/github.com/godbus/dbus/doc.go @@ -0,0 +1,63 @@ +/* +Package dbus implements bindings to the D-Bus message bus system. + +To use the message bus API, you first need to connect to a bus (usually the +session or system bus). The acquired connection then can be used to call methods +on remote objects and emit or receive signals. Using the Export method, you can +arrange D-Bus methods calls to be directly translated to method calls on a Go +value. + +Conversion Rules + +For outgoing messages, Go types are automatically converted to the +corresponding D-Bus types. The following types are directly encoded as their +respective D-Bus equivalents: + + Go type | D-Bus type + ------------+----------- + byte | BYTE + bool | BOOLEAN + int16 | INT16 + uint16 | UINT16 + int32 | INT32 + uint32 | UINT32 + int64 | INT64 + uint64 | UINT64 + float64 | DOUBLE + string | STRING + ObjectPath | OBJECT_PATH + Signature | SIGNATURE + Variant | VARIANT + UnixFDIndex | UNIX_FD + +Slices and arrays encode as ARRAYs of their element type. + +Maps encode as DICTs, provided that their key type can be used as a key for +a DICT. + +Structs other than Variant and Signature encode as a STRUCT containing their +exported fields. Fields whose tags contain `dbus:"-"` and unexported fields will +be skipped. + +Pointers encode as the value they're pointed to. + +Trying to encode any other type or a slice, map or struct containing an +unsupported type will result in an InvalidTypeError. + +For incoming messages, the inverse of these rules are used, with the exception +of STRUCTs. Incoming STRUCTS are represented as a slice of empty interfaces +containing the struct fields in the correct order. The Store function can be +used to convert such values to Go structs. + +Unix FD passing + +Handling Unix file descriptors deserves special mention. To use them, you should +first check that they are supported on a connection by calling SupportsUnixFDs. +If it returns true, all method of Connection will translate messages containing +UnixFD's to messages that are accompanied by the given file descriptors with the +UnixFD values being substituted by the correct indices. Similarily, the indices +of incoming messages are automatically resolved. It shouldn't be necessary to use +UnixFDIndex. + +*/ +package dbus diff --git a/vendor/github.com/godbus/dbus/encoder.go b/vendor/github.com/godbus/dbus/encoder.go new file mode 100644 index 000000000..9f0a9e89e --- /dev/null +++ b/vendor/github.com/godbus/dbus/encoder.go @@ -0,0 +1,208 @@ +package dbus + +import ( + "bytes" + "encoding/binary" + "io" + "reflect" +) + +// An encoder encodes values to the D-Bus wire format. +type encoder struct { + out io.Writer + order binary.ByteOrder + pos int +} + +// NewEncoder returns a new encoder that writes to out in the given byte order. +func newEncoder(out io.Writer, order binary.ByteOrder) *encoder { + return newEncoderAtOffset(out, 0, order) +} + +// newEncoderAtOffset returns a new encoder that writes to out in the given +// byte order. Specify the offset to initialize pos for proper alignment +// computation. +func newEncoderAtOffset(out io.Writer, offset int, order binary.ByteOrder) *encoder { + enc := new(encoder) + enc.out = out + enc.order = order + enc.pos = offset + return enc +} + +// Aligns the next output to be on a multiple of n. Panics on write errors. +func (enc *encoder) align(n int) { + pad := enc.padding(0, n) + if pad > 0 { + empty := make([]byte, pad) + if _, err := enc.out.Write(empty); err != nil { + panic(err) + } + enc.pos += pad + } +} + +// pad returns the number of bytes of padding, based on current position and additional offset. +// and alignment. +func (enc *encoder) padding(offset, algn int) int { + abs := enc.pos + offset + if abs%algn != 0 { + newabs := (abs + algn - 1) & ^(algn - 1) + return newabs - abs + } + return 0 +} + +// Calls binary.Write(enc.out, enc.order, v) and panics on write errors. +func (enc *encoder) binwrite(v interface{}) { + if err := binary.Write(enc.out, enc.order, v); err != nil { + panic(err) + } +} + +// Encode encodes the given values to the underyling reader. All written values +// are aligned properly as required by the D-Bus spec. +func (enc *encoder) Encode(vs ...interface{}) (err error) { + defer func() { + err, _ = recover().(error) + }() + for _, v := range vs { + enc.encode(reflect.ValueOf(v), 0) + } + return nil +} + +// encode encodes the given value to the writer and panics on error. depth holds +// the depth of the container nesting. +func (enc *encoder) encode(v reflect.Value, depth int) { + enc.align(alignment(v.Type())) + switch v.Kind() { + case reflect.Uint8: + var b [1]byte + b[0] = byte(v.Uint()) + if _, err := enc.out.Write(b[:]); err != nil { + panic(err) + } + enc.pos++ + case reflect.Bool: + if v.Bool() { + enc.encode(reflect.ValueOf(uint32(1)), depth) + } else { + enc.encode(reflect.ValueOf(uint32(0)), depth) + } + case reflect.Int16: + enc.binwrite(int16(v.Int())) + enc.pos += 2 + case reflect.Uint16: + enc.binwrite(uint16(v.Uint())) + enc.pos += 2 + case reflect.Int32: + enc.binwrite(int32(v.Int())) + enc.pos += 4 + case reflect.Uint32: + enc.binwrite(uint32(v.Uint())) + enc.pos += 4 + case reflect.Int64: + enc.binwrite(v.Int()) + enc.pos += 8 + case reflect.Uint64: + enc.binwrite(v.Uint()) + enc.pos += 8 + case reflect.Float64: + enc.binwrite(v.Float()) + enc.pos += 8 + case reflect.String: + enc.encode(reflect.ValueOf(uint32(len(v.String()))), depth) + b := make([]byte, v.Len()+1) + copy(b, v.String()) + b[len(b)-1] = 0 + n, err := enc.out.Write(b) + if err != nil { + panic(err) + } + enc.pos += n + case reflect.Ptr: + enc.encode(v.Elem(), depth) + case reflect.Slice, reflect.Array: + if depth >= 64 { + panic(FormatError("input exceeds container depth limit")) + } + // Lookahead offset: 4 bytes for uint32 length (with alignment), + // plus alignment for elements. + n := enc.padding(0, 4) + 4 + offset := enc.pos + n + enc.padding(n, alignment(v.Type().Elem())) + + var buf bytes.Buffer + bufenc := newEncoderAtOffset(&buf, offset, enc.order) + + for i := 0; i < v.Len(); i++ { + bufenc.encode(v.Index(i), depth+1) + } + enc.encode(reflect.ValueOf(uint32(buf.Len())), depth) + length := buf.Len() + enc.align(alignment(v.Type().Elem())) + if _, err := buf.WriteTo(enc.out); err != nil { + panic(err) + } + enc.pos += length + case reflect.Struct: + if depth >= 64 && v.Type() != signatureType { + panic(FormatError("input exceeds container depth limit")) + } + switch t := v.Type(); t { + case signatureType: + str := v.Field(0) + enc.encode(reflect.ValueOf(byte(str.Len())), depth+1) + b := make([]byte, str.Len()+1) + copy(b, str.String()) + b[len(b)-1] = 0 + n, err := enc.out.Write(b) + if err != nil { + panic(err) + } + enc.pos += n + case variantType: + variant := v.Interface().(Variant) + enc.encode(reflect.ValueOf(variant.sig), depth+1) + enc.encode(reflect.ValueOf(variant.value), depth+1) + default: + for i := 0; i < v.Type().NumField(); i++ { + field := t.Field(i) + if field.PkgPath == "" && field.Tag.Get("dbus") != "-" { + enc.encode(v.Field(i), depth+1) + } + } + } + case reflect.Map: + // Maps are arrays of structures, so they actually increase the depth by + // 2. + if depth >= 63 { + panic(FormatError("input exceeds container depth limit")) + } + if !isKeyType(v.Type().Key()) { + panic(InvalidTypeError{v.Type()}) + } + keys := v.MapKeys() + // Lookahead offset: 4 bytes for uint32 length (with alignment), + // plus 8-byte alignment + n := enc.padding(0, 4) + 4 + offset := enc.pos + n + enc.padding(n, 8) + + var buf bytes.Buffer + bufenc := newEncoderAtOffset(&buf, offset, enc.order) + for _, k := range keys { + bufenc.align(8) + bufenc.encode(k, depth+2) + bufenc.encode(v.MapIndex(k), depth+2) + } + enc.encode(reflect.ValueOf(uint32(buf.Len())), depth) + length := buf.Len() + enc.align(8) + if _, err := buf.WriteTo(enc.out); err != nil { + panic(err) + } + enc.pos += length + default: + panic(InvalidTypeError{v.Type()}) + } +} diff --git a/vendor/github.com/godbus/dbus/export.go b/vendor/github.com/godbus/dbus/export.go new file mode 100644 index 000000000..c6440a741 --- /dev/null +++ b/vendor/github.com/godbus/dbus/export.go @@ -0,0 +1,411 @@ +package dbus + +import ( + "errors" + "fmt" + "reflect" + "strings" +) + +var ( + errmsgInvalidArg = Error{ + "org.freedesktop.DBus.Error.InvalidArgs", + []interface{}{"Invalid type / number of args"}, + } + errmsgNoObject = Error{ + "org.freedesktop.DBus.Error.NoSuchObject", + []interface{}{"No such object"}, + } + errmsgUnknownMethod = Error{ + "org.freedesktop.DBus.Error.UnknownMethod", + []interface{}{"Unknown / invalid method"}, + } +) + +// exportWithMapping represents an exported struct along with a method name +// mapping to allow for exporting lower-case methods, etc. +type exportWithMapping struct { + export interface{} + + // Method name mapping; key -> struct method, value -> dbus method. + mapping map[string]string + + // Whether or not this export is for the entire subtree + includeSubtree bool +} + +// Sender is a type which can be used in exported methods to receive the message +// sender. +type Sender string + +func exportedMethod(export exportWithMapping, name string) reflect.Value { + if export.export == nil { + return reflect.Value{} + } + + // If a mapping was included in the export, check the map to see if we + // should be looking for a different method in the export. + if export.mapping != nil { + for key, value := range export.mapping { + if value == name { + name = key + break + } + + // Catch the case where a method is aliased but the client is calling + // the original, e.g. the "Foo" method was exported mapped to + // "foo," and dbus client called the original "Foo." + if key == name { + return reflect.Value{} + } + } + } + + value := reflect.ValueOf(export.export) + m := value.MethodByName(name) + + // Catch the case of attempting to call an unexported method + method, ok := value.Type().MethodByName(name) + + if !m.IsValid() || !ok || method.PkgPath != "" { + return reflect.Value{} + } + t := m.Type() + if t.NumOut() == 0 || + t.Out(t.NumOut()-1) != reflect.TypeOf(&errmsgInvalidArg) { + + return reflect.Value{} + } + return m +} + +// searchHandlers will look through all registered handlers looking for one +// to handle the given path. If a verbatim one isn't found, it will check for +// a subtree registration for the path as well. +func (conn *Conn) searchHandlers(path ObjectPath) (map[string]exportWithMapping, bool) { + conn.handlersLck.RLock() + defer conn.handlersLck.RUnlock() + + handlers, ok := conn.handlers[path] + if ok { + return handlers, ok + } + + // If handlers weren't found for this exact path, look for a matching subtree + // registration + handlers = make(map[string]exportWithMapping) + path = path[:strings.LastIndex(string(path), "/")] + for len(path) > 0 { + var subtreeHandlers map[string]exportWithMapping + subtreeHandlers, ok = conn.handlers[path] + if ok { + for iface, handler := range subtreeHandlers { + // Only include this handler if it registered for the subtree + if handler.includeSubtree { + handlers[iface] = handler + } + } + + break + } + + path = path[:strings.LastIndex(string(path), "/")] + } + + return handlers, ok +} + +// handleCall handles the given method call (i.e. looks if it's one of the +// pre-implemented ones and searches for a corresponding handler if not). +func (conn *Conn) handleCall(msg *Message) { + name := msg.Headers[FieldMember].value.(string) + path := msg.Headers[FieldPath].value.(ObjectPath) + ifaceName, hasIface := msg.Headers[FieldInterface].value.(string) + sender, hasSender := msg.Headers[FieldSender].value.(string) + serial := msg.serial + if ifaceName == "org.freedesktop.DBus.Peer" { + switch name { + case "Ping": + conn.sendReply(sender, serial) + case "GetMachineId": + conn.sendReply(sender, serial, conn.uuid) + default: + conn.sendError(errmsgUnknownMethod, sender, serial) + } + return + } + if len(name) == 0 { + conn.sendError(errmsgUnknownMethod, sender, serial) + } + + // Find the exported handler (if any) for this path + handlers, ok := conn.searchHandlers(path) + if !ok { + conn.sendError(errmsgNoObject, sender, serial) + return + } + + var m reflect.Value + if hasIface { + iface := handlers[ifaceName] + m = exportedMethod(iface, name) + } else { + for _, v := range handlers { + m = exportedMethod(v, name) + if m.IsValid() { + break + } + } + } + + if !m.IsValid() { + conn.sendError(errmsgUnknownMethod, sender, serial) + return + } + + t := m.Type() + vs := msg.Body + pointers := make([]interface{}, t.NumIn()) + decode := make([]interface{}, 0, len(vs)) + for i := 0; i < t.NumIn(); i++ { + tp := t.In(i) + val := reflect.New(tp) + pointers[i] = val.Interface() + if tp == reflect.TypeOf((*Sender)(nil)).Elem() { + val.Elem().SetString(sender) + } else if tp == reflect.TypeOf((*Message)(nil)).Elem() { + val.Elem().Set(reflect.ValueOf(*msg)) + } else { + decode = append(decode, pointers[i]) + } + } + + if len(decode) != len(vs) { + conn.sendError(errmsgInvalidArg, sender, serial) + return + } + + if err := Store(vs, decode...); err != nil { + conn.sendError(errmsgInvalidArg, sender, serial) + return + } + + // Extract parameters + params := make([]reflect.Value, len(pointers)) + for i := 0; i < len(pointers); i++ { + params[i] = reflect.ValueOf(pointers[i]).Elem() + } + + // Call method + ret := m.Call(params) + if em := ret[t.NumOut()-1].Interface().(*Error); em != nil { + conn.sendError(*em, sender, serial) + return + } + + if msg.Flags&FlagNoReplyExpected == 0 { + reply := new(Message) + reply.Type = TypeMethodReply + reply.serial = conn.getSerial() + reply.Headers = make(map[HeaderField]Variant) + if hasSender { + reply.Headers[FieldDestination] = msg.Headers[FieldSender] + } + reply.Headers[FieldReplySerial] = MakeVariant(msg.serial) + reply.Body = make([]interface{}, len(ret)-1) + for i := 0; i < len(ret)-1; i++ { + reply.Body[i] = ret[i].Interface() + } + if len(ret) != 1 { + reply.Headers[FieldSignature] = MakeVariant(SignatureOf(reply.Body...)) + } + conn.outLck.RLock() + if !conn.closed { + conn.out <- reply + } + conn.outLck.RUnlock() + } +} + +// Emit emits the given signal on the message bus. The name parameter must be +// formatted as "interface.member", e.g., "org.freedesktop.DBus.NameLost". +func (conn *Conn) Emit(path ObjectPath, name string, values ...interface{}) error { + if !path.IsValid() { + return errors.New("dbus: invalid object path") + } + i := strings.LastIndex(name, ".") + if i == -1 { + return errors.New("dbus: invalid method name") + } + iface := name[:i] + member := name[i+1:] + if !isValidMember(member) { + return errors.New("dbus: invalid method name") + } + if !isValidInterface(iface) { + return errors.New("dbus: invalid interface name") + } + msg := new(Message) + msg.Type = TypeSignal + msg.serial = conn.getSerial() + msg.Headers = make(map[HeaderField]Variant) + msg.Headers[FieldInterface] = MakeVariant(iface) + msg.Headers[FieldMember] = MakeVariant(member) + msg.Headers[FieldPath] = MakeVariant(path) + msg.Body = values + if len(values) > 0 { + msg.Headers[FieldSignature] = MakeVariant(SignatureOf(values...)) + } + conn.outLck.RLock() + defer conn.outLck.RUnlock() + if conn.closed { + return ErrClosed + } + conn.out <- msg + return nil +} + +// Export registers the given value to be exported as an object on the +// message bus. +// +// If a method call on the given path and interface is received, an exported +// method with the same name is called with v as the receiver if the +// parameters match and the last return value is of type *Error. If this +// *Error is not nil, it is sent back to the caller as an error. +// Otherwise, a method reply is sent with the other return values as its body. +// +// Any parameters with the special type Sender are set to the sender of the +// dbus message when the method is called. Parameters of this type do not +// contribute to the dbus signature of the method (i.e. the method is exposed +// as if the parameters of type Sender were not there). +// +// Similarly, any parameters with the type Message are set to the raw message +// received on the bus. Again, parameters of this type do not contribute to the +// dbus signature of the method. +// +// Every method call is executed in a new goroutine, so the method may be called +// in multiple goroutines at once. +// +// Method calls on the interface org.freedesktop.DBus.Peer will be automatically +// handled for every object. +// +// Passing nil as the first parameter will cause conn to cease handling calls on +// the given combination of path and interface. +// +// Export returns an error if path is not a valid path name. +func (conn *Conn) Export(v interface{}, path ObjectPath, iface string) error { + return conn.ExportWithMap(v, nil, path, iface) +} + +// ExportWithMap works exactly like Export but provides the ability to remap +// method names (e.g. export a lower-case method). +// +// The keys in the map are the real method names (exported on the struct), and +// the values are the method names to be exported on DBus. +func (conn *Conn) ExportWithMap(v interface{}, mapping map[string]string, path ObjectPath, iface string) error { + return conn.exportWithMap(v, mapping, path, iface, false) +} + +// ExportSubtree works exactly like Export but registers the given value for +// an entire subtree rather under the root path provided. +// +// In order to make this useful, one parameter in each of the value's exported +// methods should be a Message, in which case it will contain the raw message +// (allowing one to get access to the path that caused the method to be called). +// +// Note that more specific export paths take precedence over less specific. For +// example, a method call using the ObjectPath /foo/bar/baz will call a method +// exported on /foo/bar before a method exported on /foo. +func (conn *Conn) ExportSubtree(v interface{}, path ObjectPath, iface string) error { + return conn.ExportSubtreeWithMap(v, nil, path, iface) +} + +// ExportSubtreeWithMap works exactly like ExportSubtree but provides the +// ability to remap method names (e.g. export a lower-case method). +// +// The keys in the map are the real method names (exported on the struct), and +// the values are the method names to be exported on DBus. +func (conn *Conn) ExportSubtreeWithMap(v interface{}, mapping map[string]string, path ObjectPath, iface string) error { + return conn.exportWithMap(v, mapping, path, iface, true) +} + +// exportWithMap is the worker function for all exports/registrations. +func (conn *Conn) exportWithMap(v interface{}, mapping map[string]string, path ObjectPath, iface string, includeSubtree bool) error { + if !path.IsValid() { + return fmt.Errorf(`dbus: Invalid path name: "%s"`, path) + } + + conn.handlersLck.Lock() + defer conn.handlersLck.Unlock() + + // Remove a previous export if the interface is nil + if v == nil { + if _, ok := conn.handlers[path]; ok { + delete(conn.handlers[path], iface) + if len(conn.handlers[path]) == 0 { + delete(conn.handlers, path) + } + } + + return nil + } + + // If this is the first handler for this path, make a new map to hold all + // handlers for this path. + if _, ok := conn.handlers[path]; !ok { + conn.handlers[path] = make(map[string]exportWithMapping) + } + + // Finally, save this handler + conn.handlers[path][iface] = exportWithMapping{export: v, mapping: mapping, includeSubtree: includeSubtree} + + return nil +} + +// ReleaseName calls org.freedesktop.DBus.ReleaseName and awaits a response. +func (conn *Conn) ReleaseName(name string) (ReleaseNameReply, error) { + var r uint32 + err := conn.busObj.Call("org.freedesktop.DBus.ReleaseName", 0, name).Store(&r) + if err != nil { + return 0, err + } + return ReleaseNameReply(r), nil +} + +// RequestName calls org.freedesktop.DBus.RequestName and awaits a response. +func (conn *Conn) RequestName(name string, flags RequestNameFlags) (RequestNameReply, error) { + var r uint32 + err := conn.busObj.Call("org.freedesktop.DBus.RequestName", 0, name, flags).Store(&r) + if err != nil { + return 0, err + } + return RequestNameReply(r), nil +} + +// ReleaseNameReply is the reply to a ReleaseName call. +type ReleaseNameReply uint32 + +const ( + ReleaseNameReplyReleased ReleaseNameReply = 1 + iota + ReleaseNameReplyNonExistent + ReleaseNameReplyNotOwner +) + +// RequestNameFlags represents the possible flags for a RequestName call. +type RequestNameFlags uint32 + +const ( + NameFlagAllowReplacement RequestNameFlags = 1 << iota + NameFlagReplaceExisting + NameFlagDoNotQueue +) + +// RequestNameReply is the reply to a RequestName call. +type RequestNameReply uint32 + +const ( + RequestNameReplyPrimaryOwner RequestNameReply = 1 + iota + RequestNameReplyInQueue + RequestNameReplyExists + RequestNameReplyAlreadyOwner +) diff --git a/vendor/github.com/godbus/dbus/homedir.go b/vendor/github.com/godbus/dbus/homedir.go new file mode 100644 index 000000000..0b745f931 --- /dev/null +++ b/vendor/github.com/godbus/dbus/homedir.go @@ -0,0 +1,28 @@ +package dbus + +import ( + "os" + "sync" +) + +var ( + homeDir string + homeDirLock sync.Mutex +) + +func getHomeDir() string { + homeDirLock.Lock() + defer homeDirLock.Unlock() + + if homeDir != "" { + return homeDir + } + + homeDir = os.Getenv("HOME") + if homeDir != "" { + return homeDir + } + + homeDir = lookupHomeDir() + return homeDir +} diff --git a/vendor/github.com/godbus/dbus/homedir_dynamic.go b/vendor/github.com/godbus/dbus/homedir_dynamic.go new file mode 100644 index 000000000..2732081e7 --- /dev/null +++ b/vendor/github.com/godbus/dbus/homedir_dynamic.go @@ -0,0 +1,15 @@ +// +build !static_build + +package dbus + +import ( + "os/user" +) + +func lookupHomeDir() string { + u, err := user.Current() + if err != nil { + return "/" + } + return u.HomeDir +} diff --git a/vendor/github.com/godbus/dbus/homedir_static.go b/vendor/github.com/godbus/dbus/homedir_static.go new file mode 100644 index 000000000..b9d9cb552 --- /dev/null +++ b/vendor/github.com/godbus/dbus/homedir_static.go @@ -0,0 +1,45 @@ +// +build static_build + +package dbus + +import ( + "bufio" + "os" + "strconv" + "strings" +) + +func lookupHomeDir() string { + myUid := os.Getuid() + + f, err := os.Open("/etc/passwd") + if err != nil { + return "/" + } + defer f.Close() + + s := bufio.NewScanner(f) + + for s.Scan() { + if err := s.Err(); err != nil { + break + } + + line := strings.TrimSpace(s.Text()) + if line == "" { + continue + } + + parts := strings.Split(line, ":") + + if len(parts) >= 6 { + uid, err := strconv.Atoi(parts[2]) + if err == nil && uid == myUid { + return parts[5] + } + } + } + + // Default to / if we can't get a better value + return "/" +} diff --git a/vendor/github.com/godbus/dbus/message.go b/vendor/github.com/godbus/dbus/message.go new file mode 100644 index 000000000..075d6e38b --- /dev/null +++ b/vendor/github.com/godbus/dbus/message.go @@ -0,0 +1,346 @@ +package dbus + +import ( + "bytes" + "encoding/binary" + "errors" + "io" + "reflect" + "strconv" +) + +const protoVersion byte = 1 + +// Flags represents the possible flags of a D-Bus message. +type Flags byte + +const ( + // FlagNoReplyExpected signals that the message is not expected to generate + // a reply. If this flag is set on outgoing messages, any possible reply + // will be discarded. + FlagNoReplyExpected Flags = 1 << iota + // FlagNoAutoStart signals that the message bus should not automatically + // start an application when handling this message. + FlagNoAutoStart +) + +// Type represents the possible types of a D-Bus message. +type Type byte + +const ( + TypeMethodCall Type = 1 + iota + TypeMethodReply + TypeError + TypeSignal + typeMax +) + +func (t Type) String() string { + switch t { + case TypeMethodCall: + return "method call" + case TypeMethodReply: + return "reply" + case TypeError: + return "error" + case TypeSignal: + return "signal" + } + return "invalid" +} + +// HeaderField represents the possible byte codes for the headers +// of a D-Bus message. +type HeaderField byte + +const ( + FieldPath HeaderField = 1 + iota + FieldInterface + FieldMember + FieldErrorName + FieldReplySerial + FieldDestination + FieldSender + FieldSignature + FieldUnixFDs + fieldMax +) + +// An InvalidMessageError describes the reason why a D-Bus message is regarded as +// invalid. +type InvalidMessageError string + +func (e InvalidMessageError) Error() string { + return "dbus: invalid message: " + string(e) +} + +// fieldType are the types of the various header fields. +var fieldTypes = [fieldMax]reflect.Type{ + FieldPath: objectPathType, + FieldInterface: stringType, + FieldMember: stringType, + FieldErrorName: stringType, + FieldReplySerial: uint32Type, + FieldDestination: stringType, + FieldSender: stringType, + FieldSignature: signatureType, + FieldUnixFDs: uint32Type, +} + +// requiredFields lists the header fields that are required by the different +// message types. +var requiredFields = [typeMax][]HeaderField{ + TypeMethodCall: {FieldPath, FieldMember}, + TypeMethodReply: {FieldReplySerial}, + TypeError: {FieldErrorName, FieldReplySerial}, + TypeSignal: {FieldPath, FieldInterface, FieldMember}, +} + +// Message represents a single D-Bus message. +type Message struct { + Type + Flags + Headers map[HeaderField]Variant + Body []interface{} + + serial uint32 +} + +type header struct { + Field byte + Variant +} + +// DecodeMessage tries to decode a single message in the D-Bus wire format +// from the given reader. The byte order is figured out from the first byte. +// The possibly returned error can be an error of the underlying reader, an +// InvalidMessageError or a FormatError. +func DecodeMessage(rd io.Reader) (msg *Message, err error) { + var order binary.ByteOrder + var hlength, length uint32 + var typ, flags, proto byte + var headers []header + + b := make([]byte, 1) + _, err = rd.Read(b) + if err != nil { + return + } + switch b[0] { + case 'l': + order = binary.LittleEndian + case 'B': + order = binary.BigEndian + default: + return nil, InvalidMessageError("invalid byte order") + } + + dec := newDecoder(rd, order) + dec.pos = 1 + + msg = new(Message) + vs, err := dec.Decode(Signature{"yyyuu"}) + if err != nil { + return nil, err + } + if err = Store(vs, &typ, &flags, &proto, &length, &msg.serial); err != nil { + return nil, err + } + msg.Type = Type(typ) + msg.Flags = Flags(flags) + + // get the header length separately because we need it later + b = make([]byte, 4) + _, err = io.ReadFull(rd, b) + if err != nil { + return nil, err + } + binary.Read(bytes.NewBuffer(b), order, &hlength) + if hlength+length+16 > 1<<27 { + return nil, InvalidMessageError("message is too long") + } + dec = newDecoder(io.MultiReader(bytes.NewBuffer(b), rd), order) + dec.pos = 12 + vs, err = dec.Decode(Signature{"a(yv)"}) + if err != nil { + return nil, err + } + if err = Store(vs, &headers); err != nil { + return nil, err + } + + msg.Headers = make(map[HeaderField]Variant) + for _, v := range headers { + msg.Headers[HeaderField(v.Field)] = v.Variant + } + + dec.align(8) + body := make([]byte, int(length)) + if length != 0 { + _, err := io.ReadFull(rd, body) + if err != nil { + return nil, err + } + } + + if err = msg.IsValid(); err != nil { + return nil, err + } + sig, _ := msg.Headers[FieldSignature].value.(Signature) + if sig.str != "" { + buf := bytes.NewBuffer(body) + dec = newDecoder(buf, order) + vs, err := dec.Decode(sig) + if err != nil { + return nil, err + } + msg.Body = vs + } + + return +} + +// EncodeTo encodes and sends a message to the given writer. The byte order must +// be either binary.LittleEndian or binary.BigEndian. If the message is not +// valid or an error occurs when writing, an error is returned. +func (msg *Message) EncodeTo(out io.Writer, order binary.ByteOrder) error { + if err := msg.IsValid(); err != nil { + return err + } + var vs [7]interface{} + switch order { + case binary.LittleEndian: + vs[0] = byte('l') + case binary.BigEndian: + vs[0] = byte('B') + default: + return errors.New("dbus: invalid byte order") + } + body := new(bytes.Buffer) + enc := newEncoder(body, order) + if len(msg.Body) != 0 { + enc.Encode(msg.Body...) + } + vs[1] = msg.Type + vs[2] = msg.Flags + vs[3] = protoVersion + vs[4] = uint32(len(body.Bytes())) + vs[5] = msg.serial + headers := make([]header, 0, len(msg.Headers)) + for k, v := range msg.Headers { + headers = append(headers, header{byte(k), v}) + } + vs[6] = headers + var buf bytes.Buffer + enc = newEncoder(&buf, order) + enc.Encode(vs[:]...) + enc.align(8) + body.WriteTo(&buf) + if buf.Len() > 1<<27 { + return InvalidMessageError("message is too long") + } + if _, err := buf.WriteTo(out); err != nil { + return err + } + return nil +} + +// IsValid checks whether msg is a valid message and returns an +// InvalidMessageError if it is not. +func (msg *Message) IsValid() error { + if msg.Flags & ^(FlagNoAutoStart|FlagNoReplyExpected) != 0 { + return InvalidMessageError("invalid flags") + } + if msg.Type == 0 || msg.Type >= typeMax { + return InvalidMessageError("invalid message type") + } + for k, v := range msg.Headers { + if k == 0 || k >= fieldMax { + return InvalidMessageError("invalid header") + } + if reflect.TypeOf(v.value) != fieldTypes[k] { + return InvalidMessageError("invalid type of header field") + } + } + for _, v := range requiredFields[msg.Type] { + if _, ok := msg.Headers[v]; !ok { + return InvalidMessageError("missing required header") + } + } + if path, ok := msg.Headers[FieldPath]; ok { + if !path.value.(ObjectPath).IsValid() { + return InvalidMessageError("invalid path name") + } + } + if iface, ok := msg.Headers[FieldInterface]; ok { + if !isValidInterface(iface.value.(string)) { + return InvalidMessageError("invalid interface name") + } + } + if member, ok := msg.Headers[FieldMember]; ok { + if !isValidMember(member.value.(string)) { + return InvalidMessageError("invalid member name") + } + } + if errname, ok := msg.Headers[FieldErrorName]; ok { + if !isValidInterface(errname.value.(string)) { + return InvalidMessageError("invalid error name") + } + } + if len(msg.Body) != 0 { + if _, ok := msg.Headers[FieldSignature]; !ok { + return InvalidMessageError("missing signature") + } + } + return nil +} + +// Serial returns the message's serial number. The returned value is only valid +// for messages received by eavesdropping. +func (msg *Message) Serial() uint32 { + return msg.serial +} + +// String returns a string representation of a message similar to the format of +// dbus-monitor. +func (msg *Message) String() string { + if err := msg.IsValid(); err != nil { + return "" + } + s := msg.Type.String() + if v, ok := msg.Headers[FieldSender]; ok { + s += " from " + v.value.(string) + } + if v, ok := msg.Headers[FieldDestination]; ok { + s += " to " + v.value.(string) + } + s += " serial " + strconv.FormatUint(uint64(msg.serial), 10) + if v, ok := msg.Headers[FieldReplySerial]; ok { + s += " reply_serial " + strconv.FormatUint(uint64(v.value.(uint32)), 10) + } + if v, ok := msg.Headers[FieldUnixFDs]; ok { + s += " unixfds " + strconv.FormatUint(uint64(v.value.(uint32)), 10) + } + if v, ok := msg.Headers[FieldPath]; ok { + s += " path " + string(v.value.(ObjectPath)) + } + if v, ok := msg.Headers[FieldInterface]; ok { + s += " interface " + v.value.(string) + } + if v, ok := msg.Headers[FieldErrorName]; ok { + s += " error " + v.value.(string) + } + if v, ok := msg.Headers[FieldMember]; ok { + s += " member " + v.value.(string) + } + if len(msg.Body) != 0 { + s += "\n" + } + for i, v := range msg.Body { + s += " " + MakeVariant(v).String() + if i != len(msg.Body)-1 { + s += "\n" + } + } + return s +} diff --git a/vendor/github.com/godbus/dbus/object.go b/vendor/github.com/godbus/dbus/object.go new file mode 100644 index 000000000..7ef45da4c --- /dev/null +++ b/vendor/github.com/godbus/dbus/object.go @@ -0,0 +1,126 @@ +package dbus + +import ( + "errors" + "strings" +) + +// BusObject is the interface of a remote object on which methods can be +// invoked. +type BusObject interface { + Call(method string, flags Flags, args ...interface{}) *Call + Go(method string, flags Flags, ch chan *Call, args ...interface{}) *Call + GetProperty(p string) (Variant, error) + Destination() string + Path() ObjectPath +} + +// Object represents a remote object on which methods can be invoked. +type Object struct { + conn *Conn + dest string + path ObjectPath +} + +// Call calls a method with (*Object).Go and waits for its reply. +func (o *Object) Call(method string, flags Flags, args ...interface{}) *Call { + return <-o.Go(method, flags, make(chan *Call, 1), args...).Done +} + +// Go calls a method with the given arguments asynchronously. It returns a +// Call structure representing this method call. The passed channel will +// return the same value once the call is done. If ch is nil, a new channel +// will be allocated. Otherwise, ch has to be buffered or Go will panic. +// +// If the flags include FlagNoReplyExpected, ch is ignored and a Call structure +// is returned of which only the Err member is valid. +// +// If the method parameter contains a dot ('.'), the part before the last dot +// specifies the interface on which the method is called. +func (o *Object) Go(method string, flags Flags, ch chan *Call, args ...interface{}) *Call { + iface := "" + i := strings.LastIndex(method, ".") + if i != -1 { + iface = method[:i] + } + method = method[i+1:] + msg := new(Message) + msg.Type = TypeMethodCall + msg.serial = o.conn.getSerial() + msg.Flags = flags & (FlagNoAutoStart | FlagNoReplyExpected) + msg.Headers = make(map[HeaderField]Variant) + msg.Headers[FieldPath] = MakeVariant(o.path) + msg.Headers[FieldDestination] = MakeVariant(o.dest) + msg.Headers[FieldMember] = MakeVariant(method) + if iface != "" { + msg.Headers[FieldInterface] = MakeVariant(iface) + } + msg.Body = args + if len(args) > 0 { + msg.Headers[FieldSignature] = MakeVariant(SignatureOf(args...)) + } + if msg.Flags&FlagNoReplyExpected == 0 { + if ch == nil { + ch = make(chan *Call, 10) + } else if cap(ch) == 0 { + panic("dbus: unbuffered channel passed to (*Object).Go") + } + call := &Call{ + Destination: o.dest, + Path: o.path, + Method: method, + Args: args, + Done: ch, + } + o.conn.callsLck.Lock() + o.conn.calls[msg.serial] = call + o.conn.callsLck.Unlock() + o.conn.outLck.RLock() + if o.conn.closed { + call.Err = ErrClosed + call.Done <- call + } else { + o.conn.out <- msg + } + o.conn.outLck.RUnlock() + return call + } + o.conn.outLck.RLock() + defer o.conn.outLck.RUnlock() + if o.conn.closed { + return &Call{Err: ErrClosed} + } + o.conn.out <- msg + return &Call{Err: nil} +} + +// GetProperty calls org.freedesktop.DBus.Properties.GetProperty on the given +// object. The property name must be given in interface.member notation. +func (o *Object) GetProperty(p string) (Variant, error) { + idx := strings.LastIndex(p, ".") + if idx == -1 || idx+1 == len(p) { + return Variant{}, errors.New("dbus: invalid property " + p) + } + + iface := p[:idx] + prop := p[idx+1:] + + result := Variant{} + err := o.Call("org.freedesktop.DBus.Properties.Get", 0, iface, prop).Store(&result) + + if err != nil { + return Variant{}, err + } + + return result, nil +} + +// Destination returns the destination that calls on o are sent to. +func (o *Object) Destination() string { + return o.dest +} + +// Path returns the path that calls on o are sent to. +func (o *Object) Path() ObjectPath { + return o.path +} diff --git a/vendor/github.com/godbus/dbus/sig.go b/vendor/github.com/godbus/dbus/sig.go new file mode 100644 index 000000000..f45b53ce1 --- /dev/null +++ b/vendor/github.com/godbus/dbus/sig.go @@ -0,0 +1,257 @@ +package dbus + +import ( + "fmt" + "reflect" + "strings" +) + +var sigToType = map[byte]reflect.Type{ + 'y': byteType, + 'b': boolType, + 'n': int16Type, + 'q': uint16Type, + 'i': int32Type, + 'u': uint32Type, + 'x': int64Type, + 't': uint64Type, + 'd': float64Type, + 's': stringType, + 'g': signatureType, + 'o': objectPathType, + 'v': variantType, + 'h': unixFDIndexType, +} + +// Signature represents a correct type signature as specified by the D-Bus +// specification. The zero value represents the empty signature, "". +type Signature struct { + str string +} + +// SignatureOf returns the concatenation of all the signatures of the given +// values. It panics if one of them is not representable in D-Bus. +func SignatureOf(vs ...interface{}) Signature { + var s string + for _, v := range vs { + s += getSignature(reflect.TypeOf(v)) + } + return Signature{s} +} + +// SignatureOfType returns the signature of the given type. It panics if the +// type is not representable in D-Bus. +func SignatureOfType(t reflect.Type) Signature { + return Signature{getSignature(t)} +} + +// getSignature returns the signature of the given type and panics on unknown types. +func getSignature(t reflect.Type) string { + // handle simple types first + switch t.Kind() { + case reflect.Uint8: + return "y" + case reflect.Bool: + return "b" + case reflect.Int16: + return "n" + case reflect.Uint16: + return "q" + case reflect.Int32: + if t == unixFDType { + return "h" + } + return "i" + case reflect.Uint32: + if t == unixFDIndexType { + return "h" + } + return "u" + case reflect.Int64: + return "x" + case reflect.Uint64: + return "t" + case reflect.Float64: + return "d" + case reflect.Ptr: + return getSignature(t.Elem()) + case reflect.String: + if t == objectPathType { + return "o" + } + return "s" + case reflect.Struct: + if t == variantType { + return "v" + } else if t == signatureType { + return "g" + } + var s string + for i := 0; i < t.NumField(); i++ { + field := t.Field(i) + if field.PkgPath == "" && field.Tag.Get("dbus") != "-" { + s += getSignature(t.Field(i).Type) + } + } + return "(" + s + ")" + case reflect.Array, reflect.Slice: + return "a" + getSignature(t.Elem()) + case reflect.Map: + if !isKeyType(t.Key()) { + panic(InvalidTypeError{t}) + } + return "a{" + getSignature(t.Key()) + getSignature(t.Elem()) + "}" + } + panic(InvalidTypeError{t}) +} + +// ParseSignature returns the signature represented by this string, or a +// SignatureError if the string is not a valid signature. +func ParseSignature(s string) (sig Signature, err error) { + if len(s) == 0 { + return + } + if len(s) > 255 { + return Signature{""}, SignatureError{s, "too long"} + } + sig.str = s + for err == nil && len(s) != 0 { + err, s = validSingle(s, 0) + } + if err != nil { + sig = Signature{""} + } + + return +} + +// ParseSignatureMust behaves like ParseSignature, except that it panics if s +// is not valid. +func ParseSignatureMust(s string) Signature { + sig, err := ParseSignature(s) + if err != nil { + panic(err) + } + return sig +} + +// Empty retruns whether the signature is the empty signature. +func (s Signature) Empty() bool { + return s.str == "" +} + +// Single returns whether the signature represents a single, complete type. +func (s Signature) Single() bool { + err, r := validSingle(s.str, 0) + return err != nil && r == "" +} + +// String returns the signature's string representation. +func (s Signature) String() string { + return s.str +} + +// A SignatureError indicates that a signature passed to a function or received +// on a connection is not a valid signature. +type SignatureError struct { + Sig string + Reason string +} + +func (e SignatureError) Error() string { + return fmt.Sprintf("dbus: invalid signature: %q (%s)", e.Sig, e.Reason) +} + +// Try to read a single type from this string. If it was successfull, err is nil +// and rem is the remaining unparsed part. Otherwise, err is a non-nil +// SignatureError and rem is "". depth is the current recursion depth which may +// not be greater than 64 and should be given as 0 on the first call. +func validSingle(s string, depth int) (err error, rem string) { + if s == "" { + return SignatureError{Sig: s, Reason: "empty signature"}, "" + } + if depth > 64 { + return SignatureError{Sig: s, Reason: "container nesting too deep"}, "" + } + switch s[0] { + case 'y', 'b', 'n', 'q', 'i', 'u', 'x', 't', 'd', 's', 'g', 'o', 'v', 'h': + return nil, s[1:] + case 'a': + if len(s) > 1 && s[1] == '{' { + i := findMatching(s[1:], '{', '}') + if i == -1 { + return SignatureError{Sig: s, Reason: "unmatched '{'"}, "" + } + i++ + rem = s[i+1:] + s = s[2:i] + if err, _ = validSingle(s[:1], depth+1); err != nil { + return err, "" + } + err, nr := validSingle(s[1:], depth+1) + if err != nil { + return err, "" + } + if nr != "" { + return SignatureError{Sig: s, Reason: "too many types in dict"}, "" + } + return nil, rem + } + return validSingle(s[1:], depth+1) + case '(': + i := findMatching(s, '(', ')') + if i == -1 { + return SignatureError{Sig: s, Reason: "unmatched ')'"}, "" + } + rem = s[i+1:] + s = s[1:i] + for err == nil && s != "" { + err, s = validSingle(s, depth+1) + } + if err != nil { + rem = "" + } + return + } + return SignatureError{Sig: s, Reason: "invalid type character"}, "" +} + +func findMatching(s string, left, right rune) int { + n := 0 + for i, v := range s { + if v == left { + n++ + } else if v == right { + n-- + } + if n == 0 { + return i + } + } + return -1 +} + +// typeFor returns the type of the given signature. It ignores any left over +// characters and panics if s doesn't start with a valid type signature. +func typeFor(s string) (t reflect.Type) { + err, _ := validSingle(s, 0) + if err != nil { + panic(err) + } + + if t, ok := sigToType[s[0]]; ok { + return t + } + switch s[0] { + case 'a': + if s[1] == '{' { + i := strings.LastIndex(s, "}") + t = reflect.MapOf(sigToType[s[2]], typeFor(s[3:i])) + } else { + t = reflect.SliceOf(typeFor(s[1:])) + } + case '(': + t = interfacesType + } + return +} diff --git a/vendor/github.com/godbus/dbus/transport_darwin.go b/vendor/github.com/godbus/dbus/transport_darwin.go new file mode 100644 index 000000000..1bba0d6bf --- /dev/null +++ b/vendor/github.com/godbus/dbus/transport_darwin.go @@ -0,0 +1,6 @@ +package dbus + +func (t *unixTransport) SendNullByte() error { + _, err := t.Write([]byte{0}) + return err +} diff --git a/vendor/github.com/godbus/dbus/transport_generic.go b/vendor/github.com/godbus/dbus/transport_generic.go new file mode 100644 index 000000000..46f8f49d6 --- /dev/null +++ b/vendor/github.com/godbus/dbus/transport_generic.go @@ -0,0 +1,35 @@ +package dbus + +import ( + "encoding/binary" + "errors" + "io" +) + +type genericTransport struct { + io.ReadWriteCloser +} + +func (t genericTransport) SendNullByte() error { + _, err := t.Write([]byte{0}) + return err +} + +func (t genericTransport) SupportsUnixFDs() bool { + return false +} + +func (t genericTransport) EnableUnixFDs() {} + +func (t genericTransport) ReadMessage() (*Message, error) { + return DecodeMessage(t) +} + +func (t genericTransport) SendMessage(msg *Message) error { + for _, v := range msg.Body { + if _, ok := v.(UnixFD); ok { + return errors.New("dbus: unix fd passing not enabled") + } + } + return msg.EncodeTo(t, binary.LittleEndian) +} diff --git a/vendor/github.com/godbus/dbus/transport_unix.go b/vendor/github.com/godbus/dbus/transport_unix.go new file mode 100644 index 000000000..3fafeabb1 --- /dev/null +++ b/vendor/github.com/godbus/dbus/transport_unix.go @@ -0,0 +1,196 @@ +//+build !windows + +package dbus + +import ( + "bytes" + "encoding/binary" + "errors" + "io" + "net" + "syscall" +) + +type oobReader struct { + conn *net.UnixConn + oob []byte + buf [4096]byte +} + +func (o *oobReader) Read(b []byte) (n int, err error) { + n, oobn, flags, _, err := o.conn.ReadMsgUnix(b, o.buf[:]) + if err != nil { + return n, err + } + if flags&syscall.MSG_CTRUNC != 0 { + return n, errors.New("dbus: control data truncated (too many fds received)") + } + o.oob = append(o.oob, o.buf[:oobn]...) + return n, nil +} + +type unixTransport struct { + *net.UnixConn + hasUnixFDs bool +} + +func newUnixTransport(keys string) (transport, error) { + var err error + + t := new(unixTransport) + abstract := getKey(keys, "abstract") + path := getKey(keys, "path") + switch { + case abstract == "" && path == "": + return nil, errors.New("dbus: invalid address (neither path nor abstract set)") + case abstract != "" && path == "": + t.UnixConn, err = net.DialUnix("unix", nil, &net.UnixAddr{Name: "@" + abstract, Net: "unix"}) + if err != nil { + return nil, err + } + return t, nil + case abstract == "" && path != "": + t.UnixConn, err = net.DialUnix("unix", nil, &net.UnixAddr{Name: path, Net: "unix"}) + if err != nil { + return nil, err + } + return t, nil + default: + return nil, errors.New("dbus: invalid address (both path and abstract set)") + } +} + +func init() { + transports["unix"] = newUnixTransport +} + +func (t *unixTransport) EnableUnixFDs() { + t.hasUnixFDs = true +} + +func (t *unixTransport) ReadMessage() (*Message, error) { + var ( + blen, hlen uint32 + csheader [16]byte + headers []header + order binary.ByteOrder + unixfds uint32 + ) + // To be sure that all bytes of out-of-band data are read, we use a special + // reader that uses ReadUnix on the underlying connection instead of Read + // and gathers the out-of-band data in a buffer. + rd := &oobReader{conn: t.UnixConn} + // read the first 16 bytes (the part of the header that has a constant size), + // from which we can figure out the length of the rest of the message + if _, err := io.ReadFull(rd, csheader[:]); err != nil { + return nil, err + } + switch csheader[0] { + case 'l': + order = binary.LittleEndian + case 'B': + order = binary.BigEndian + default: + return nil, InvalidMessageError("invalid byte order") + } + // csheader[4:8] -> length of message body, csheader[12:16] -> length of + // header fields (without alignment) + binary.Read(bytes.NewBuffer(csheader[4:8]), order, &blen) + binary.Read(bytes.NewBuffer(csheader[12:]), order, &hlen) + if hlen%8 != 0 { + hlen += 8 - (hlen % 8) + } + + // decode headers and look for unix fds + headerdata := make([]byte, hlen+4) + copy(headerdata, csheader[12:]) + if _, err := io.ReadFull(t, headerdata[4:]); err != nil { + return nil, err + } + dec := newDecoder(bytes.NewBuffer(headerdata), order) + dec.pos = 12 + vs, err := dec.Decode(Signature{"a(yv)"}) + if err != nil { + return nil, err + } + Store(vs, &headers) + for _, v := range headers { + if v.Field == byte(FieldUnixFDs) { + unixfds, _ = v.Variant.value.(uint32) + } + } + all := make([]byte, 16+hlen+blen) + copy(all, csheader[:]) + copy(all[16:], headerdata[4:]) + if _, err := io.ReadFull(rd, all[16+hlen:]); err != nil { + return nil, err + } + if unixfds != 0 { + if !t.hasUnixFDs { + return nil, errors.New("dbus: got unix fds on unsupported transport") + } + // read the fds from the OOB data + scms, err := syscall.ParseSocketControlMessage(rd.oob) + if err != nil { + return nil, err + } + if len(scms) != 1 { + return nil, errors.New("dbus: received more than one socket control message") + } + fds, err := syscall.ParseUnixRights(&scms[0]) + if err != nil { + return nil, err + } + msg, err := DecodeMessage(bytes.NewBuffer(all)) + if err != nil { + return nil, err + } + // substitute the values in the message body (which are indices for the + // array receiver via OOB) with the actual values + for i, v := range msg.Body { + if j, ok := v.(UnixFDIndex); ok { + if uint32(j) >= unixfds { + return nil, InvalidMessageError("invalid index for unix fd") + } + msg.Body[i] = UnixFD(fds[j]) + } + } + return msg, nil + } + return DecodeMessage(bytes.NewBuffer(all)) +} + +func (t *unixTransport) SendMessage(msg *Message) error { + fds := make([]int, 0) + for i, v := range msg.Body { + if fd, ok := v.(UnixFD); ok { + msg.Body[i] = UnixFDIndex(len(fds)) + fds = append(fds, int(fd)) + } + } + if len(fds) != 0 { + if !t.hasUnixFDs { + return errors.New("dbus: unix fd passing not enabled") + } + msg.Headers[FieldUnixFDs] = MakeVariant(uint32(len(fds))) + oob := syscall.UnixRights(fds...) + buf := new(bytes.Buffer) + msg.EncodeTo(buf, binary.LittleEndian) + n, oobn, err := t.UnixConn.WriteMsgUnix(buf.Bytes(), oob, nil) + if err != nil { + return err + } + if n != buf.Len() || oobn != len(oob) { + return io.ErrShortWrite + } + } else { + if err := msg.EncodeTo(t, binary.LittleEndian); err != nil { + return nil + } + } + return nil +} + +func (t *unixTransport) SupportsUnixFDs() bool { + return true +} diff --git a/vendor/github.com/godbus/dbus/transport_unixcred_dragonfly.go b/vendor/github.com/godbus/dbus/transport_unixcred_dragonfly.go new file mode 100644 index 000000000..a8cd39395 --- /dev/null +++ b/vendor/github.com/godbus/dbus/transport_unixcred_dragonfly.go @@ -0,0 +1,95 @@ +// The UnixCredentials system call is currently only implemented on Linux +// http://golang.org/src/pkg/syscall/sockcmsg_linux.go +// https://golang.org/s/go1.4-syscall +// http://code.google.com/p/go/source/browse/unix/sockcmsg_linux.go?repo=sys + +// Local implementation of the UnixCredentials system call for DragonFly BSD + +package dbus + +/* +#include +*/ +import "C" + +import ( + "io" + "os" + "syscall" + "unsafe" +) + +// http://golang.org/src/pkg/syscall/ztypes_linux_amd64.go +// http://golang.org/src/pkg/syscall/ztypes_dragonfly_amd64.go +type Ucred struct { + Pid int32 + Uid uint32 + Gid uint32 +} + +// http://golang.org/src/pkg/syscall/types_linux.go +// http://golang.org/src/pkg/syscall/types_dragonfly.go +// https://github.com/DragonFlyBSD/DragonFlyBSD/blob/master/sys/sys/ucred.h +const ( + SizeofUcred = C.sizeof_struct_ucred +) + +// http://golang.org/src/pkg/syscall/sockcmsg_unix.go +func cmsgAlignOf(salen int) int { + // From http://golang.org/src/pkg/syscall/sockcmsg_unix.go + //salign := sizeofPtr + // NOTE: It seems like 64-bit Darwin and DragonFly BSD kernels + // still require 32-bit aligned access to network subsystem. + //if darwin64Bit || dragonfly64Bit { + // salign = 4 + //} + salign := 4 + return (salen + salign - 1) & ^(salign - 1) +} + +// http://golang.org/src/pkg/syscall/sockcmsg_unix.go +func cmsgData(h *syscall.Cmsghdr) unsafe.Pointer { + return unsafe.Pointer(uintptr(unsafe.Pointer(h)) + uintptr(cmsgAlignOf(syscall.SizeofCmsghdr))) +} + +// http://golang.org/src/pkg/syscall/sockcmsg_linux.go +// UnixCredentials encodes credentials into a socket control message +// for sending to another process. This can be used for +// authentication. +func UnixCredentials(ucred *Ucred) []byte { + b := make([]byte, syscall.CmsgSpace(SizeofUcred)) + h := (*syscall.Cmsghdr)(unsafe.Pointer(&b[0])) + h.Level = syscall.SOL_SOCKET + h.Type = syscall.SCM_CREDS + h.SetLen(syscall.CmsgLen(SizeofUcred)) + *((*Ucred)(cmsgData(h))) = *ucred + return b +} + +// http://golang.org/src/pkg/syscall/sockcmsg_linux.go +// ParseUnixCredentials decodes a socket control message that contains +// credentials in a Ucred structure. To receive such a message, the +// SO_PASSCRED option must be enabled on the socket. +func ParseUnixCredentials(m *syscall.SocketControlMessage) (*Ucred, error) { + if m.Header.Level != syscall.SOL_SOCKET { + return nil, syscall.EINVAL + } + if m.Header.Type != syscall.SCM_CREDS { + return nil, syscall.EINVAL + } + ucred := *(*Ucred)(unsafe.Pointer(&m.Data[0])) + return &ucred, nil +} + +func (t *unixTransport) SendNullByte() error { + ucred := &Ucred{Pid: int32(os.Getpid()), Uid: uint32(os.Getuid()), Gid: uint32(os.Getgid())} + b := UnixCredentials(ucred) + _, oobn, err := t.UnixConn.WriteMsgUnix([]byte{0}, b, nil) + if err != nil { + return err + } + if oobn != len(b) { + return io.ErrShortWrite + } + return nil +} diff --git a/vendor/github.com/godbus/dbus/transport_unixcred_linux.go b/vendor/github.com/godbus/dbus/transport_unixcred_linux.go new file mode 100644 index 000000000..d9dfdf698 --- /dev/null +++ b/vendor/github.com/godbus/dbus/transport_unixcred_linux.go @@ -0,0 +1,25 @@ +// The UnixCredentials system call is currently only implemented on Linux +// http://golang.org/src/pkg/syscall/sockcmsg_linux.go +// https://golang.org/s/go1.4-syscall +// http://code.google.com/p/go/source/browse/unix/sockcmsg_linux.go?repo=sys + +package dbus + +import ( + "io" + "os" + "syscall" +) + +func (t *unixTransport) SendNullByte() error { + ucred := &syscall.Ucred{Pid: int32(os.Getpid()), Uid: uint32(os.Getuid()), Gid: uint32(os.Getgid())} + b := syscall.UnixCredentials(ucred) + _, oobn, err := t.UnixConn.WriteMsgUnix([]byte{0}, b, nil) + if err != nil { + return err + } + if oobn != len(b) { + return io.ErrShortWrite + } + return nil +} diff --git a/vendor/github.com/godbus/dbus/variant.go b/vendor/github.com/godbus/dbus/variant.go new file mode 100644 index 000000000..b7b13ae90 --- /dev/null +++ b/vendor/github.com/godbus/dbus/variant.go @@ -0,0 +1,139 @@ +package dbus + +import ( + "bytes" + "fmt" + "reflect" + "sort" + "strconv" +) + +// Variant represents the D-Bus variant type. +type Variant struct { + sig Signature + value interface{} +} + +// MakeVariant converts the given value to a Variant. It panics if v cannot be +// represented as a D-Bus type. +func MakeVariant(v interface{}) Variant { + return Variant{SignatureOf(v), v} +} + +// ParseVariant parses the given string as a variant as described at +// https://developer.gnome.org/glib/unstable/gvariant-text.html. If sig is not +// empty, it is taken to be the expected signature for the variant. +func ParseVariant(s string, sig Signature) (Variant, error) { + tokens := varLex(s) + p := &varParser{tokens: tokens} + n, err := varMakeNode(p) + if err != nil { + return Variant{}, err + } + if sig.str == "" { + sig, err = varInfer(n) + if err != nil { + return Variant{}, err + } + } + v, err := n.Value(sig) + if err != nil { + return Variant{}, err + } + return MakeVariant(v), nil +} + +// format returns a formatted version of v and whether this string can be parsed +// unambigously. +func (v Variant) format() (string, bool) { + switch v.sig.str[0] { + case 'b', 'i': + return fmt.Sprint(v.value), true + case 'n', 'q', 'u', 'x', 't', 'd', 'h': + return fmt.Sprint(v.value), false + case 's': + return strconv.Quote(v.value.(string)), true + case 'o': + return strconv.Quote(string(v.value.(ObjectPath))), false + case 'g': + return strconv.Quote(v.value.(Signature).str), false + case 'v': + s, unamb := v.value.(Variant).format() + if !unamb { + return "<@" + v.value.(Variant).sig.str + " " + s + ">", true + } + return "<" + s + ">", true + case 'y': + return fmt.Sprintf("%#x", v.value.(byte)), false + } + rv := reflect.ValueOf(v.value) + switch rv.Kind() { + case reflect.Slice: + if rv.Len() == 0 { + return "[]", false + } + unamb := true + buf := bytes.NewBuffer([]byte("[")) + for i := 0; i < rv.Len(); i++ { + // TODO: slooow + s, b := MakeVariant(rv.Index(i).Interface()).format() + unamb = unamb && b + buf.WriteString(s) + if i != rv.Len()-1 { + buf.WriteString(", ") + } + } + buf.WriteByte(']') + return buf.String(), unamb + case reflect.Map: + if rv.Len() == 0 { + return "{}", false + } + unamb := true + var buf bytes.Buffer + kvs := make([]string, rv.Len()) + for i, k := range rv.MapKeys() { + s, b := MakeVariant(k.Interface()).format() + unamb = unamb && b + buf.Reset() + buf.WriteString(s) + buf.WriteString(": ") + s, b = MakeVariant(rv.MapIndex(k).Interface()).format() + unamb = unamb && b + buf.WriteString(s) + kvs[i] = buf.String() + } + buf.Reset() + buf.WriteByte('{') + sort.Strings(kvs) + for i, kv := range kvs { + if i > 0 { + buf.WriteString(", ") + } + buf.WriteString(kv) + } + buf.WriteByte('}') + return buf.String(), unamb + } + return `"INVALID"`, true +} + +// Signature returns the D-Bus signature of the underlying value of v. +func (v Variant) Signature() Signature { + return v.sig +} + +// String returns the string representation of the underlying value of v as +// described at https://developer.gnome.org/glib/unstable/gvariant-text.html. +func (v Variant) String() string { + s, unamb := v.format() + if !unamb { + return "@" + v.sig.str + " " + s + } + return s +} + +// Value returns the underlying value of v. +func (v Variant) Value() interface{} { + return v.value +} diff --git a/vendor/github.com/godbus/dbus/variant_lexer.go b/vendor/github.com/godbus/dbus/variant_lexer.go new file mode 100644 index 000000000..332007d6f --- /dev/null +++ b/vendor/github.com/godbus/dbus/variant_lexer.go @@ -0,0 +1,284 @@ +package dbus + +import ( + "fmt" + "strings" + "unicode" + "unicode/utf8" +) + +// Heavily inspired by the lexer from text/template. + +type varToken struct { + typ varTokenType + val string +} + +type varTokenType byte + +const ( + tokEOF varTokenType = iota + tokError + tokNumber + tokString + tokBool + tokArrayStart + tokArrayEnd + tokDictStart + tokDictEnd + tokVariantStart + tokVariantEnd + tokComma + tokColon + tokType + tokByteString +) + +type varLexer struct { + input string + start int + pos int + width int + tokens []varToken +} + +type lexState func(*varLexer) lexState + +func varLex(s string) []varToken { + l := &varLexer{input: s} + l.run() + return l.tokens +} + +func (l *varLexer) accept(valid string) bool { + if strings.IndexRune(valid, l.next()) >= 0 { + return true + } + l.backup() + return false +} + +func (l *varLexer) backup() { + l.pos -= l.width +} + +func (l *varLexer) emit(t varTokenType) { + l.tokens = append(l.tokens, varToken{t, l.input[l.start:l.pos]}) + l.start = l.pos +} + +func (l *varLexer) errorf(format string, v ...interface{}) lexState { + l.tokens = append(l.tokens, varToken{ + tokError, + fmt.Sprintf(format, v...), + }) + return nil +} + +func (l *varLexer) ignore() { + l.start = l.pos +} + +func (l *varLexer) next() rune { + var r rune + + if l.pos >= len(l.input) { + l.width = 0 + return -1 + } + r, l.width = utf8.DecodeRuneInString(l.input[l.pos:]) + l.pos += l.width + return r +} + +func (l *varLexer) run() { + for state := varLexNormal; state != nil; { + state = state(l) + } +} + +func (l *varLexer) peek() rune { + r := l.next() + l.backup() + return r +} + +func varLexNormal(l *varLexer) lexState { + for { + r := l.next() + switch { + case r == -1: + l.emit(tokEOF) + return nil + case r == '[': + l.emit(tokArrayStart) + case r == ']': + l.emit(tokArrayEnd) + case r == '{': + l.emit(tokDictStart) + case r == '}': + l.emit(tokDictEnd) + case r == '<': + l.emit(tokVariantStart) + case r == '>': + l.emit(tokVariantEnd) + case r == ':': + l.emit(tokColon) + case r == ',': + l.emit(tokComma) + case r == '\'' || r == '"': + l.backup() + return varLexString + case r == '@': + l.backup() + return varLexType + case unicode.IsSpace(r): + l.ignore() + case unicode.IsNumber(r) || r == '+' || r == '-': + l.backup() + return varLexNumber + case r == 'b': + pos := l.start + if n := l.peek(); n == '"' || n == '\'' { + return varLexByteString + } + // not a byte string; try to parse it as a type or bool below + l.pos = pos + 1 + l.width = 1 + fallthrough + default: + // either a bool or a type. Try bools first. + l.backup() + if l.pos+4 <= len(l.input) { + if l.input[l.pos:l.pos+4] == "true" { + l.pos += 4 + l.emit(tokBool) + continue + } + } + if l.pos+5 <= len(l.input) { + if l.input[l.pos:l.pos+5] == "false" { + l.pos += 5 + l.emit(tokBool) + continue + } + } + // must be a type. + return varLexType + } + } +} + +var varTypeMap = map[string]string{ + "boolean": "b", + "byte": "y", + "int16": "n", + "uint16": "q", + "int32": "i", + "uint32": "u", + "int64": "x", + "uint64": "t", + "double": "f", + "string": "s", + "objectpath": "o", + "signature": "g", +} + +func varLexByteString(l *varLexer) lexState { + q := l.next() +Loop: + for { + switch l.next() { + case '\\': + if r := l.next(); r != -1 { + break + } + fallthrough + case -1: + return l.errorf("unterminated bytestring") + case q: + break Loop + } + } + l.emit(tokByteString) + return varLexNormal +} + +func varLexNumber(l *varLexer) lexState { + l.accept("+-") + digits := "0123456789" + if l.accept("0") { + if l.accept("x") { + digits = "0123456789abcdefABCDEF" + } else { + digits = "01234567" + } + } + for strings.IndexRune(digits, l.next()) >= 0 { + } + l.backup() + if l.accept(".") { + for strings.IndexRune(digits, l.next()) >= 0 { + } + l.backup() + } + if l.accept("eE") { + l.accept("+-") + for strings.IndexRune("0123456789", l.next()) >= 0 { + } + l.backup() + } + if r := l.peek(); unicode.IsLetter(r) { + l.next() + return l.errorf("bad number syntax: %q", l.input[l.start:l.pos]) + } + l.emit(tokNumber) + return varLexNormal +} + +func varLexString(l *varLexer) lexState { + q := l.next() +Loop: + for { + switch l.next() { + case '\\': + if r := l.next(); r != -1 { + break + } + fallthrough + case -1: + return l.errorf("unterminated string") + case q: + break Loop + } + } + l.emit(tokString) + return varLexNormal +} + +func varLexType(l *varLexer) lexState { + at := l.accept("@") + for { + r := l.next() + if r == -1 { + break + } + if unicode.IsSpace(r) { + l.backup() + break + } + } + if at { + if _, err := ParseSignature(l.input[l.start+1 : l.pos]); err != nil { + return l.errorf("%s", err) + } + } else { + if _, ok := varTypeMap[l.input[l.start:l.pos]]; ok { + l.emit(tokType) + return varLexNormal + } + return l.errorf("unrecognized type %q", l.input[l.start:l.pos]) + } + l.emit(tokType) + return varLexNormal +} diff --git a/vendor/github.com/godbus/dbus/variant_parser.go b/vendor/github.com/godbus/dbus/variant_parser.go new file mode 100644 index 000000000..d20f5da6d --- /dev/null +++ b/vendor/github.com/godbus/dbus/variant_parser.go @@ -0,0 +1,817 @@ +package dbus + +import ( + "bytes" + "errors" + "fmt" + "io" + "reflect" + "strconv" + "strings" + "unicode/utf8" +) + +type varParser struct { + tokens []varToken + i int +} + +func (p *varParser) backup() { + p.i-- +} + +func (p *varParser) next() varToken { + if p.i < len(p.tokens) { + t := p.tokens[p.i] + p.i++ + return t + } + return varToken{typ: tokEOF} +} + +type varNode interface { + Infer() (Signature, error) + String() string + Sigs() sigSet + Value(Signature) (interface{}, error) +} + +func varMakeNode(p *varParser) (varNode, error) { + var sig Signature + + for { + t := p.next() + switch t.typ { + case tokEOF: + return nil, io.ErrUnexpectedEOF + case tokError: + return nil, errors.New(t.val) + case tokNumber: + return varMakeNumNode(t, sig) + case tokString: + return varMakeStringNode(t, sig) + case tokBool: + if sig.str != "" && sig.str != "b" { + return nil, varTypeError{t.val, sig} + } + b, err := strconv.ParseBool(t.val) + if err != nil { + return nil, err + } + return boolNode(b), nil + case tokArrayStart: + return varMakeArrayNode(p, sig) + case tokVariantStart: + return varMakeVariantNode(p, sig) + case tokDictStart: + return varMakeDictNode(p, sig) + case tokType: + if sig.str != "" { + return nil, errors.New("unexpected type annotation") + } + if t.val[0] == '@' { + sig.str = t.val[1:] + } else { + sig.str = varTypeMap[t.val] + } + case tokByteString: + if sig.str != "" && sig.str != "ay" { + return nil, varTypeError{t.val, sig} + } + b, err := varParseByteString(t.val) + if err != nil { + return nil, err + } + return byteStringNode(b), nil + default: + return nil, fmt.Errorf("unexpected %q", t.val) + } + } +} + +type varTypeError struct { + val string + sig Signature +} + +func (e varTypeError) Error() string { + return fmt.Sprintf("dbus: can't parse %q as type %q", e.val, e.sig.str) +} + +type sigSet map[Signature]bool + +func (s sigSet) Empty() bool { + return len(s) == 0 +} + +func (s sigSet) Intersect(s2 sigSet) sigSet { + r := make(sigSet) + for k := range s { + if s2[k] { + r[k] = true + } + } + return r +} + +func (s sigSet) Single() (Signature, bool) { + if len(s) == 1 { + for k := range s { + return k, true + } + } + return Signature{}, false +} + +func (s sigSet) ToArray() sigSet { + r := make(sigSet, len(s)) + for k := range s { + r[Signature{"a" + k.str}] = true + } + return r +} + +type numNode struct { + sig Signature + str string + val interface{} +} + +var numSigSet = sigSet{ + Signature{"y"}: true, + Signature{"n"}: true, + Signature{"q"}: true, + Signature{"i"}: true, + Signature{"u"}: true, + Signature{"x"}: true, + Signature{"t"}: true, + Signature{"d"}: true, +} + +func (n numNode) Infer() (Signature, error) { + if strings.ContainsAny(n.str, ".e") { + return Signature{"d"}, nil + } + return Signature{"i"}, nil +} + +func (n numNode) String() string { + return n.str +} + +func (n numNode) Sigs() sigSet { + if n.sig.str != "" { + return sigSet{n.sig: true} + } + if strings.ContainsAny(n.str, ".e") { + return sigSet{Signature{"d"}: true} + } + return numSigSet +} + +func (n numNode) Value(sig Signature) (interface{}, error) { + if n.sig.str != "" && n.sig != sig { + return nil, varTypeError{n.str, sig} + } + if n.val != nil { + return n.val, nil + } + return varNumAs(n.str, sig) +} + +func varMakeNumNode(tok varToken, sig Signature) (varNode, error) { + if sig.str == "" { + return numNode{str: tok.val}, nil + } + num, err := varNumAs(tok.val, sig) + if err != nil { + return nil, err + } + return numNode{sig: sig, val: num}, nil +} + +func varNumAs(s string, sig Signature) (interface{}, error) { + isUnsigned := false + size := 32 + switch sig.str { + case "n": + size = 16 + case "i": + case "x": + size = 64 + case "y": + size = 8 + isUnsigned = true + case "q": + size = 16 + isUnsigned = true + case "u": + isUnsigned = true + case "t": + size = 64 + isUnsigned = true + case "d": + d, err := strconv.ParseFloat(s, 64) + if err != nil { + return nil, err + } + return d, nil + default: + return nil, varTypeError{s, sig} + } + base := 10 + if strings.HasPrefix(s, "0x") { + base = 16 + s = s[2:] + } + if strings.HasPrefix(s, "0") && len(s) != 1 { + base = 8 + s = s[1:] + } + if isUnsigned { + i, err := strconv.ParseUint(s, base, size) + if err != nil { + return nil, err + } + var v interface{} = i + switch sig.str { + case "y": + v = byte(i) + case "q": + v = uint16(i) + case "u": + v = uint32(i) + } + return v, nil + } + i, err := strconv.ParseInt(s, base, size) + if err != nil { + return nil, err + } + var v interface{} = i + switch sig.str { + case "n": + v = int16(i) + case "i": + v = int32(i) + } + return v, nil +} + +type stringNode struct { + sig Signature + str string // parsed + val interface{} // has correct type +} + +var stringSigSet = sigSet{ + Signature{"s"}: true, + Signature{"g"}: true, + Signature{"o"}: true, +} + +func (n stringNode) Infer() (Signature, error) { + return Signature{"s"}, nil +} + +func (n stringNode) String() string { + return n.str +} + +func (n stringNode) Sigs() sigSet { + if n.sig.str != "" { + return sigSet{n.sig: true} + } + return stringSigSet +} + +func (n stringNode) Value(sig Signature) (interface{}, error) { + if n.sig.str != "" && n.sig != sig { + return nil, varTypeError{n.str, sig} + } + if n.val != nil { + return n.val, nil + } + switch { + case sig.str == "g": + return Signature{n.str}, nil + case sig.str == "o": + return ObjectPath(n.str), nil + case sig.str == "s": + return n.str, nil + default: + return nil, varTypeError{n.str, sig} + } +} + +func varMakeStringNode(tok varToken, sig Signature) (varNode, error) { + if sig.str != "" && sig.str != "s" && sig.str != "g" && sig.str != "o" { + return nil, fmt.Errorf("invalid type %q for string", sig.str) + } + s, err := varParseString(tok.val) + if err != nil { + return nil, err + } + n := stringNode{str: s} + if sig.str == "" { + return stringNode{str: s}, nil + } + n.sig = sig + switch sig.str { + case "o": + n.val = ObjectPath(s) + case "g": + n.val = Signature{s} + case "s": + n.val = s + } + return n, nil +} + +func varParseString(s string) (string, error) { + // quotes are guaranteed to be there + s = s[1 : len(s)-1] + buf := new(bytes.Buffer) + for len(s) != 0 { + r, size := utf8.DecodeRuneInString(s) + if r == utf8.RuneError && size == 1 { + return "", errors.New("invalid UTF-8") + } + s = s[size:] + if r != '\\' { + buf.WriteRune(r) + continue + } + r, size = utf8.DecodeRuneInString(s) + if r == utf8.RuneError && size == 1 { + return "", errors.New("invalid UTF-8") + } + s = s[size:] + switch r { + case 'a': + buf.WriteRune(0x7) + case 'b': + buf.WriteRune(0x8) + case 'f': + buf.WriteRune(0xc) + case 'n': + buf.WriteRune('\n') + case 'r': + buf.WriteRune('\r') + case 't': + buf.WriteRune('\t') + case '\n': + case 'u': + if len(s) < 4 { + return "", errors.New("short unicode escape") + } + r, err := strconv.ParseUint(s[:4], 16, 32) + if err != nil { + return "", err + } + buf.WriteRune(rune(r)) + s = s[4:] + case 'U': + if len(s) < 8 { + return "", errors.New("short unicode escape") + } + r, err := strconv.ParseUint(s[:8], 16, 32) + if err != nil { + return "", err + } + buf.WriteRune(rune(r)) + s = s[8:] + default: + buf.WriteRune(r) + } + } + return buf.String(), nil +} + +var boolSigSet = sigSet{Signature{"b"}: true} + +type boolNode bool + +func (boolNode) Infer() (Signature, error) { + return Signature{"b"}, nil +} + +func (b boolNode) String() string { + if b { + return "true" + } + return "false" +} + +func (boolNode) Sigs() sigSet { + return boolSigSet +} + +func (b boolNode) Value(sig Signature) (interface{}, error) { + if sig.str != "b" { + return nil, varTypeError{b.String(), sig} + } + return bool(b), nil +} + +type arrayNode struct { + set sigSet + children []varNode + val interface{} +} + +func (n arrayNode) Infer() (Signature, error) { + for _, v := range n.children { + csig, err := varInfer(v) + if err != nil { + continue + } + return Signature{"a" + csig.str}, nil + } + return Signature{}, fmt.Errorf("can't infer type for %q", n.String()) +} + +func (n arrayNode) String() string { + s := "[" + for i, v := range n.children { + s += v.String() + if i != len(n.children)-1 { + s += ", " + } + } + return s + "]" +} + +func (n arrayNode) Sigs() sigSet { + return n.set +} + +func (n arrayNode) Value(sig Signature) (interface{}, error) { + if n.set.Empty() { + // no type information whatsoever, so this must be an empty slice + return reflect.MakeSlice(typeFor(sig.str), 0, 0).Interface(), nil + } + if !n.set[sig] { + return nil, varTypeError{n.String(), sig} + } + s := reflect.MakeSlice(typeFor(sig.str), len(n.children), len(n.children)) + for i, v := range n.children { + rv, err := v.Value(Signature{sig.str[1:]}) + if err != nil { + return nil, err + } + s.Index(i).Set(reflect.ValueOf(rv)) + } + return s.Interface(), nil +} + +func varMakeArrayNode(p *varParser, sig Signature) (varNode, error) { + var n arrayNode + if sig.str != "" { + n.set = sigSet{sig: true} + } + if t := p.next(); t.typ == tokArrayEnd { + return n, nil + } else { + p.backup() + } +Loop: + for { + t := p.next() + switch t.typ { + case tokEOF: + return nil, io.ErrUnexpectedEOF + case tokError: + return nil, errors.New(t.val) + } + p.backup() + cn, err := varMakeNode(p) + if err != nil { + return nil, err + } + if cset := cn.Sigs(); !cset.Empty() { + if n.set.Empty() { + n.set = cset.ToArray() + } else { + nset := cset.ToArray().Intersect(n.set) + if nset.Empty() { + return nil, fmt.Errorf("can't parse %q with given type information", cn.String()) + } + n.set = nset + } + } + n.children = append(n.children, cn) + switch t := p.next(); t.typ { + case tokEOF: + return nil, io.ErrUnexpectedEOF + case tokError: + return nil, errors.New(t.val) + case tokArrayEnd: + break Loop + case tokComma: + continue + default: + return nil, fmt.Errorf("unexpected %q", t.val) + } + } + return n, nil +} + +type variantNode struct { + n varNode +} + +var variantSet = sigSet{ + Signature{"v"}: true, +} + +func (variantNode) Infer() (Signature, error) { + return Signature{"v"}, nil +} + +func (n variantNode) String() string { + return "<" + n.n.String() + ">" +} + +func (variantNode) Sigs() sigSet { + return variantSet +} + +func (n variantNode) Value(sig Signature) (interface{}, error) { + if sig.str != "v" { + return nil, varTypeError{n.String(), sig} + } + sig, err := varInfer(n.n) + if err != nil { + return nil, err + } + v, err := n.n.Value(sig) + if err != nil { + return nil, err + } + return MakeVariant(v), nil +} + +func varMakeVariantNode(p *varParser, sig Signature) (varNode, error) { + n, err := varMakeNode(p) + if err != nil { + return nil, err + } + if t := p.next(); t.typ != tokVariantEnd { + return nil, fmt.Errorf("unexpected %q", t.val) + } + vn := variantNode{n} + if sig.str != "" && sig.str != "v" { + return nil, varTypeError{vn.String(), sig} + } + return variantNode{n}, nil +} + +type dictEntry struct { + key, val varNode +} + +type dictNode struct { + kset, vset sigSet + children []dictEntry + val interface{} +} + +func (n dictNode) Infer() (Signature, error) { + for _, v := range n.children { + ksig, err := varInfer(v.key) + if err != nil { + continue + } + vsig, err := varInfer(v.val) + if err != nil { + continue + } + return Signature{"a{" + ksig.str + vsig.str + "}"}, nil + } + return Signature{}, fmt.Errorf("can't infer type for %q", n.String()) +} + +func (n dictNode) String() string { + s := "{" + for i, v := range n.children { + s += v.key.String() + ": " + v.val.String() + if i != len(n.children)-1 { + s += ", " + } + } + return s + "}" +} + +func (n dictNode) Sigs() sigSet { + r := sigSet{} + for k := range n.kset { + for v := range n.vset { + sig := "a{" + k.str + v.str + "}" + r[Signature{sig}] = true + } + } + return r +} + +func (n dictNode) Value(sig Signature) (interface{}, error) { + set := n.Sigs() + if set.Empty() { + // no type information -> empty dict + return reflect.MakeMap(typeFor(sig.str)).Interface(), nil + } + if !set[sig] { + return nil, varTypeError{n.String(), sig} + } + m := reflect.MakeMap(typeFor(sig.str)) + ksig := Signature{sig.str[2:3]} + vsig := Signature{sig.str[3 : len(sig.str)-1]} + for _, v := range n.children { + kv, err := v.key.Value(ksig) + if err != nil { + return nil, err + } + vv, err := v.val.Value(vsig) + if err != nil { + return nil, err + } + m.SetMapIndex(reflect.ValueOf(kv), reflect.ValueOf(vv)) + } + return m.Interface(), nil +} + +func varMakeDictNode(p *varParser, sig Signature) (varNode, error) { + var n dictNode + + if sig.str != "" { + if len(sig.str) < 5 { + return nil, fmt.Errorf("invalid signature %q for dict type", sig) + } + ksig := Signature{string(sig.str[2])} + vsig := Signature{sig.str[3 : len(sig.str)-1]} + n.kset = sigSet{ksig: true} + n.vset = sigSet{vsig: true} + } + if t := p.next(); t.typ == tokDictEnd { + return n, nil + } else { + p.backup() + } +Loop: + for { + t := p.next() + switch t.typ { + case tokEOF: + return nil, io.ErrUnexpectedEOF + case tokError: + return nil, errors.New(t.val) + } + p.backup() + kn, err := varMakeNode(p) + if err != nil { + return nil, err + } + if kset := kn.Sigs(); !kset.Empty() { + if n.kset.Empty() { + n.kset = kset + } else { + n.kset = kset.Intersect(n.kset) + if n.kset.Empty() { + return nil, fmt.Errorf("can't parse %q with given type information", kn.String()) + } + } + } + t = p.next() + switch t.typ { + case tokEOF: + return nil, io.ErrUnexpectedEOF + case tokError: + return nil, errors.New(t.val) + case tokColon: + default: + return nil, fmt.Errorf("unexpected %q", t.val) + } + t = p.next() + switch t.typ { + case tokEOF: + return nil, io.ErrUnexpectedEOF + case tokError: + return nil, errors.New(t.val) + } + p.backup() + vn, err := varMakeNode(p) + if err != nil { + return nil, err + } + if vset := vn.Sigs(); !vset.Empty() { + if n.vset.Empty() { + n.vset = vset + } else { + n.vset = n.vset.Intersect(vset) + if n.vset.Empty() { + return nil, fmt.Errorf("can't parse %q with given type information", vn.String()) + } + } + } + n.children = append(n.children, dictEntry{kn, vn}) + t = p.next() + switch t.typ { + case tokEOF: + return nil, io.ErrUnexpectedEOF + case tokError: + return nil, errors.New(t.val) + case tokDictEnd: + break Loop + case tokComma: + continue + default: + return nil, fmt.Errorf("unexpected %q", t.val) + } + } + return n, nil +} + +type byteStringNode []byte + +var byteStringSet = sigSet{ + Signature{"ay"}: true, +} + +func (byteStringNode) Infer() (Signature, error) { + return Signature{"ay"}, nil +} + +func (b byteStringNode) String() string { + return string(b) +} + +func (b byteStringNode) Sigs() sigSet { + return byteStringSet +} + +func (b byteStringNode) Value(sig Signature) (interface{}, error) { + if sig.str != "ay" { + return nil, varTypeError{b.String(), sig} + } + return []byte(b), nil +} + +func varParseByteString(s string) ([]byte, error) { + // quotes and b at start are guaranteed to be there + b := make([]byte, 0, 1) + s = s[2 : len(s)-1] + for len(s) != 0 { + c := s[0] + s = s[1:] + if c != '\\' { + b = append(b, c) + continue + } + c = s[0] + s = s[1:] + switch c { + case 'a': + b = append(b, 0x7) + case 'b': + b = append(b, 0x8) + case 'f': + b = append(b, 0xc) + case 'n': + b = append(b, '\n') + case 'r': + b = append(b, '\r') + case 't': + b = append(b, '\t') + case 'x': + if len(s) < 2 { + return nil, errors.New("short escape") + } + n, err := strconv.ParseUint(s[:2], 16, 8) + if err != nil { + return nil, err + } + b = append(b, byte(n)) + s = s[2:] + case '0': + if len(s) < 3 { + return nil, errors.New("short escape") + } + n, err := strconv.ParseUint(s[:3], 8, 8) + if err != nil { + return nil, err + } + b = append(b, byte(n)) + s = s[3:] + default: + b = append(b, c) + } + } + return append(b, 0), nil +} + +func varInfer(n varNode) (Signature, error) { + if sig, ok := n.Sigs().Single(); ok { + return sig, nil + } + return n.Infer() +} diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/mflag/LICENSE b/vendor/github.com/gorilla/context/LICENSE similarity index 82% rename from vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/mflag/LICENSE rename to vendor/github.com/gorilla/context/LICENSE index ac74d8f04..0e5fb8728 100644 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/mflag/LICENSE +++ b/vendor/github.com/gorilla/context/LICENSE @@ -1,16 +1,16 @@ -Copyright (c) 2014-2015 The Docker & Go Authors. All rights reserved. +Copyright (c) 2012 Rodrigo Moraes. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: - * Redistributions of source code must retain the above copyright + * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. - * Redistributions in binary form must reproduce the above + * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. - * Neither the name of Google Inc. nor the names of its + * Neither the name of Google Inc. nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. diff --git a/vendor/github.com/gorilla/context/README.md b/vendor/github.com/gorilla/context/README.md new file mode 100644 index 000000000..08f86693b --- /dev/null +++ b/vendor/github.com/gorilla/context/README.md @@ -0,0 +1,10 @@ +context +======= +[![Build Status](https://travis-ci.org/gorilla/context.png?branch=master)](https://travis-ci.org/gorilla/context) + +gorilla/context is a general purpose registry for global request variables. + +> Note: gorilla/context, having been born well before `context.Context` existed, does not play well +> with the shallow copying of the request that [`http.Request.WithContext`](https://golang.org/pkg/net/http/#Request.WithContext) (added to net/http Go 1.7 onwards) performs. You should either use *just* gorilla/context, or moving forward, the new `http.Request.Context()`. + +Read the full documentation here: http://www.gorillatoolkit.org/pkg/context diff --git a/vendor/github.com/gorilla/context/context.go b/vendor/github.com/gorilla/context/context.go new file mode 100644 index 000000000..81cb128b1 --- /dev/null +++ b/vendor/github.com/gorilla/context/context.go @@ -0,0 +1,143 @@ +// Copyright 2012 The Gorilla Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package context + +import ( + "net/http" + "sync" + "time" +) + +var ( + mutex sync.RWMutex + data = make(map[*http.Request]map[interface{}]interface{}) + datat = make(map[*http.Request]int64) +) + +// Set stores a value for a given key in a given request. +func Set(r *http.Request, key, val interface{}) { + mutex.Lock() + if data[r] == nil { + data[r] = make(map[interface{}]interface{}) + datat[r] = time.Now().Unix() + } + data[r][key] = val + mutex.Unlock() +} + +// Get returns a value stored for a given key in a given request. +func Get(r *http.Request, key interface{}) interface{} { + mutex.RLock() + if ctx := data[r]; ctx != nil { + value := ctx[key] + mutex.RUnlock() + return value + } + mutex.RUnlock() + return nil +} + +// GetOk returns stored value and presence state like multi-value return of map access. +func GetOk(r *http.Request, key interface{}) (interface{}, bool) { + mutex.RLock() + if _, ok := data[r]; ok { + value, ok := data[r][key] + mutex.RUnlock() + return value, ok + } + mutex.RUnlock() + return nil, false +} + +// GetAll returns all stored values for the request as a map. Nil is returned for invalid requests. +func GetAll(r *http.Request) map[interface{}]interface{} { + mutex.RLock() + if context, ok := data[r]; ok { + result := make(map[interface{}]interface{}, len(context)) + for k, v := range context { + result[k] = v + } + mutex.RUnlock() + return result + } + mutex.RUnlock() + return nil +} + +// GetAllOk returns all stored values for the request as a map and a boolean value that indicates if +// the request was registered. +func GetAllOk(r *http.Request) (map[interface{}]interface{}, bool) { + mutex.RLock() + context, ok := data[r] + result := make(map[interface{}]interface{}, len(context)) + for k, v := range context { + result[k] = v + } + mutex.RUnlock() + return result, ok +} + +// Delete removes a value stored for a given key in a given request. +func Delete(r *http.Request, key interface{}) { + mutex.Lock() + if data[r] != nil { + delete(data[r], key) + } + mutex.Unlock() +} + +// Clear removes all values stored for a given request. +// +// This is usually called by a handler wrapper to clean up request +// variables at the end of a request lifetime. See ClearHandler(). +func Clear(r *http.Request) { + mutex.Lock() + clear(r) + mutex.Unlock() +} + +// clear is Clear without the lock. +func clear(r *http.Request) { + delete(data, r) + delete(datat, r) +} + +// Purge removes request data stored for longer than maxAge, in seconds. +// It returns the amount of requests removed. +// +// If maxAge <= 0, all request data is removed. +// +// This is only used for sanity check: in case context cleaning was not +// properly set some request data can be kept forever, consuming an increasing +// amount of memory. In case this is detected, Purge() must be called +// periodically until the problem is fixed. +func Purge(maxAge int) int { + mutex.Lock() + count := 0 + if maxAge <= 0 { + count = len(data) + data = make(map[*http.Request]map[interface{}]interface{}) + datat = make(map[*http.Request]int64) + } else { + min := time.Now().Unix() - int64(maxAge) + for r := range data { + if datat[r] < min { + clear(r) + count++ + } + } + } + mutex.Unlock() + return count +} + +// ClearHandler wraps an http.Handler and clears request values at the end +// of a request lifetime. +func ClearHandler(h http.Handler) http.Handler { + return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + defer Clear(r) + h.ServeHTTP(w, r) + }) +} diff --git a/vendor/github.com/gorilla/context/doc.go b/vendor/github.com/gorilla/context/doc.go new file mode 100644 index 000000000..448d1bfca --- /dev/null +++ b/vendor/github.com/gorilla/context/doc.go @@ -0,0 +1,88 @@ +// Copyright 2012 The Gorilla Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +/* +Package context stores values shared during a request lifetime. + +Note: gorilla/context, having been born well before `context.Context` existed, +does not play well > with the shallow copying of the request that +[`http.Request.WithContext`](https://golang.org/pkg/net/http/#Request.WithContext) +(added to net/http Go 1.7 onwards) performs. You should either use *just* +gorilla/context, or moving forward, the new `http.Request.Context()`. + +For example, a router can set variables extracted from the URL and later +application handlers can access those values, or it can be used to store +sessions values to be saved at the end of a request. There are several +others common uses. + +The idea was posted by Brad Fitzpatrick to the go-nuts mailing list: + + http://groups.google.com/group/golang-nuts/msg/e2d679d303aa5d53 + +Here's the basic usage: first define the keys that you will need. The key +type is interface{} so a key can be of any type that supports equality. +Here we define a key using a custom int type to avoid name collisions: + + package foo + + import ( + "github.com/gorilla/context" + ) + + type key int + + const MyKey key = 0 + +Then set a variable. Variables are bound to an http.Request object, so you +need a request instance to set a value: + + context.Set(r, MyKey, "bar") + +The application can later access the variable using the same key you provided: + + func MyHandler(w http.ResponseWriter, r *http.Request) { + // val is "bar". + val := context.Get(r, foo.MyKey) + + // returns ("bar", true) + val, ok := context.GetOk(r, foo.MyKey) + // ... + } + +And that's all about the basic usage. We discuss some other ideas below. + +Any type can be stored in the context. To enforce a given type, make the key +private and wrap Get() and Set() to accept and return values of a specific +type: + + type key int + + const mykey key = 0 + + // GetMyKey returns a value for this package from the request values. + func GetMyKey(r *http.Request) SomeType { + if rv := context.Get(r, mykey); rv != nil { + return rv.(SomeType) + } + return nil + } + + // SetMyKey sets a value for this package in the request values. + func SetMyKey(r *http.Request, val SomeType) { + context.Set(r, mykey, val) + } + +Variables must be cleared at the end of a request, to remove all values +that were stored. This can be done in an http.Handler, after a request was +served. Just call Clear() passing the request: + + context.Clear(r) + +...or use ClearHandler(), which conveniently wraps an http.Handler to clear +variables at the end of a request lifetime. + +The Routers from the packages gorilla/mux and gorilla/pat call Clear() +so if you are using either of them you don't need to clear the context manually. +*/ +package context diff --git a/vendor/github.com/gorilla/mux/AUTHORS b/vendor/github.com/gorilla/mux/AUTHORS new file mode 100644 index 000000000..b722392ee --- /dev/null +++ b/vendor/github.com/gorilla/mux/AUTHORS @@ -0,0 +1,8 @@ +# This is the official list of gorilla/mux authors for copyright purposes. +# +# Please keep the list sorted. + +Google LLC (https://opensource.google.com/) +Kamil Kisielk +Matt Silverlock +Rodrigo Moraes (https://github.com/moraes) diff --git a/vendor/github.com/gorilla/mux/ISSUE_TEMPLATE.md b/vendor/github.com/gorilla/mux/ISSUE_TEMPLATE.md new file mode 100644 index 000000000..196deda23 --- /dev/null +++ b/vendor/github.com/gorilla/mux/ISSUE_TEMPLATE.md @@ -0,0 +1,10 @@ +**What version of Go are you running?** (Paste the output of `go version`) + + +**What version of gorilla/mux are you at?** (Paste the output of `git rev-parse HEAD` inside `$GOPATH/src/github.com/gorilla/mux`) + + +**Describe your problem** (and what you have tried so far) + + +**Paste a minimal, runnable, reproduction of your issue below** (use backticks to format it) diff --git a/vendor/github.com/docker/docker/pkg/symlink/LICENSE.BSD b/vendor/github.com/gorilla/mux/LICENSE similarity index 82% rename from vendor/github.com/docker/docker/pkg/symlink/LICENSE.BSD rename to vendor/github.com/gorilla/mux/LICENSE index 9b4f4a294..6903df638 100644 --- a/vendor/github.com/docker/docker/pkg/symlink/LICENSE.BSD +++ b/vendor/github.com/gorilla/mux/LICENSE @@ -1,16 +1,16 @@ -Copyright (c) 2014-2016 The Docker & Go Authors. All rights reserved. +Copyright (c) 2012-2018 The Gorilla Authors. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: - * Redistributions of source code must retain the above copyright + * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. - * Redistributions in binary form must reproduce the above + * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. - * Neither the name of Google Inc. nor the names of its + * Neither the name of Google Inc. nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. diff --git a/vendor/github.com/gorilla/mux/README.md b/vendor/github.com/gorilla/mux/README.md new file mode 100644 index 000000000..bb1d0fd0e --- /dev/null +++ b/vendor/github.com/gorilla/mux/README.md @@ -0,0 +1,649 @@ +# gorilla/mux + +[![GoDoc](https://godoc.org/github.com/gorilla/mux?status.svg)](https://godoc.org/github.com/gorilla/mux) +[![Build Status](https://travis-ci.org/gorilla/mux.svg?branch=master)](https://travis-ci.org/gorilla/mux) +[![Sourcegraph](https://sourcegraph.com/github.com/gorilla/mux/-/badge.svg)](https://sourcegraph.com/github.com/gorilla/mux?badge) + +![Gorilla Logo](http://www.gorillatoolkit.org/static/images/gorilla-icon-64.png) + +http://www.gorillatoolkit.org/pkg/mux + +Package `gorilla/mux` implements a request router and dispatcher for matching incoming requests to +their respective handler. + +The name mux stands for "HTTP request multiplexer". Like the standard `http.ServeMux`, `mux.Router` matches incoming requests against a list of registered routes and calls a handler for the route that matches the URL or other conditions. The main features are: + +* It implements the `http.Handler` interface so it is compatible with the standard `http.ServeMux`. +* Requests can be matched based on URL host, path, path prefix, schemes, header and query values, HTTP methods or using custom matchers. +* URL hosts, paths and query values can have variables with an optional regular expression. +* Registered URLs can be built, or "reversed", which helps maintaining references to resources. +* Routes can be used as subrouters: nested routes are only tested if the parent route matches. This is useful to define groups of routes that share common conditions like a host, a path prefix or other repeated attributes. As a bonus, this optimizes request matching. + +--- + +* [Install](#install) +* [Examples](#examples) +* [Matching Routes](#matching-routes) +* [Static Files](#static-files) +* [Registered URLs](#registered-urls) +* [Walking Routes](#walking-routes) +* [Graceful Shutdown](#graceful-shutdown) +* [Middleware](#middleware) +* [Testing Handlers](#testing-handlers) +* [Full Example](#full-example) + +--- + +## Install + +With a [correctly configured](https://golang.org/doc/install#testing) Go toolchain: + +```sh +go get -u github.com/gorilla/mux +``` + +## Examples + +Let's start registering a couple of URL paths and handlers: + +```go +func main() { + r := mux.NewRouter() + r.HandleFunc("/", HomeHandler) + r.HandleFunc("/products", ProductsHandler) + r.HandleFunc("/articles", ArticlesHandler) + http.Handle("/", r) +} +``` + +Here we register three routes mapping URL paths to handlers. This is equivalent to how `http.HandleFunc()` works: if an incoming request URL matches one of the paths, the corresponding handler is called passing (`http.ResponseWriter`, `*http.Request`) as parameters. + +Paths can have variables. They are defined using the format `{name}` or `{name:pattern}`. If a regular expression pattern is not defined, the matched variable will be anything until the next slash. For example: + +```go +r := mux.NewRouter() +r.HandleFunc("/products/{key}", ProductHandler) +r.HandleFunc("/articles/{category}/", ArticlesCategoryHandler) +r.HandleFunc("/articles/{category}/{id:[0-9]+}", ArticleHandler) +``` + +The names are used to create a map of route variables which can be retrieved calling `mux.Vars()`: + +```go +func ArticlesCategoryHandler(w http.ResponseWriter, r *http.Request) { + vars := mux.Vars(r) + w.WriteHeader(http.StatusOK) + fmt.Fprintf(w, "Category: %v\n", vars["category"]) +} +``` + +And this is all you need to know about the basic usage. More advanced options are explained below. + +### Matching Routes + +Routes can also be restricted to a domain or subdomain. Just define a host pattern to be matched. They can also have variables: + +```go +r := mux.NewRouter() +// Only matches if domain is "www.example.com". +r.Host("www.example.com") +// Matches a dynamic subdomain. +r.Host("{subdomain:[a-z]+}.domain.com") +``` + +There are several other matchers that can be added. To match path prefixes: + +```go +r.PathPrefix("/products/") +``` + +...or HTTP methods: + +```go +r.Methods("GET", "POST") +``` + +...or URL schemes: + +```go +r.Schemes("https") +``` + +...or header values: + +```go +r.Headers("X-Requested-With", "XMLHttpRequest") +``` + +...or query values: + +```go +r.Queries("key", "value") +``` + +...or to use a custom matcher function: + +```go +r.MatcherFunc(func(r *http.Request, rm *RouteMatch) bool { + return r.ProtoMajor == 0 +}) +``` + +...and finally, it is possible to combine several matchers in a single route: + +```go +r.HandleFunc("/products", ProductsHandler). + Host("www.example.com"). + Methods("GET"). + Schemes("http") +``` + +Routes are tested in the order they were added to the router. If two routes match, the first one wins: + +```go +r := mux.NewRouter() +r.HandleFunc("/specific", specificHandler) +r.PathPrefix("/").Handler(catchAllHandler) +``` + +Setting the same matching conditions again and again can be boring, so we have a way to group several routes that share the same requirements. We call it "subrouting". + +For example, let's say we have several URLs that should only match when the host is `www.example.com`. Create a route for that host and get a "subrouter" from it: + +```go +r := mux.NewRouter() +s := r.Host("www.example.com").Subrouter() +``` + +Then register routes in the subrouter: + +```go +s.HandleFunc("/products/", ProductsHandler) +s.HandleFunc("/products/{key}", ProductHandler) +s.HandleFunc("/articles/{category}/{id:[0-9]+}", ArticleHandler) +``` + +The three URL paths we registered above will only be tested if the domain is `www.example.com`, because the subrouter is tested first. This is not only convenient, but also optimizes request matching. You can create subrouters combining any attribute matchers accepted by a route. + +Subrouters can be used to create domain or path "namespaces": you define subrouters in a central place and then parts of the app can register its paths relatively to a given subrouter. + +There's one more thing about subroutes. When a subrouter has a path prefix, the inner routes use it as base for their paths: + +```go +r := mux.NewRouter() +s := r.PathPrefix("/products").Subrouter() +// "/products/" +s.HandleFunc("/", ProductsHandler) +// "/products/{key}/" +s.HandleFunc("/{key}/", ProductHandler) +// "/products/{key}/details" +s.HandleFunc("/{key}/details", ProductDetailsHandler) +``` + + +### Static Files + +Note that the path provided to `PathPrefix()` represents a "wildcard": calling +`PathPrefix("/static/").Handler(...)` means that the handler will be passed any +request that matches "/static/\*". This makes it easy to serve static files with mux: + +```go +func main() { + var dir string + + flag.StringVar(&dir, "dir", ".", "the directory to serve files from. Defaults to the current dir") + flag.Parse() + r := mux.NewRouter() + + // This will serve files under http://localhost:8000/static/ + r.PathPrefix("/static/").Handler(http.StripPrefix("/static/", http.FileServer(http.Dir(dir)))) + + srv := &http.Server{ + Handler: r, + Addr: "127.0.0.1:8000", + // Good practice: enforce timeouts for servers you create! + WriteTimeout: 15 * time.Second, + ReadTimeout: 15 * time.Second, + } + + log.Fatal(srv.ListenAndServe()) +} +``` + +### Registered URLs + +Now let's see how to build registered URLs. + +Routes can be named. All routes that define a name can have their URLs built, or "reversed". We define a name calling `Name()` on a route. For example: + +```go +r := mux.NewRouter() +r.HandleFunc("/articles/{category}/{id:[0-9]+}", ArticleHandler). + Name("article") +``` + +To build a URL, get the route and call the `URL()` method, passing a sequence of key/value pairs for the route variables. For the previous route, we would do: + +```go +url, err := r.Get("article").URL("category", "technology", "id", "42") +``` + +...and the result will be a `url.URL` with the following path: + +``` +"/articles/technology/42" +``` + +This also works for host and query value variables: + +```go +r := mux.NewRouter() +r.Host("{subdomain}.domain.com"). + Path("/articles/{category}/{id:[0-9]+}"). + Queries("filter", "{filter}"). + HandlerFunc(ArticleHandler). + Name("article") + +// url.String() will be "http://news.domain.com/articles/technology/42?filter=gorilla" +url, err := r.Get("article").URL("subdomain", "news", + "category", "technology", + "id", "42", + "filter", "gorilla") +``` + +All variables defined in the route are required, and their values must conform to the corresponding patterns. These requirements guarantee that a generated URL will always match a registered route -- the only exception is for explicitly defined "build-only" routes which never match. + +Regex support also exists for matching Headers within a route. For example, we could do: + +```go +r.HeadersRegexp("Content-Type", "application/(text|json)") +``` + +...and the route will match both requests with a Content-Type of `application/json` as well as `application/text` + +There's also a way to build only the URL host or path for a route: use the methods `URLHost()` or `URLPath()` instead. For the previous route, we would do: + +```go +// "http://news.domain.com/" +host, err := r.Get("article").URLHost("subdomain", "news") + +// "/articles/technology/42" +path, err := r.Get("article").URLPath("category", "technology", "id", "42") +``` + +And if you use subrouters, host and path defined separately can be built as well: + +```go +r := mux.NewRouter() +s := r.Host("{subdomain}.domain.com").Subrouter() +s.Path("/articles/{category}/{id:[0-9]+}"). + HandlerFunc(ArticleHandler). + Name("article") + +// "http://news.domain.com/articles/technology/42" +url, err := r.Get("article").URL("subdomain", "news", + "category", "technology", + "id", "42") +``` + +### Walking Routes + +The `Walk` function on `mux.Router` can be used to visit all of the routes that are registered on a router. For example, +the following prints all of the registered routes: + +```go +package main + +import ( + "fmt" + "net/http" + "strings" + + "github.com/gorilla/mux" +) + +func handler(w http.ResponseWriter, r *http.Request) { + return +} + +func main() { + r := mux.NewRouter() + r.HandleFunc("/", handler) + r.HandleFunc("/products", handler).Methods("POST") + r.HandleFunc("/articles", handler).Methods("GET") + r.HandleFunc("/articles/{id}", handler).Methods("GET", "PUT") + r.HandleFunc("/authors", handler).Queries("surname", "{surname}") + err := r.Walk(func(route *mux.Route, router *mux.Router, ancestors []*mux.Route) error { + pathTemplate, err := route.GetPathTemplate() + if err == nil { + fmt.Println("ROUTE:", pathTemplate) + } + pathRegexp, err := route.GetPathRegexp() + if err == nil { + fmt.Println("Path regexp:", pathRegexp) + } + queriesTemplates, err := route.GetQueriesTemplates() + if err == nil { + fmt.Println("Queries templates:", strings.Join(queriesTemplates, ",")) + } + queriesRegexps, err := route.GetQueriesRegexp() + if err == nil { + fmt.Println("Queries regexps:", strings.Join(queriesRegexps, ",")) + } + methods, err := route.GetMethods() + if err == nil { + fmt.Println("Methods:", strings.Join(methods, ",")) + } + fmt.Println() + return nil + }) + + if err != nil { + fmt.Println(err) + } + + http.Handle("/", r) +} +``` + +### Graceful Shutdown + +Go 1.8 introduced the ability to [gracefully shutdown](https://golang.org/doc/go1.8#http_shutdown) a `*http.Server`. Here's how to do that alongside `mux`: + +```go +package main + +import ( + "context" + "flag" + "log" + "net/http" + "os" + "os/signal" + "time" + + "github.com/gorilla/mux" +) + +func main() { + var wait time.Duration + flag.DurationVar(&wait, "graceful-timeout", time.Second * 15, "the duration for which the server gracefully wait for existing connections to finish - e.g. 15s or 1m") + flag.Parse() + + r := mux.NewRouter() + // Add your routes as needed + + srv := &http.Server{ + Addr: "0.0.0.0:8080", + // Good practice to set timeouts to avoid Slowloris attacks. + WriteTimeout: time.Second * 15, + ReadTimeout: time.Second * 15, + IdleTimeout: time.Second * 60, + Handler: r, // Pass our instance of gorilla/mux in. + } + + // Run our server in a goroutine so that it doesn't block. + go func() { + if err := srv.ListenAndServe(); err != nil { + log.Println(err) + } + }() + + c := make(chan os.Signal, 1) + // We'll accept graceful shutdowns when quit via SIGINT (Ctrl+C) + // SIGKILL, SIGQUIT or SIGTERM (Ctrl+/) will not be caught. + signal.Notify(c, os.Interrupt) + + // Block until we receive our signal. + <-c + + // Create a deadline to wait for. + ctx, cancel := context.WithTimeout(context.Background(), wait) + defer cancel() + // Doesn't block if no connections, but will otherwise wait + // until the timeout deadline. + srv.Shutdown(ctx) + // Optionally, you could run srv.Shutdown in a goroutine and block on + // <-ctx.Done() if your application should wait for other services + // to finalize based on context cancellation. + log.Println("shutting down") + os.Exit(0) +} +``` + +### Middleware + +Mux supports the addition of middlewares to a [Router](https://godoc.org/github.com/gorilla/mux#Router), which are executed in the order they are added if a match is found, including its subrouters. +Middlewares are (typically) small pieces of code which take one request, do something with it, and pass it down to another middleware or the final handler. Some common use cases for middleware are request logging, header manipulation, or `ResponseWriter` hijacking. + +Mux middlewares are defined using the de facto standard type: + +```go +type MiddlewareFunc func(http.Handler) http.Handler +``` + +Typically, the returned handler is a closure which does something with the http.ResponseWriter and http.Request passed to it, and then calls the handler passed as parameter to the MiddlewareFunc. This takes advantage of closures being able access variables from the context where they are created, while retaining the signature enforced by the receivers. + +A very basic middleware which logs the URI of the request being handled could be written as: + +```go +func loggingMiddleware(next http.Handler) http.Handler { + return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + // Do stuff here + log.Println(r.RequestURI) + // Call the next handler, which can be another middleware in the chain, or the final handler. + next.ServeHTTP(w, r) + }) +} +``` + +Middlewares can be added to a router using `Router.Use()`: + +```go +r := mux.NewRouter() +r.HandleFunc("/", handler) +r.Use(loggingMiddleware) +``` + +A more complex authentication middleware, which maps session token to users, could be written as: + +```go +// Define our struct +type authenticationMiddleware struct { + tokenUsers map[string]string +} + +// Initialize it somewhere +func (amw *authenticationMiddleware) Populate() { + amw.tokenUsers["00000000"] = "user0" + amw.tokenUsers["aaaaaaaa"] = "userA" + amw.tokenUsers["05f717e5"] = "randomUser" + amw.tokenUsers["deadbeef"] = "user0" +} + +// Middleware function, which will be called for each request +func (amw *authenticationMiddleware) Middleware(next http.Handler) http.Handler { + return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + token := r.Header.Get("X-Session-Token") + + if user, found := amw.tokenUsers[token]; found { + // We found the token in our map + log.Printf("Authenticated user %s\n", user) + // Pass down the request to the next middleware (or final handler) + next.ServeHTTP(w, r) + } else { + // Write an error and stop the handler chain + http.Error(w, "Forbidden", http.StatusForbidden) + } + }) +} +``` + +```go +r := mux.NewRouter() +r.HandleFunc("/", handler) + +amw := authenticationMiddleware{} +amw.Populate() + +r.Use(amw.Middleware) +``` + +Note: The handler chain will be stopped if your middleware doesn't call `next.ServeHTTP()` with the corresponding parameters. This can be used to abort a request if the middleware writer wants to. Middlewares _should_ write to `ResponseWriter` if they _are_ going to terminate the request, and they _should not_ write to `ResponseWriter` if they _are not_ going to terminate it. + +### Testing Handlers + +Testing handlers in a Go web application is straightforward, and _mux_ doesn't complicate this any further. Given two files: `endpoints.go` and `endpoints_test.go`, here's how we'd test an application using _mux_. + +First, our simple HTTP handler: + +```go +// endpoints.go +package main + +func HealthCheckHandler(w http.ResponseWriter, r *http.Request) { + // A very simple health check. + w.WriteHeader(http.StatusOK) + w.Header().Set("Content-Type", "application/json") + + // In the future we could report back on the status of our DB, or our cache + // (e.g. Redis) by performing a simple PING, and include them in the response. + io.WriteString(w, `{"alive": true}`) +} + +func main() { + r := mux.NewRouter() + r.HandleFunc("/health", HealthCheckHandler) + + log.Fatal(http.ListenAndServe("localhost:8080", r)) +} +``` + +Our test code: + +```go +// endpoints_test.go +package main + +import ( + "net/http" + "net/http/httptest" + "testing" +) + +func TestHealthCheckHandler(t *testing.T) { + // Create a request to pass to our handler. We don't have any query parameters for now, so we'll + // pass 'nil' as the third parameter. + req, err := http.NewRequest("GET", "/health", nil) + if err != nil { + t.Fatal(err) + } + + // We create a ResponseRecorder (which satisfies http.ResponseWriter) to record the response. + rr := httptest.NewRecorder() + handler := http.HandlerFunc(HealthCheckHandler) + + // Our handlers satisfy http.Handler, so we can call their ServeHTTP method + // directly and pass in our Request and ResponseRecorder. + handler.ServeHTTP(rr, req) + + // Check the status code is what we expect. + if status := rr.Code; status != http.StatusOK { + t.Errorf("handler returned wrong status code: got %v want %v", + status, http.StatusOK) + } + + // Check the response body is what we expect. + expected := `{"alive": true}` + if rr.Body.String() != expected { + t.Errorf("handler returned unexpected body: got %v want %v", + rr.Body.String(), expected) + } +} +``` + +In the case that our routes have [variables](#examples), we can pass those in the request. We could write +[table-driven tests](https://dave.cheney.net/2013/06/09/writing-table-driven-tests-in-go) to test multiple +possible route variables as needed. + +```go +// endpoints.go +func main() { + r := mux.NewRouter() + // A route with a route variable: + r.HandleFunc("/metrics/{type}", MetricsHandler) + + log.Fatal(http.ListenAndServe("localhost:8080", r)) +} +``` + +Our test file, with a table-driven test of `routeVariables`: + +```go +// endpoints_test.go +func TestMetricsHandler(t *testing.T) { + tt := []struct{ + routeVariable string + shouldPass bool + }{ + {"goroutines", true}, + {"heap", true}, + {"counters", true}, + {"queries", true}, + {"adhadaeqm3k", false}, + } + + for _, tc := range tt { + path := fmt.Sprintf("/metrics/%s", tc.routeVariable) + req, err := http.NewRequest("GET", path, nil) + if err != nil { + t.Fatal(err) + } + + rr := httptest.NewRecorder() + + // Need to create a router that we can pass the request through so that the vars will be added to the context + router := mux.NewRouter() + router.HandleFunc("/metrics/{type}", MetricsHandler) + router.ServeHTTP(rr, req) + + // In this case, our MetricsHandler returns a non-200 response + // for a route variable it doesn't know about. + if rr.Code == http.StatusOK && !tc.shouldPass { + t.Errorf("handler should have failed on routeVariable %s: got %v want %v", + tc.routeVariable, rr.Code, http.StatusOK) + } + } +} +``` + +## Full Example + +Here's a complete, runnable example of a small `mux` based server: + +```go +package main + +import ( + "net/http" + "log" + "github.com/gorilla/mux" +) + +func YourHandler(w http.ResponseWriter, r *http.Request) { + w.Write([]byte("Gorilla!\n")) +} + +func main() { + r := mux.NewRouter() + // Routes consist of a path and a handler function. + r.HandleFunc("/", YourHandler) + + // Bind to a port and pass our router in + log.Fatal(http.ListenAndServe(":8000", r)) +} +``` + +## License + +BSD licensed. See the LICENSE file for details. diff --git a/vendor/github.com/gorilla/mux/context_gorilla.go b/vendor/github.com/gorilla/mux/context_gorilla.go new file mode 100644 index 000000000..d7adaa8fa --- /dev/null +++ b/vendor/github.com/gorilla/mux/context_gorilla.go @@ -0,0 +1,26 @@ +// +build !go1.7 + +package mux + +import ( + "net/http" + + "github.com/gorilla/context" +) + +func contextGet(r *http.Request, key interface{}) interface{} { + return context.Get(r, key) +} + +func contextSet(r *http.Request, key, val interface{}) *http.Request { + if val == nil { + return r + } + + context.Set(r, key, val) + return r +} + +func contextClear(r *http.Request) { + context.Clear(r) +} diff --git a/vendor/github.com/gorilla/mux/context_native.go b/vendor/github.com/gorilla/mux/context_native.go new file mode 100644 index 000000000..209cbea7d --- /dev/null +++ b/vendor/github.com/gorilla/mux/context_native.go @@ -0,0 +1,24 @@ +// +build go1.7 + +package mux + +import ( + "context" + "net/http" +) + +func contextGet(r *http.Request, key interface{}) interface{} { + return r.Context().Value(key) +} + +func contextSet(r *http.Request, key, val interface{}) *http.Request { + if val == nil { + return r + } + + return r.WithContext(context.WithValue(r.Context(), key, val)) +} + +func contextClear(r *http.Request) { + return +} diff --git a/vendor/github.com/gorilla/mux/doc.go b/vendor/github.com/gorilla/mux/doc.go new file mode 100644 index 000000000..38957deea --- /dev/null +++ b/vendor/github.com/gorilla/mux/doc.go @@ -0,0 +1,306 @@ +// Copyright 2012 The Gorilla Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +/* +Package mux implements a request router and dispatcher. + +The name mux stands for "HTTP request multiplexer". Like the standard +http.ServeMux, mux.Router matches incoming requests against a list of +registered routes and calls a handler for the route that matches the URL +or other conditions. The main features are: + + * Requests can be matched based on URL host, path, path prefix, schemes, + header and query values, HTTP methods or using custom matchers. + * URL hosts, paths and query values can have variables with an optional + regular expression. + * Registered URLs can be built, or "reversed", which helps maintaining + references to resources. + * Routes can be used as subrouters: nested routes are only tested if the + parent route matches. This is useful to define groups of routes that + share common conditions like a host, a path prefix or other repeated + attributes. As a bonus, this optimizes request matching. + * It implements the http.Handler interface so it is compatible with the + standard http.ServeMux. + +Let's start registering a couple of URL paths and handlers: + + func main() { + r := mux.NewRouter() + r.HandleFunc("/", HomeHandler) + r.HandleFunc("/products", ProductsHandler) + r.HandleFunc("/articles", ArticlesHandler) + http.Handle("/", r) + } + +Here we register three routes mapping URL paths to handlers. This is +equivalent to how http.HandleFunc() works: if an incoming request URL matches +one of the paths, the corresponding handler is called passing +(http.ResponseWriter, *http.Request) as parameters. + +Paths can have variables. They are defined using the format {name} or +{name:pattern}. If a regular expression pattern is not defined, the matched +variable will be anything until the next slash. For example: + + r := mux.NewRouter() + r.HandleFunc("/products/{key}", ProductHandler) + r.HandleFunc("/articles/{category}/", ArticlesCategoryHandler) + r.HandleFunc("/articles/{category}/{id:[0-9]+}", ArticleHandler) + +Groups can be used inside patterns, as long as they are non-capturing (?:re). For example: + + r.HandleFunc("/articles/{category}/{sort:(?:asc|desc|new)}", ArticlesCategoryHandler) + +The names are used to create a map of route variables which can be retrieved +calling mux.Vars(): + + vars := mux.Vars(request) + category := vars["category"] + +Note that if any capturing groups are present, mux will panic() during parsing. To prevent +this, convert any capturing groups to non-capturing, e.g. change "/{sort:(asc|desc)}" to +"/{sort:(?:asc|desc)}". This is a change from prior versions which behaved unpredictably +when capturing groups were present. + +And this is all you need to know about the basic usage. More advanced options +are explained below. + +Routes can also be restricted to a domain or subdomain. Just define a host +pattern to be matched. They can also have variables: + + r := mux.NewRouter() + // Only matches if domain is "www.example.com". + r.Host("www.example.com") + // Matches a dynamic subdomain. + r.Host("{subdomain:[a-z]+}.domain.com") + +There are several other matchers that can be added. To match path prefixes: + + r.PathPrefix("/products/") + +...or HTTP methods: + + r.Methods("GET", "POST") + +...or URL schemes: + + r.Schemes("https") + +...or header values: + + r.Headers("X-Requested-With", "XMLHttpRequest") + +...or query values: + + r.Queries("key", "value") + +...or to use a custom matcher function: + + r.MatcherFunc(func(r *http.Request, rm *RouteMatch) bool { + return r.ProtoMajor == 0 + }) + +...and finally, it is possible to combine several matchers in a single route: + + r.HandleFunc("/products", ProductsHandler). + Host("www.example.com"). + Methods("GET"). + Schemes("http") + +Setting the same matching conditions again and again can be boring, so we have +a way to group several routes that share the same requirements. +We call it "subrouting". + +For example, let's say we have several URLs that should only match when the +host is "www.example.com". Create a route for that host and get a "subrouter" +from it: + + r := mux.NewRouter() + s := r.Host("www.example.com").Subrouter() + +Then register routes in the subrouter: + + s.HandleFunc("/products/", ProductsHandler) + s.HandleFunc("/products/{key}", ProductHandler) + s.HandleFunc("/articles/{category}/{id:[0-9]+}"), ArticleHandler) + +The three URL paths we registered above will only be tested if the domain is +"www.example.com", because the subrouter is tested first. This is not +only convenient, but also optimizes request matching. You can create +subrouters combining any attribute matchers accepted by a route. + +Subrouters can be used to create domain or path "namespaces": you define +subrouters in a central place and then parts of the app can register its +paths relatively to a given subrouter. + +There's one more thing about subroutes. When a subrouter has a path prefix, +the inner routes use it as base for their paths: + + r := mux.NewRouter() + s := r.PathPrefix("/products").Subrouter() + // "/products/" + s.HandleFunc("/", ProductsHandler) + // "/products/{key}/" + s.HandleFunc("/{key}/", ProductHandler) + // "/products/{key}/details" + s.HandleFunc("/{key}/details", ProductDetailsHandler) + +Note that the path provided to PathPrefix() represents a "wildcard": calling +PathPrefix("/static/").Handler(...) means that the handler will be passed any +request that matches "/static/*". This makes it easy to serve static files with mux: + + func main() { + var dir string + + flag.StringVar(&dir, "dir", ".", "the directory to serve files from. Defaults to the current dir") + flag.Parse() + r := mux.NewRouter() + + // This will serve files under http://localhost:8000/static/ + r.PathPrefix("/static/").Handler(http.StripPrefix("/static/", http.FileServer(http.Dir(dir)))) + + srv := &http.Server{ + Handler: r, + Addr: "127.0.0.1:8000", + // Good practice: enforce timeouts for servers you create! + WriteTimeout: 15 * time.Second, + ReadTimeout: 15 * time.Second, + } + + log.Fatal(srv.ListenAndServe()) + } + +Now let's see how to build registered URLs. + +Routes can be named. All routes that define a name can have their URLs built, +or "reversed". We define a name calling Name() on a route. For example: + + r := mux.NewRouter() + r.HandleFunc("/articles/{category}/{id:[0-9]+}", ArticleHandler). + Name("article") + +To build a URL, get the route and call the URL() method, passing a sequence of +key/value pairs for the route variables. For the previous route, we would do: + + url, err := r.Get("article").URL("category", "technology", "id", "42") + +...and the result will be a url.URL with the following path: + + "/articles/technology/42" + +This also works for host and query value variables: + + r := mux.NewRouter() + r.Host("{subdomain}.domain.com"). + Path("/articles/{category}/{id:[0-9]+}"). + Queries("filter", "{filter}"). + HandlerFunc(ArticleHandler). + Name("article") + + // url.String() will be "http://news.domain.com/articles/technology/42?filter=gorilla" + url, err := r.Get("article").URL("subdomain", "news", + "category", "technology", + "id", "42", + "filter", "gorilla") + +All variables defined in the route are required, and their values must +conform to the corresponding patterns. These requirements guarantee that a +generated URL will always match a registered route -- the only exception is +for explicitly defined "build-only" routes which never match. + +Regex support also exists for matching Headers within a route. For example, we could do: + + r.HeadersRegexp("Content-Type", "application/(text|json)") + +...and the route will match both requests with a Content-Type of `application/json` as well as +`application/text` + +There's also a way to build only the URL host or path for a route: +use the methods URLHost() or URLPath() instead. For the previous route, +we would do: + + // "http://news.domain.com/" + host, err := r.Get("article").URLHost("subdomain", "news") + + // "/articles/technology/42" + path, err := r.Get("article").URLPath("category", "technology", "id", "42") + +And if you use subrouters, host and path defined separately can be built +as well: + + r := mux.NewRouter() + s := r.Host("{subdomain}.domain.com").Subrouter() + s.Path("/articles/{category}/{id:[0-9]+}"). + HandlerFunc(ArticleHandler). + Name("article") + + // "http://news.domain.com/articles/technology/42" + url, err := r.Get("article").URL("subdomain", "news", + "category", "technology", + "id", "42") + +Mux supports the addition of middlewares to a Router, which are executed in the order they are added if a match is found, including its subrouters. Middlewares are (typically) small pieces of code which take one request, do something with it, and pass it down to another middleware or the final handler. Some common use cases for middleware are request logging, header manipulation, or ResponseWriter hijacking. + + type MiddlewareFunc func(http.Handler) http.Handler + +Typically, the returned handler is a closure which does something with the http.ResponseWriter and http.Request passed to it, and then calls the handler passed as parameter to the MiddlewareFunc (closures can access variables from the context where they are created). + +A very basic middleware which logs the URI of the request being handled could be written as: + + func simpleMw(next http.Handler) http.Handler { + return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + // Do stuff here + log.Println(r.RequestURI) + // Call the next handler, which can be another middleware in the chain, or the final handler. + next.ServeHTTP(w, r) + }) + } + +Middlewares can be added to a router using `Router.Use()`: + + r := mux.NewRouter() + r.HandleFunc("/", handler) + r.Use(simpleMw) + +A more complex authentication middleware, which maps session token to users, could be written as: + + // Define our struct + type authenticationMiddleware struct { + tokenUsers map[string]string + } + + // Initialize it somewhere + func (amw *authenticationMiddleware) Populate() { + amw.tokenUsers["00000000"] = "user0" + amw.tokenUsers["aaaaaaaa"] = "userA" + amw.tokenUsers["05f717e5"] = "randomUser" + amw.tokenUsers["deadbeef"] = "user0" + } + + // Middleware function, which will be called for each request + func (amw *authenticationMiddleware) Middleware(next http.Handler) http.Handler { + return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + token := r.Header.Get("X-Session-Token") + + if user, found := amw.tokenUsers[token]; found { + // We found the token in our map + log.Printf("Authenticated user %s\n", user) + next.ServeHTTP(w, r) + } else { + http.Error(w, "Forbidden", http.StatusForbidden) + } + }) + } + + r := mux.NewRouter() + r.HandleFunc("/", handler) + + amw := authenticationMiddleware{} + amw.Populate() + + r.Use(amw.Middleware) + +Note: The handler chain will be stopped if your middleware doesn't call `next.ServeHTTP()` with the corresponding parameters. This can be used to abort a request if the middleware writer wants to. + +*/ +package mux diff --git a/vendor/github.com/gorilla/mux/middleware.go b/vendor/github.com/gorilla/mux/middleware.go new file mode 100644 index 000000000..ceb812cee --- /dev/null +++ b/vendor/github.com/gorilla/mux/middleware.go @@ -0,0 +1,72 @@ +package mux + +import ( + "net/http" + "strings" +) + +// MiddlewareFunc is a function which receives an http.Handler and returns another http.Handler. +// Typically, the returned handler is a closure which does something with the http.ResponseWriter and http.Request passed +// to it, and then calls the handler passed as parameter to the MiddlewareFunc. +type MiddlewareFunc func(http.Handler) http.Handler + +// middleware interface is anything which implements a MiddlewareFunc named Middleware. +type middleware interface { + Middleware(handler http.Handler) http.Handler +} + +// Middleware allows MiddlewareFunc to implement the middleware interface. +func (mw MiddlewareFunc) Middleware(handler http.Handler) http.Handler { + return mw(handler) +} + +// Use appends a MiddlewareFunc to the chain. Middleware can be used to intercept or otherwise modify requests and/or responses, and are executed in the order that they are applied to the Router. +func (r *Router) Use(mwf ...MiddlewareFunc) { + for _, fn := range mwf { + r.middlewares = append(r.middlewares, fn) + } +} + +// useInterface appends a middleware to the chain. Middleware can be used to intercept or otherwise modify requests and/or responses, and are executed in the order that they are applied to the Router. +func (r *Router) useInterface(mw middleware) { + r.middlewares = append(r.middlewares, mw) +} + +// CORSMethodMiddleware sets the Access-Control-Allow-Methods response header +// on a request, by matching routes based only on paths. It also handles +// OPTIONS requests, by settings Access-Control-Allow-Methods, and then +// returning without calling the next http handler. +func CORSMethodMiddleware(r *Router) MiddlewareFunc { + return func(next http.Handler) http.Handler { + return http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) { + var allMethods []string + + err := r.Walk(func(route *Route, _ *Router, _ []*Route) error { + for _, m := range route.matchers { + if _, ok := m.(*routeRegexp); ok { + if m.Match(req, &RouteMatch{}) { + methods, err := route.GetMethods() + if err != nil { + return err + } + + allMethods = append(allMethods, methods...) + } + break + } + } + return nil + }) + + if err == nil { + w.Header().Set("Access-Control-Allow-Methods", strings.Join(append(allMethods, "OPTIONS"), ",")) + + if req.Method == "OPTIONS" { + return + } + } + + next.ServeHTTP(w, req) + }) + } +} diff --git a/vendor/github.com/gorilla/mux/mux.go b/vendor/github.com/gorilla/mux/mux.go new file mode 100644 index 000000000..4bbafa51d --- /dev/null +++ b/vendor/github.com/gorilla/mux/mux.go @@ -0,0 +1,588 @@ +// Copyright 2012 The Gorilla Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package mux + +import ( + "errors" + "fmt" + "net/http" + "path" + "regexp" +) + +var ( + // ErrMethodMismatch is returned when the method in the request does not match + // the method defined against the route. + ErrMethodMismatch = errors.New("method is not allowed") + // ErrNotFound is returned when no route match is found. + ErrNotFound = errors.New("no matching route was found") +) + +// NewRouter returns a new router instance. +func NewRouter() *Router { + return &Router{namedRoutes: make(map[string]*Route), KeepContext: false} +} + +// Router registers routes to be matched and dispatches a handler. +// +// It implements the http.Handler interface, so it can be registered to serve +// requests: +// +// var router = mux.NewRouter() +// +// func main() { +// http.Handle("/", router) +// } +// +// Or, for Google App Engine, register it in a init() function: +// +// func init() { +// http.Handle("/", router) +// } +// +// This will send all incoming requests to the router. +type Router struct { + // Configurable Handler to be used when no route matches. + NotFoundHandler http.Handler + + // Configurable Handler to be used when the request method does not match the route. + MethodNotAllowedHandler http.Handler + + // Parent route, if this is a subrouter. + parent parentRoute + // Routes to be matched, in order. + routes []*Route + // Routes by name for URL building. + namedRoutes map[string]*Route + // See Router.StrictSlash(). This defines the flag for new routes. + strictSlash bool + // See Router.SkipClean(). This defines the flag for new routes. + skipClean bool + // If true, do not clear the request context after handling the request. + // This has no effect when go1.7+ is used, since the context is stored + // on the request itself. + KeepContext bool + // see Router.UseEncodedPath(). This defines a flag for all routes. + useEncodedPath bool + // Slice of middlewares to be called after a match is found + middlewares []middleware +} + +// Match attempts to match the given request against the router's registered routes. +// +// If the request matches a route of this router or one of its subrouters the Route, +// Handler, and Vars fields of the the match argument are filled and this function +// returns true. +// +// If the request does not match any of this router's or its subrouters' routes +// then this function returns false. If available, a reason for the match failure +// will be filled in the match argument's MatchErr field. If the match failure type +// (eg: not found) has a registered handler, the handler is assigned to the Handler +// field of the match argument. +func (r *Router) Match(req *http.Request, match *RouteMatch) bool { + for _, route := range r.routes { + if route.Match(req, match) { + // Build middleware chain if no error was found + if match.MatchErr == nil { + for i := len(r.middlewares) - 1; i >= 0; i-- { + match.Handler = r.middlewares[i].Middleware(match.Handler) + } + } + return true + } + } + + if match.MatchErr == ErrMethodMismatch { + if r.MethodNotAllowedHandler != nil { + match.Handler = r.MethodNotAllowedHandler + return true + } + + return false + } + + // Closest match for a router (includes sub-routers) + if r.NotFoundHandler != nil { + match.Handler = r.NotFoundHandler + match.MatchErr = ErrNotFound + return true + } + + match.MatchErr = ErrNotFound + return false +} + +// ServeHTTP dispatches the handler registered in the matched route. +// +// When there is a match, the route variables can be retrieved calling +// mux.Vars(request). +func (r *Router) ServeHTTP(w http.ResponseWriter, req *http.Request) { + if !r.skipClean { + path := req.URL.Path + if r.useEncodedPath { + path = req.URL.EscapedPath() + } + // Clean path to canonical form and redirect. + if p := cleanPath(path); p != path { + + // Added 3 lines (Philip Schlump) - It was dropping the query string and #whatever from query. + // This matches with fix in go 1.2 r.c. 4 for same problem. Go Issue: + // http://code.google.com/p/go/issues/detail?id=5252 + url := *req.URL + url.Path = p + p = url.String() + + w.Header().Set("Location", p) + w.WriteHeader(http.StatusMovedPermanently) + return + } + } + var match RouteMatch + var handler http.Handler + if r.Match(req, &match) { + handler = match.Handler + req = setVars(req, match.Vars) + req = setCurrentRoute(req, match.Route) + } + + if handler == nil && match.MatchErr == ErrMethodMismatch { + handler = methodNotAllowedHandler() + } + + if handler == nil { + handler = http.NotFoundHandler() + } + + if !r.KeepContext { + defer contextClear(req) + } + + handler.ServeHTTP(w, req) +} + +// Get returns a route registered with the given name. +func (r *Router) Get(name string) *Route { + return r.getNamedRoutes()[name] +} + +// GetRoute returns a route registered with the given name. This method +// was renamed to Get() and remains here for backwards compatibility. +func (r *Router) GetRoute(name string) *Route { + return r.getNamedRoutes()[name] +} + +// StrictSlash defines the trailing slash behavior for new routes. The initial +// value is false. +// +// When true, if the route path is "/path/", accessing "/path" will perform a redirect +// to the former and vice versa. In other words, your application will always +// see the path as specified in the route. +// +// When false, if the route path is "/path", accessing "/path/" will not match +// this route and vice versa. +// +// The re-direct is a HTTP 301 (Moved Permanently). Note that when this is set for +// routes with a non-idempotent method (e.g. POST, PUT), the subsequent re-directed +// request will be made as a GET by most clients. Use middleware or client settings +// to modify this behaviour as needed. +// +// Special case: when a route sets a path prefix using the PathPrefix() method, +// strict slash is ignored for that route because the redirect behavior can't +// be determined from a prefix alone. However, any subrouters created from that +// route inherit the original StrictSlash setting. +func (r *Router) StrictSlash(value bool) *Router { + r.strictSlash = value + return r +} + +// SkipClean defines the path cleaning behaviour for new routes. The initial +// value is false. Users should be careful about which routes are not cleaned +// +// When true, if the route path is "/path//to", it will remain with the double +// slash. This is helpful if you have a route like: /fetch/http://xkcd.com/534/ +// +// When false, the path will be cleaned, so /fetch/http://xkcd.com/534/ will +// become /fetch/http/xkcd.com/534 +func (r *Router) SkipClean(value bool) *Router { + r.skipClean = value + return r +} + +// UseEncodedPath tells the router to match the encoded original path +// to the routes. +// For eg. "/path/foo%2Fbar/to" will match the path "/path/{var}/to". +// +// If not called, the router will match the unencoded path to the routes. +// For eg. "/path/foo%2Fbar/to" will match the path "/path/foo/bar/to" +func (r *Router) UseEncodedPath() *Router { + r.useEncodedPath = true + return r +} + +// ---------------------------------------------------------------------------- +// parentRoute +// ---------------------------------------------------------------------------- + +func (r *Router) getBuildScheme() string { + if r.parent != nil { + return r.parent.getBuildScheme() + } + return "" +} + +// getNamedRoutes returns the map where named routes are registered. +func (r *Router) getNamedRoutes() map[string]*Route { + if r.namedRoutes == nil { + if r.parent != nil { + r.namedRoutes = r.parent.getNamedRoutes() + } else { + r.namedRoutes = make(map[string]*Route) + } + } + return r.namedRoutes +} + +// getRegexpGroup returns regexp definitions from the parent route, if any. +func (r *Router) getRegexpGroup() *routeRegexpGroup { + if r.parent != nil { + return r.parent.getRegexpGroup() + } + return nil +} + +func (r *Router) buildVars(m map[string]string) map[string]string { + if r.parent != nil { + m = r.parent.buildVars(m) + } + return m +} + +// ---------------------------------------------------------------------------- +// Route factories +// ---------------------------------------------------------------------------- + +// NewRoute registers an empty route. +func (r *Router) NewRoute() *Route { + route := &Route{parent: r, strictSlash: r.strictSlash, skipClean: r.skipClean, useEncodedPath: r.useEncodedPath} + r.routes = append(r.routes, route) + return route +} + +// Handle registers a new route with a matcher for the URL path. +// See Route.Path() and Route.Handler(). +func (r *Router) Handle(path string, handler http.Handler) *Route { + return r.NewRoute().Path(path).Handler(handler) +} + +// HandleFunc registers a new route with a matcher for the URL path. +// See Route.Path() and Route.HandlerFunc(). +func (r *Router) HandleFunc(path string, f func(http.ResponseWriter, + *http.Request)) *Route { + return r.NewRoute().Path(path).HandlerFunc(f) +} + +// Headers registers a new route with a matcher for request header values. +// See Route.Headers(). +func (r *Router) Headers(pairs ...string) *Route { + return r.NewRoute().Headers(pairs...) +} + +// Host registers a new route with a matcher for the URL host. +// See Route.Host(). +func (r *Router) Host(tpl string) *Route { + return r.NewRoute().Host(tpl) +} + +// MatcherFunc registers a new route with a custom matcher function. +// See Route.MatcherFunc(). +func (r *Router) MatcherFunc(f MatcherFunc) *Route { + return r.NewRoute().MatcherFunc(f) +} + +// Methods registers a new route with a matcher for HTTP methods. +// See Route.Methods(). +func (r *Router) Methods(methods ...string) *Route { + return r.NewRoute().Methods(methods...) +} + +// Path registers a new route with a matcher for the URL path. +// See Route.Path(). +func (r *Router) Path(tpl string) *Route { + return r.NewRoute().Path(tpl) +} + +// PathPrefix registers a new route with a matcher for the URL path prefix. +// See Route.PathPrefix(). +func (r *Router) PathPrefix(tpl string) *Route { + return r.NewRoute().PathPrefix(tpl) +} + +// Queries registers a new route with a matcher for URL query values. +// See Route.Queries(). +func (r *Router) Queries(pairs ...string) *Route { + return r.NewRoute().Queries(pairs...) +} + +// Schemes registers a new route with a matcher for URL schemes. +// See Route.Schemes(). +func (r *Router) Schemes(schemes ...string) *Route { + return r.NewRoute().Schemes(schemes...) +} + +// BuildVarsFunc registers a new route with a custom function for modifying +// route variables before building a URL. +func (r *Router) BuildVarsFunc(f BuildVarsFunc) *Route { + return r.NewRoute().BuildVarsFunc(f) +} + +// Walk walks the router and all its sub-routers, calling walkFn for each route +// in the tree. The routes are walked in the order they were added. Sub-routers +// are explored depth-first. +func (r *Router) Walk(walkFn WalkFunc) error { + return r.walk(walkFn, []*Route{}) +} + +// SkipRouter is used as a return value from WalkFuncs to indicate that the +// router that walk is about to descend down to should be skipped. +var SkipRouter = errors.New("skip this router") + +// WalkFunc is the type of the function called for each route visited by Walk. +// At every invocation, it is given the current route, and the current router, +// and a list of ancestor routes that lead to the current route. +type WalkFunc func(route *Route, router *Router, ancestors []*Route) error + +func (r *Router) walk(walkFn WalkFunc, ancestors []*Route) error { + for _, t := range r.routes { + err := walkFn(t, r, ancestors) + if err == SkipRouter { + continue + } + if err != nil { + return err + } + for _, sr := range t.matchers { + if h, ok := sr.(*Router); ok { + ancestors = append(ancestors, t) + err := h.walk(walkFn, ancestors) + if err != nil { + return err + } + ancestors = ancestors[:len(ancestors)-1] + } + } + if h, ok := t.handler.(*Router); ok { + ancestors = append(ancestors, t) + err := h.walk(walkFn, ancestors) + if err != nil { + return err + } + ancestors = ancestors[:len(ancestors)-1] + } + } + return nil +} + +// ---------------------------------------------------------------------------- +// Context +// ---------------------------------------------------------------------------- + +// RouteMatch stores information about a matched route. +type RouteMatch struct { + Route *Route + Handler http.Handler + Vars map[string]string + + // MatchErr is set to appropriate matching error + // It is set to ErrMethodMismatch if there is a mismatch in + // the request method and route method + MatchErr error +} + +type contextKey int + +const ( + varsKey contextKey = iota + routeKey +) + +// Vars returns the route variables for the current request, if any. +func Vars(r *http.Request) map[string]string { + if rv := contextGet(r, varsKey); rv != nil { + return rv.(map[string]string) + } + return nil +} + +// CurrentRoute returns the matched route for the current request, if any. +// This only works when called inside the handler of the matched route +// because the matched route is stored in the request context which is cleared +// after the handler returns, unless the KeepContext option is set on the +// Router. +func CurrentRoute(r *http.Request) *Route { + if rv := contextGet(r, routeKey); rv != nil { + return rv.(*Route) + } + return nil +} + +func setVars(r *http.Request, val interface{}) *http.Request { + return contextSet(r, varsKey, val) +} + +func setCurrentRoute(r *http.Request, val interface{}) *http.Request { + return contextSet(r, routeKey, val) +} + +// ---------------------------------------------------------------------------- +// Helpers +// ---------------------------------------------------------------------------- + +// cleanPath returns the canonical path for p, eliminating . and .. elements. +// Borrowed from the net/http package. +func cleanPath(p string) string { + if p == "" { + return "/" + } + if p[0] != '/' { + p = "/" + p + } + np := path.Clean(p) + // path.Clean removes trailing slash except for root; + // put the trailing slash back if necessary. + if p[len(p)-1] == '/' && np != "/" { + np += "/" + } + + return np +} + +// uniqueVars returns an error if two slices contain duplicated strings. +func uniqueVars(s1, s2 []string) error { + for _, v1 := range s1 { + for _, v2 := range s2 { + if v1 == v2 { + return fmt.Errorf("mux: duplicated route variable %q", v2) + } + } + } + return nil +} + +// checkPairs returns the count of strings passed in, and an error if +// the count is not an even number. +func checkPairs(pairs ...string) (int, error) { + length := len(pairs) + if length%2 != 0 { + return length, fmt.Errorf( + "mux: number of parameters must be multiple of 2, got %v", pairs) + } + return length, nil +} + +// mapFromPairsToString converts variadic string parameters to a +// string to string map. +func mapFromPairsToString(pairs ...string) (map[string]string, error) { + length, err := checkPairs(pairs...) + if err != nil { + return nil, err + } + m := make(map[string]string, length/2) + for i := 0; i < length; i += 2 { + m[pairs[i]] = pairs[i+1] + } + return m, nil +} + +// mapFromPairsToRegex converts variadic string parameters to a +// string to regex map. +func mapFromPairsToRegex(pairs ...string) (map[string]*regexp.Regexp, error) { + length, err := checkPairs(pairs...) + if err != nil { + return nil, err + } + m := make(map[string]*regexp.Regexp, length/2) + for i := 0; i < length; i += 2 { + regex, err := regexp.Compile(pairs[i+1]) + if err != nil { + return nil, err + } + m[pairs[i]] = regex + } + return m, nil +} + +// matchInArray returns true if the given string value is in the array. +func matchInArray(arr []string, value string) bool { + for _, v := range arr { + if v == value { + return true + } + } + return false +} + +// matchMapWithString returns true if the given key/value pairs exist in a given map. +func matchMapWithString(toCheck map[string]string, toMatch map[string][]string, canonicalKey bool) bool { + for k, v := range toCheck { + // Check if key exists. + if canonicalKey { + k = http.CanonicalHeaderKey(k) + } + if values := toMatch[k]; values == nil { + return false + } else if v != "" { + // If value was defined as an empty string we only check that the + // key exists. Otherwise we also check for equality. + valueExists := false + for _, value := range values { + if v == value { + valueExists = true + break + } + } + if !valueExists { + return false + } + } + } + return true +} + +// matchMapWithRegex returns true if the given key/value pairs exist in a given map compiled against +// the given regex +func matchMapWithRegex(toCheck map[string]*regexp.Regexp, toMatch map[string][]string, canonicalKey bool) bool { + for k, v := range toCheck { + // Check if key exists. + if canonicalKey { + k = http.CanonicalHeaderKey(k) + } + if values := toMatch[k]; values == nil { + return false + } else if v != nil { + // If value was defined as an empty string we only check that the + // key exists. Otherwise we also check for equality. + valueExists := false + for _, value := range values { + if v.MatchString(value) { + valueExists = true + break + } + } + if !valueExists { + return false + } + } + } + return true +} + +// methodNotAllowed replies to the request with an HTTP status code 405. +func methodNotAllowed(w http.ResponseWriter, r *http.Request) { + w.WriteHeader(http.StatusMethodNotAllowed) +} + +// methodNotAllowedHandler returns a simple request handler +// that replies to each request with a status code 405. +func methodNotAllowedHandler() http.Handler { return http.HandlerFunc(methodNotAllowed) } diff --git a/vendor/github.com/gorilla/mux/regexp.go b/vendor/github.com/gorilla/mux/regexp.go new file mode 100644 index 000000000..2b57e5627 --- /dev/null +++ b/vendor/github.com/gorilla/mux/regexp.go @@ -0,0 +1,332 @@ +// Copyright 2012 The Gorilla Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package mux + +import ( + "bytes" + "fmt" + "net/http" + "net/url" + "regexp" + "strconv" + "strings" +) + +type routeRegexpOptions struct { + strictSlash bool + useEncodedPath bool +} + +type regexpType int + +const ( + regexpTypePath regexpType = 0 + regexpTypeHost regexpType = 1 + regexpTypePrefix regexpType = 2 + regexpTypeQuery regexpType = 3 +) + +// newRouteRegexp parses a route template and returns a routeRegexp, +// used to match a host, a path or a query string. +// +// It will extract named variables, assemble a regexp to be matched, create +// a "reverse" template to build URLs and compile regexps to validate variable +// values used in URL building. +// +// Previously we accepted only Python-like identifiers for variable +// names ([a-zA-Z_][a-zA-Z0-9_]*), but currently the only restriction is that +// name and pattern can't be empty, and names can't contain a colon. +func newRouteRegexp(tpl string, typ regexpType, options routeRegexpOptions) (*routeRegexp, error) { + // Check if it is well-formed. + idxs, errBraces := braceIndices(tpl) + if errBraces != nil { + return nil, errBraces + } + // Backup the original. + template := tpl + // Now let's parse it. + defaultPattern := "[^/]+" + if typ == regexpTypeQuery { + defaultPattern = ".*" + } else if typ == regexpTypeHost { + defaultPattern = "[^.]+" + } + // Only match strict slash if not matching + if typ != regexpTypePath { + options.strictSlash = false + } + // Set a flag for strictSlash. + endSlash := false + if options.strictSlash && strings.HasSuffix(tpl, "/") { + tpl = tpl[:len(tpl)-1] + endSlash = true + } + varsN := make([]string, len(idxs)/2) + varsR := make([]*regexp.Regexp, len(idxs)/2) + pattern := bytes.NewBufferString("") + pattern.WriteByte('^') + reverse := bytes.NewBufferString("") + var end int + var err error + for i := 0; i < len(idxs); i += 2 { + // Set all values we are interested in. + raw := tpl[end:idxs[i]] + end = idxs[i+1] + parts := strings.SplitN(tpl[idxs[i]+1:end-1], ":", 2) + name := parts[0] + patt := defaultPattern + if len(parts) == 2 { + patt = parts[1] + } + // Name or pattern can't be empty. + if name == "" || patt == "" { + return nil, fmt.Errorf("mux: missing name or pattern in %q", + tpl[idxs[i]:end]) + } + // Build the regexp pattern. + fmt.Fprintf(pattern, "%s(?P<%s>%s)", regexp.QuoteMeta(raw), varGroupName(i/2), patt) + + // Build the reverse template. + fmt.Fprintf(reverse, "%s%%s", raw) + + // Append variable name and compiled pattern. + varsN[i/2] = name + varsR[i/2], err = regexp.Compile(fmt.Sprintf("^%s$", patt)) + if err != nil { + return nil, err + } + } + // Add the remaining. + raw := tpl[end:] + pattern.WriteString(regexp.QuoteMeta(raw)) + if options.strictSlash { + pattern.WriteString("[/]?") + } + if typ == regexpTypeQuery { + // Add the default pattern if the query value is empty + if queryVal := strings.SplitN(template, "=", 2)[1]; queryVal == "" { + pattern.WriteString(defaultPattern) + } + } + if typ != regexpTypePrefix { + pattern.WriteByte('$') + } + reverse.WriteString(raw) + if endSlash { + reverse.WriteByte('/') + } + // Compile full regexp. + reg, errCompile := regexp.Compile(pattern.String()) + if errCompile != nil { + return nil, errCompile + } + + // Check for capturing groups which used to work in older versions + if reg.NumSubexp() != len(idxs)/2 { + panic(fmt.Sprintf("route %s contains capture groups in its regexp. ", template) + + "Only non-capturing groups are accepted: e.g. (?:pattern) instead of (pattern)") + } + + // Done! + return &routeRegexp{ + template: template, + regexpType: typ, + options: options, + regexp: reg, + reverse: reverse.String(), + varsN: varsN, + varsR: varsR, + }, nil +} + +// routeRegexp stores a regexp to match a host or path and information to +// collect and validate route variables. +type routeRegexp struct { + // The unmodified template. + template string + // The type of match + regexpType regexpType + // Options for matching + options routeRegexpOptions + // Expanded regexp. + regexp *regexp.Regexp + // Reverse template. + reverse string + // Variable names. + varsN []string + // Variable regexps (validators). + varsR []*regexp.Regexp +} + +// Match matches the regexp against the URL host or path. +func (r *routeRegexp) Match(req *http.Request, match *RouteMatch) bool { + if r.regexpType != regexpTypeHost { + if r.regexpType == regexpTypeQuery { + return r.matchQueryString(req) + } + path := req.URL.Path + if r.options.useEncodedPath { + path = req.URL.EscapedPath() + } + return r.regexp.MatchString(path) + } + + return r.regexp.MatchString(getHost(req)) +} + +// url builds a URL part using the given values. +func (r *routeRegexp) url(values map[string]string) (string, error) { + urlValues := make([]interface{}, len(r.varsN)) + for k, v := range r.varsN { + value, ok := values[v] + if !ok { + return "", fmt.Errorf("mux: missing route variable %q", v) + } + if r.regexpType == regexpTypeQuery { + value = url.QueryEscape(value) + } + urlValues[k] = value + } + rv := fmt.Sprintf(r.reverse, urlValues...) + if !r.regexp.MatchString(rv) { + // The URL is checked against the full regexp, instead of checking + // individual variables. This is faster but to provide a good error + // message, we check individual regexps if the URL doesn't match. + for k, v := range r.varsN { + if !r.varsR[k].MatchString(values[v]) { + return "", fmt.Errorf( + "mux: variable %q doesn't match, expected %q", values[v], + r.varsR[k].String()) + } + } + } + return rv, nil +} + +// getURLQuery returns a single query parameter from a request URL. +// For a URL with foo=bar&baz=ding, we return only the relevant key +// value pair for the routeRegexp. +func (r *routeRegexp) getURLQuery(req *http.Request) string { + if r.regexpType != regexpTypeQuery { + return "" + } + templateKey := strings.SplitN(r.template, "=", 2)[0] + for key, vals := range req.URL.Query() { + if key == templateKey && len(vals) > 0 { + return key + "=" + vals[0] + } + } + return "" +} + +func (r *routeRegexp) matchQueryString(req *http.Request) bool { + return r.regexp.MatchString(r.getURLQuery(req)) +} + +// braceIndices returns the first level curly brace indices from a string. +// It returns an error in case of unbalanced braces. +func braceIndices(s string) ([]int, error) { + var level, idx int + var idxs []int + for i := 0; i < len(s); i++ { + switch s[i] { + case '{': + if level++; level == 1 { + idx = i + } + case '}': + if level--; level == 0 { + idxs = append(idxs, idx, i+1) + } else if level < 0 { + return nil, fmt.Errorf("mux: unbalanced braces in %q", s) + } + } + } + if level != 0 { + return nil, fmt.Errorf("mux: unbalanced braces in %q", s) + } + return idxs, nil +} + +// varGroupName builds a capturing group name for the indexed variable. +func varGroupName(idx int) string { + return "v" + strconv.Itoa(idx) +} + +// ---------------------------------------------------------------------------- +// routeRegexpGroup +// ---------------------------------------------------------------------------- + +// routeRegexpGroup groups the route matchers that carry variables. +type routeRegexpGroup struct { + host *routeRegexp + path *routeRegexp + queries []*routeRegexp +} + +// setMatch extracts the variables from the URL once a route matches. +func (v *routeRegexpGroup) setMatch(req *http.Request, m *RouteMatch, r *Route) { + // Store host variables. + if v.host != nil { + host := getHost(req) + matches := v.host.regexp.FindStringSubmatchIndex(host) + if len(matches) > 0 { + extractVars(host, matches, v.host.varsN, m.Vars) + } + } + path := req.URL.Path + if r.useEncodedPath { + path = req.URL.EscapedPath() + } + // Store path variables. + if v.path != nil { + matches := v.path.regexp.FindStringSubmatchIndex(path) + if len(matches) > 0 { + extractVars(path, matches, v.path.varsN, m.Vars) + // Check if we should redirect. + if v.path.options.strictSlash { + p1 := strings.HasSuffix(path, "/") + p2 := strings.HasSuffix(v.path.template, "/") + if p1 != p2 { + u, _ := url.Parse(req.URL.String()) + if p1 { + u.Path = u.Path[:len(u.Path)-1] + } else { + u.Path += "/" + } + m.Handler = http.RedirectHandler(u.String(), 301) + } + } + } + } + // Store query string variables. + for _, q := range v.queries { + queryURL := q.getURLQuery(req) + matches := q.regexp.FindStringSubmatchIndex(queryURL) + if len(matches) > 0 { + extractVars(queryURL, matches, q.varsN, m.Vars) + } + } +} + +// getHost tries its best to return the request host. +func getHost(r *http.Request) string { + if r.URL.IsAbs() { + return r.URL.Host + } + host := r.Host + // Slice off any port information. + if i := strings.Index(host, ":"); i != -1 { + host = host[:i] + } + return host + +} + +func extractVars(input string, matches []int, names []string, output map[string]string) { + for i, name := range names { + output[name] = input[matches[2*i+2]:matches[2*i+3]] + } +} diff --git a/vendor/github.com/gorilla/mux/route.go b/vendor/github.com/gorilla/mux/route.go new file mode 100644 index 000000000..c8bb5c7e3 --- /dev/null +++ b/vendor/github.com/gorilla/mux/route.go @@ -0,0 +1,763 @@ +// Copyright 2012 The Gorilla Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package mux + +import ( + "errors" + "fmt" + "net/http" + "net/url" + "regexp" + "strings" +) + +// Route stores information to match a request and build URLs. +type Route struct { + // Parent where the route was registered (a Router). + parent parentRoute + // Request handler for the route. + handler http.Handler + // List of matchers. + matchers []matcher + // Manager for the variables from host and path. + regexp *routeRegexpGroup + // If true, when the path pattern is "/path/", accessing "/path" will + // redirect to the former and vice versa. + strictSlash bool + // If true, when the path pattern is "/path//to", accessing "/path//to" + // will not redirect + skipClean bool + // If true, "/path/foo%2Fbar/to" will match the path "/path/{var}/to" + useEncodedPath bool + // The scheme used when building URLs. + buildScheme string + // If true, this route never matches: it is only used to build URLs. + buildOnly bool + // The name used to build URLs. + name string + // Error resulted from building a route. + err error + + buildVarsFunc BuildVarsFunc +} + +// SkipClean reports whether path cleaning is enabled for this route via +// Router.SkipClean. +func (r *Route) SkipClean() bool { + return r.skipClean +} + +// Match matches the route against the request. +func (r *Route) Match(req *http.Request, match *RouteMatch) bool { + if r.buildOnly || r.err != nil { + return false + } + + var matchErr error + + // Match everything. + for _, m := range r.matchers { + if matched := m.Match(req, match); !matched { + if _, ok := m.(methodMatcher); ok { + matchErr = ErrMethodMismatch + continue + } + matchErr = nil + return false + } + } + + if matchErr != nil { + match.MatchErr = matchErr + return false + } + + if match.MatchErr == ErrMethodMismatch { + // We found a route which matches request method, clear MatchErr + match.MatchErr = nil + // Then override the mis-matched handler + match.Handler = r.handler + } + + // Yay, we have a match. Let's collect some info about it. + if match.Route == nil { + match.Route = r + } + if match.Handler == nil { + match.Handler = r.handler + } + if match.Vars == nil { + match.Vars = make(map[string]string) + } + + // Set variables. + if r.regexp != nil { + r.regexp.setMatch(req, match, r) + } + return true +} + +// ---------------------------------------------------------------------------- +// Route attributes +// ---------------------------------------------------------------------------- + +// GetError returns an error resulted from building the route, if any. +func (r *Route) GetError() error { + return r.err +} + +// BuildOnly sets the route to never match: it is only used to build URLs. +func (r *Route) BuildOnly() *Route { + r.buildOnly = true + return r +} + +// Handler -------------------------------------------------------------------- + +// Handler sets a handler for the route. +func (r *Route) Handler(handler http.Handler) *Route { + if r.err == nil { + r.handler = handler + } + return r +} + +// HandlerFunc sets a handler function for the route. +func (r *Route) HandlerFunc(f func(http.ResponseWriter, *http.Request)) *Route { + return r.Handler(http.HandlerFunc(f)) +} + +// GetHandler returns the handler for the route, if any. +func (r *Route) GetHandler() http.Handler { + return r.handler +} + +// Name ----------------------------------------------------------------------- + +// Name sets the name for the route, used to build URLs. +// It is an error to call Name more than once on a route. +func (r *Route) Name(name string) *Route { + if r.name != "" { + r.err = fmt.Errorf("mux: route already has name %q, can't set %q", + r.name, name) + } + if r.err == nil { + r.name = name + r.getNamedRoutes()[name] = r + } + return r +} + +// GetName returns the name for the route, if any. +func (r *Route) GetName() string { + return r.name +} + +// ---------------------------------------------------------------------------- +// Matchers +// ---------------------------------------------------------------------------- + +// matcher types try to match a request. +type matcher interface { + Match(*http.Request, *RouteMatch) bool +} + +// addMatcher adds a matcher to the route. +func (r *Route) addMatcher(m matcher) *Route { + if r.err == nil { + r.matchers = append(r.matchers, m) + } + return r +} + +// addRegexpMatcher adds a host or path matcher and builder to a route. +func (r *Route) addRegexpMatcher(tpl string, typ regexpType) error { + if r.err != nil { + return r.err + } + r.regexp = r.getRegexpGroup() + if typ == regexpTypePath || typ == regexpTypePrefix { + if len(tpl) > 0 && tpl[0] != '/' { + return fmt.Errorf("mux: path must start with a slash, got %q", tpl) + } + if r.regexp.path != nil { + tpl = strings.TrimRight(r.regexp.path.template, "/") + tpl + } + } + rr, err := newRouteRegexp(tpl, typ, routeRegexpOptions{ + strictSlash: r.strictSlash, + useEncodedPath: r.useEncodedPath, + }) + if err != nil { + return err + } + for _, q := range r.regexp.queries { + if err = uniqueVars(rr.varsN, q.varsN); err != nil { + return err + } + } + if typ == regexpTypeHost { + if r.regexp.path != nil { + if err = uniqueVars(rr.varsN, r.regexp.path.varsN); err != nil { + return err + } + } + r.regexp.host = rr + } else { + if r.regexp.host != nil { + if err = uniqueVars(rr.varsN, r.regexp.host.varsN); err != nil { + return err + } + } + if typ == regexpTypeQuery { + r.regexp.queries = append(r.regexp.queries, rr) + } else { + r.regexp.path = rr + } + } + r.addMatcher(rr) + return nil +} + +// Headers -------------------------------------------------------------------- + +// headerMatcher matches the request against header values. +type headerMatcher map[string]string + +func (m headerMatcher) Match(r *http.Request, match *RouteMatch) bool { + return matchMapWithString(m, r.Header, true) +} + +// Headers adds a matcher for request header values. +// It accepts a sequence of key/value pairs to be matched. For example: +// +// r := mux.NewRouter() +// r.Headers("Content-Type", "application/json", +// "X-Requested-With", "XMLHttpRequest") +// +// The above route will only match if both request header values match. +// If the value is an empty string, it will match any value if the key is set. +func (r *Route) Headers(pairs ...string) *Route { + if r.err == nil { + var headers map[string]string + headers, r.err = mapFromPairsToString(pairs...) + return r.addMatcher(headerMatcher(headers)) + } + return r +} + +// headerRegexMatcher matches the request against the route given a regex for the header +type headerRegexMatcher map[string]*regexp.Regexp + +func (m headerRegexMatcher) Match(r *http.Request, match *RouteMatch) bool { + return matchMapWithRegex(m, r.Header, true) +} + +// HeadersRegexp accepts a sequence of key/value pairs, where the value has regex +// support. For example: +// +// r := mux.NewRouter() +// r.HeadersRegexp("Content-Type", "application/(text|json)", +// "X-Requested-With", "XMLHttpRequest") +// +// The above route will only match if both the request header matches both regular expressions. +// If the value is an empty string, it will match any value if the key is set. +// Use the start and end of string anchors (^ and $) to match an exact value. +func (r *Route) HeadersRegexp(pairs ...string) *Route { + if r.err == nil { + var headers map[string]*regexp.Regexp + headers, r.err = mapFromPairsToRegex(pairs...) + return r.addMatcher(headerRegexMatcher(headers)) + } + return r +} + +// Host ----------------------------------------------------------------------- + +// Host adds a matcher for the URL host. +// It accepts a template with zero or more URL variables enclosed by {}. +// Variables can define an optional regexp pattern to be matched: +// +// - {name} matches anything until the next dot. +// +// - {name:pattern} matches the given regexp pattern. +// +// For example: +// +// r := mux.NewRouter() +// r.Host("www.example.com") +// r.Host("{subdomain}.domain.com") +// r.Host("{subdomain:[a-z]+}.domain.com") +// +// Variable names must be unique in a given route. They can be retrieved +// calling mux.Vars(request). +func (r *Route) Host(tpl string) *Route { + r.err = r.addRegexpMatcher(tpl, regexpTypeHost) + return r +} + +// MatcherFunc ---------------------------------------------------------------- + +// MatcherFunc is the function signature used by custom matchers. +type MatcherFunc func(*http.Request, *RouteMatch) bool + +// Match returns the match for a given request. +func (m MatcherFunc) Match(r *http.Request, match *RouteMatch) bool { + return m(r, match) +} + +// MatcherFunc adds a custom function to be used as request matcher. +func (r *Route) MatcherFunc(f MatcherFunc) *Route { + return r.addMatcher(f) +} + +// Methods -------------------------------------------------------------------- + +// methodMatcher matches the request against HTTP methods. +type methodMatcher []string + +func (m methodMatcher) Match(r *http.Request, match *RouteMatch) bool { + return matchInArray(m, r.Method) +} + +// Methods adds a matcher for HTTP methods. +// It accepts a sequence of one or more methods to be matched, e.g.: +// "GET", "POST", "PUT". +func (r *Route) Methods(methods ...string) *Route { + for k, v := range methods { + methods[k] = strings.ToUpper(v) + } + return r.addMatcher(methodMatcher(methods)) +} + +// Path ----------------------------------------------------------------------- + +// Path adds a matcher for the URL path. +// It accepts a template with zero or more URL variables enclosed by {}. The +// template must start with a "/". +// Variables can define an optional regexp pattern to be matched: +// +// - {name} matches anything until the next slash. +// +// - {name:pattern} matches the given regexp pattern. +// +// For example: +// +// r := mux.NewRouter() +// r.Path("/products/").Handler(ProductsHandler) +// r.Path("/products/{key}").Handler(ProductsHandler) +// r.Path("/articles/{category}/{id:[0-9]+}"). +// Handler(ArticleHandler) +// +// Variable names must be unique in a given route. They can be retrieved +// calling mux.Vars(request). +func (r *Route) Path(tpl string) *Route { + r.err = r.addRegexpMatcher(tpl, regexpTypePath) + return r +} + +// PathPrefix ----------------------------------------------------------------- + +// PathPrefix adds a matcher for the URL path prefix. This matches if the given +// template is a prefix of the full URL path. See Route.Path() for details on +// the tpl argument. +// +// Note that it does not treat slashes specially ("/foobar/" will be matched by +// the prefix "/foo") so you may want to use a trailing slash here. +// +// Also note that the setting of Router.StrictSlash() has no effect on routes +// with a PathPrefix matcher. +func (r *Route) PathPrefix(tpl string) *Route { + r.err = r.addRegexpMatcher(tpl, regexpTypePrefix) + return r +} + +// Query ---------------------------------------------------------------------- + +// Queries adds a matcher for URL query values. +// It accepts a sequence of key/value pairs. Values may define variables. +// For example: +// +// r := mux.NewRouter() +// r.Queries("foo", "bar", "id", "{id:[0-9]+}") +// +// The above route will only match if the URL contains the defined queries +// values, e.g.: ?foo=bar&id=42. +// +// It the value is an empty string, it will match any value if the key is set. +// +// Variables can define an optional regexp pattern to be matched: +// +// - {name} matches anything until the next slash. +// +// - {name:pattern} matches the given regexp pattern. +func (r *Route) Queries(pairs ...string) *Route { + length := len(pairs) + if length%2 != 0 { + r.err = fmt.Errorf( + "mux: number of parameters must be multiple of 2, got %v", pairs) + return nil + } + for i := 0; i < length; i += 2 { + if r.err = r.addRegexpMatcher(pairs[i]+"="+pairs[i+1], regexpTypeQuery); r.err != nil { + return r + } + } + + return r +} + +// Schemes -------------------------------------------------------------------- + +// schemeMatcher matches the request against URL schemes. +type schemeMatcher []string + +func (m schemeMatcher) Match(r *http.Request, match *RouteMatch) bool { + return matchInArray(m, r.URL.Scheme) +} + +// Schemes adds a matcher for URL schemes. +// It accepts a sequence of schemes to be matched, e.g.: "http", "https". +func (r *Route) Schemes(schemes ...string) *Route { + for k, v := range schemes { + schemes[k] = strings.ToLower(v) + } + if r.buildScheme == "" && len(schemes) > 0 { + r.buildScheme = schemes[0] + } + return r.addMatcher(schemeMatcher(schemes)) +} + +// BuildVarsFunc -------------------------------------------------------------- + +// BuildVarsFunc is the function signature used by custom build variable +// functions (which can modify route variables before a route's URL is built). +type BuildVarsFunc func(map[string]string) map[string]string + +// BuildVarsFunc adds a custom function to be used to modify build variables +// before a route's URL is built. +func (r *Route) BuildVarsFunc(f BuildVarsFunc) *Route { + r.buildVarsFunc = f + return r +} + +// Subrouter ------------------------------------------------------------------ + +// Subrouter creates a subrouter for the route. +// +// It will test the inner routes only if the parent route matched. For example: +// +// r := mux.NewRouter() +// s := r.Host("www.example.com").Subrouter() +// s.HandleFunc("/products/", ProductsHandler) +// s.HandleFunc("/products/{key}", ProductHandler) +// s.HandleFunc("/articles/{category}/{id:[0-9]+}"), ArticleHandler) +// +// Here, the routes registered in the subrouter won't be tested if the host +// doesn't match. +func (r *Route) Subrouter() *Router { + router := &Router{parent: r, strictSlash: r.strictSlash} + r.addMatcher(router) + return router +} + +// ---------------------------------------------------------------------------- +// URL building +// ---------------------------------------------------------------------------- + +// URL builds a URL for the route. +// +// It accepts a sequence of key/value pairs for the route variables. For +// example, given this route: +// +// r := mux.NewRouter() +// r.HandleFunc("/articles/{category}/{id:[0-9]+}", ArticleHandler). +// Name("article") +// +// ...a URL for it can be built using: +// +// url, err := r.Get("article").URL("category", "technology", "id", "42") +// +// ...which will return an url.URL with the following path: +// +// "/articles/technology/42" +// +// This also works for host variables: +// +// r := mux.NewRouter() +// r.Host("{subdomain}.domain.com"). +// HandleFunc("/articles/{category}/{id:[0-9]+}", ArticleHandler). +// Name("article") +// +// // url.String() will be "http://news.domain.com/articles/technology/42" +// url, err := r.Get("article").URL("subdomain", "news", +// "category", "technology", +// "id", "42") +// +// All variables defined in the route are required, and their values must +// conform to the corresponding patterns. +func (r *Route) URL(pairs ...string) (*url.URL, error) { + if r.err != nil { + return nil, r.err + } + if r.regexp == nil { + return nil, errors.New("mux: route doesn't have a host or path") + } + values, err := r.prepareVars(pairs...) + if err != nil { + return nil, err + } + var scheme, host, path string + queries := make([]string, 0, len(r.regexp.queries)) + if r.regexp.host != nil { + if host, err = r.regexp.host.url(values); err != nil { + return nil, err + } + scheme = "http" + if s := r.getBuildScheme(); s != "" { + scheme = s + } + } + if r.regexp.path != nil { + if path, err = r.regexp.path.url(values); err != nil { + return nil, err + } + } + for _, q := range r.regexp.queries { + var query string + if query, err = q.url(values); err != nil { + return nil, err + } + queries = append(queries, query) + } + return &url.URL{ + Scheme: scheme, + Host: host, + Path: path, + RawQuery: strings.Join(queries, "&"), + }, nil +} + +// URLHost builds the host part of the URL for a route. See Route.URL(). +// +// The route must have a host defined. +func (r *Route) URLHost(pairs ...string) (*url.URL, error) { + if r.err != nil { + return nil, r.err + } + if r.regexp == nil || r.regexp.host == nil { + return nil, errors.New("mux: route doesn't have a host") + } + values, err := r.prepareVars(pairs...) + if err != nil { + return nil, err + } + host, err := r.regexp.host.url(values) + if err != nil { + return nil, err + } + u := &url.URL{ + Scheme: "http", + Host: host, + } + if s := r.getBuildScheme(); s != "" { + u.Scheme = s + } + return u, nil +} + +// URLPath builds the path part of the URL for a route. See Route.URL(). +// +// The route must have a path defined. +func (r *Route) URLPath(pairs ...string) (*url.URL, error) { + if r.err != nil { + return nil, r.err + } + if r.regexp == nil || r.regexp.path == nil { + return nil, errors.New("mux: route doesn't have a path") + } + values, err := r.prepareVars(pairs...) + if err != nil { + return nil, err + } + path, err := r.regexp.path.url(values) + if err != nil { + return nil, err + } + return &url.URL{ + Path: path, + }, nil +} + +// GetPathTemplate returns the template used to build the +// route match. +// This is useful for building simple REST API documentation and for instrumentation +// against third-party services. +// An error will be returned if the route does not define a path. +func (r *Route) GetPathTemplate() (string, error) { + if r.err != nil { + return "", r.err + } + if r.regexp == nil || r.regexp.path == nil { + return "", errors.New("mux: route doesn't have a path") + } + return r.regexp.path.template, nil +} + +// GetPathRegexp returns the expanded regular expression used to match route path. +// This is useful for building simple REST API documentation and for instrumentation +// against third-party services. +// An error will be returned if the route does not define a path. +func (r *Route) GetPathRegexp() (string, error) { + if r.err != nil { + return "", r.err + } + if r.regexp == nil || r.regexp.path == nil { + return "", errors.New("mux: route does not have a path") + } + return r.regexp.path.regexp.String(), nil +} + +// GetQueriesRegexp returns the expanded regular expressions used to match the +// route queries. +// This is useful for building simple REST API documentation and for instrumentation +// against third-party services. +// An error will be returned if the route does not have queries. +func (r *Route) GetQueriesRegexp() ([]string, error) { + if r.err != nil { + return nil, r.err + } + if r.regexp == nil || r.regexp.queries == nil { + return nil, errors.New("mux: route doesn't have queries") + } + var queries []string + for _, query := range r.regexp.queries { + queries = append(queries, query.regexp.String()) + } + return queries, nil +} + +// GetQueriesTemplates returns the templates used to build the +// query matching. +// This is useful for building simple REST API documentation and for instrumentation +// against third-party services. +// An error will be returned if the route does not define queries. +func (r *Route) GetQueriesTemplates() ([]string, error) { + if r.err != nil { + return nil, r.err + } + if r.regexp == nil || r.regexp.queries == nil { + return nil, errors.New("mux: route doesn't have queries") + } + var queries []string + for _, query := range r.regexp.queries { + queries = append(queries, query.template) + } + return queries, nil +} + +// GetMethods returns the methods the route matches against +// This is useful for building simple REST API documentation and for instrumentation +// against third-party services. +// An error will be returned if route does not have methods. +func (r *Route) GetMethods() ([]string, error) { + if r.err != nil { + return nil, r.err + } + for _, m := range r.matchers { + if methods, ok := m.(methodMatcher); ok { + return []string(methods), nil + } + } + return nil, errors.New("mux: route doesn't have methods") +} + +// GetHostTemplate returns the template used to build the +// route match. +// This is useful for building simple REST API documentation and for instrumentation +// against third-party services. +// An error will be returned if the route does not define a host. +func (r *Route) GetHostTemplate() (string, error) { + if r.err != nil { + return "", r.err + } + if r.regexp == nil || r.regexp.host == nil { + return "", errors.New("mux: route doesn't have a host") + } + return r.regexp.host.template, nil +} + +// prepareVars converts the route variable pairs into a map. If the route has a +// BuildVarsFunc, it is invoked. +func (r *Route) prepareVars(pairs ...string) (map[string]string, error) { + m, err := mapFromPairsToString(pairs...) + if err != nil { + return nil, err + } + return r.buildVars(m), nil +} + +func (r *Route) buildVars(m map[string]string) map[string]string { + if r.parent != nil { + m = r.parent.buildVars(m) + } + if r.buildVarsFunc != nil { + m = r.buildVarsFunc(m) + } + return m +} + +// ---------------------------------------------------------------------------- +// parentRoute +// ---------------------------------------------------------------------------- + +// parentRoute allows routes to know about parent host and path definitions. +type parentRoute interface { + getBuildScheme() string + getNamedRoutes() map[string]*Route + getRegexpGroup() *routeRegexpGroup + buildVars(map[string]string) map[string]string +} + +func (r *Route) getBuildScheme() string { + if r.buildScheme != "" { + return r.buildScheme + } + if r.parent != nil { + return r.parent.getBuildScheme() + } + return "" +} + +// getNamedRoutes returns the map where named routes are registered. +func (r *Route) getNamedRoutes() map[string]*Route { + if r.parent == nil { + // During tests router is not always set. + r.parent = NewRouter() + } + return r.parent.getNamedRoutes() +} + +// getRegexpGroup returns regexp definitions from this route. +func (r *Route) getRegexpGroup() *routeRegexpGroup { + if r.regexp == nil { + if r.parent == nil { + // During tests router is not always set. + r.parent = NewRouter() + } + regexp := r.parent.getRegexpGroup() + if regexp == nil { + r.regexp = new(routeRegexpGroup) + } else { + // Copy. + r.regexp = &routeRegexpGroup{ + host: regexp.host, + path: regexp.path, + queries: regexp.queries, + } + } + } + return r.regexp +} diff --git a/vendor/github.com/gorilla/mux/test_helpers.go b/vendor/github.com/gorilla/mux/test_helpers.go new file mode 100644 index 000000000..32ecffde4 --- /dev/null +++ b/vendor/github.com/gorilla/mux/test_helpers.go @@ -0,0 +1,19 @@ +// Copyright 2012 The Gorilla Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package mux + +import "net/http" + +// SetURLVars sets the URL variables for the given request, to be accessed via +// mux.Vars for testing route behaviour. Arguments are not modified, a shallow +// copy is returned. +// +// This API should only be used for testing purposes; it provides a way to +// inject variables into the request context. Alternatively, URL variables +// can be set by making a route that captures the required variables, +// starting a server and sending the request to that server. +func SetURLVars(r *http.Request, val map[string]string) *http.Request { + return setVars(r, val) +} diff --git a/vendor/github.com/hashicorp/go-msgpack/codec/msgpack_test.py b/vendor/github.com/hashicorp/go-msgpack/codec/msgpack_test.py old mode 100644 new mode 100755 diff --git a/vendor/github.com/hashicorp/go-syslog/.gitignore b/vendor/github.com/hashicorp/go-syslog/.gitignore deleted file mode 100644 index 00268614f..000000000 --- a/vendor/github.com/hashicorp/go-syslog/.gitignore +++ /dev/null @@ -1,22 +0,0 @@ -# Compiled Object files, Static and Dynamic libs (Shared Objects) -*.o -*.a -*.so - -# Folders -_obj -_test - -# Architecture specific extensions/prefixes -*.[568vq] -[568vq].out - -*.cgo1.go -*.cgo2.c -_cgo_defun.c -_cgo_gotypes.go -_cgo_export.* - -_testmain.go - -*.exe diff --git a/vendor/github.com/hashicorp/hcl/appveyor.yml b/vendor/github.com/hashicorp/hcl/appveyor.yml new file mode 100644 index 000000000..3c8cdf8e9 --- /dev/null +++ b/vendor/github.com/hashicorp/hcl/appveyor.yml @@ -0,0 +1,19 @@ +version: "build-{branch}-{build}" +image: Visual Studio 2015 +clone_folder: c:\gopath\src\github.com\hashicorp\hcl +environment: + GOPATH: c:\gopath +init: + - git config --global core.autocrlf true +install: +- cmd: >- + echo %Path% + + go version + + go env + + go get -t ./... + +build_script: +- cmd: go test -v ./... diff --git a/vendor/github.com/hashicorp/logutils/.gitignore b/vendor/github.com/hashicorp/logutils/.gitignore deleted file mode 100644 index 00268614f..000000000 --- a/vendor/github.com/hashicorp/logutils/.gitignore +++ /dev/null @@ -1,22 +0,0 @@ -# Compiled Object files, Static and Dynamic libs (Shared Objects) -*.o -*.a -*.so - -# Folders -_obj -_test - -# Architecture specific extensions/prefixes -*.[568vq] -[568vq].out - -*.cgo1.go -*.cgo2.c -_cgo_defun.c -_cgo_gotypes.go -_cgo_export.* - -_testmain.go - -*.exe diff --git a/vendor/github.com/hashicorp/vault/main.go b/vendor/github.com/hashicorp/vault/main.go new file mode 100644 index 000000000..6cd34fe36 --- /dev/null +++ b/vendor/github.com/hashicorp/vault/main.go @@ -0,0 +1,11 @@ +package main // import "github.com/hashicorp/vault" + +import ( + "os" + + "github.com/hashicorp/vault/cli" +) + +func main() { + os.Exit(cli.Run(os.Args[1:])) +} diff --git a/vendor/github.com/jmespath/go-jmespath/.gitignore b/vendor/github.com/jmespath/go-jmespath/.gitignore deleted file mode 100644 index 531fcc11c..000000000 --- a/vendor/github.com/jmespath/go-jmespath/.gitignore +++ /dev/null @@ -1,4 +0,0 @@ -jpgo -jmespath-fuzz.zip -cpu.out -go-jmespath.test diff --git a/vendor/github.com/jmespath/go-jmespath/.travis.yml b/vendor/github.com/jmespath/go-jmespath/.travis.yml deleted file mode 100644 index 1f9807757..000000000 --- a/vendor/github.com/jmespath/go-jmespath/.travis.yml +++ /dev/null @@ -1,9 +0,0 @@ -language: go - -sudo: false - -go: - - 1.4 - -install: go get -v -t ./... -script: make test diff --git a/vendor/github.com/miekg/dns/.gitignore b/vendor/github.com/miekg/dns/.gitignore deleted file mode 100644 index 776cd950c..000000000 --- a/vendor/github.com/miekg/dns/.gitignore +++ /dev/null @@ -1,4 +0,0 @@ -*.6 -tags -test.out -a.out diff --git a/vendor/github.com/miekg/dns/.travis.yml b/vendor/github.com/miekg/dns/.travis.yml deleted file mode 100644 index 1f056ab7c..000000000 --- a/vendor/github.com/miekg/dns/.travis.yml +++ /dev/null @@ -1,7 +0,0 @@ -language: go -sudo: false -go: - - 1.5 - - 1.6 -script: - - go test -race -v -bench=. diff --git a/vendor/github.com/miekg/dns/labels.go b/vendor/github.com/miekg/dns/labels.go index bfd8cf561..fca5c7dd2 100644 --- a/vendor/github.com/miekg/dns/labels.go +++ b/vendor/github.com/miekg/dns/labels.go @@ -88,7 +88,7 @@ func CompareDomainName(s1, s2 string) (n int) { return } -// CountLabel counts the number of labels in the string s. +// CountLabel counts the the number of labels in the string s. // s must be a syntactically valid domain name. func CountLabel(s string) (labels int) { if s == "." { diff --git a/vendor/github.com/mitchellh/mapstructure/.travis.yml b/vendor/github.com/mitchellh/mapstructure/.travis.yml deleted file mode 100644 index 7f3fe9a96..000000000 --- a/vendor/github.com/mitchellh/mapstructure/.travis.yml +++ /dev/null @@ -1,7 +0,0 @@ -language: go - -go: - - 1.4 - -script: - - go test diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/symlink/LICENSE.APACHE b/vendor/github.com/mrunalp/fileutils/LICENSE similarity index 99% rename from vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/symlink/LICENSE.APACHE rename to vendor/github.com/mrunalp/fileutils/LICENSE index 9e4bd4dbe..27448585a 100644 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/symlink/LICENSE.APACHE +++ b/vendor/github.com/mrunalp/fileutils/LICENSE @@ -176,7 +176,7 @@ END OF TERMS AND CONDITIONS - Copyright 2014-2015 Docker, Inc. + Copyright 2014 Docker, Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. diff --git a/vendor/github.com/mrunalp/fileutils/README.md b/vendor/github.com/mrunalp/fileutils/README.md new file mode 100644 index 000000000..d15692488 --- /dev/null +++ b/vendor/github.com/mrunalp/fileutils/README.md @@ -0,0 +1,5 @@ +# fileutils + +Collection of utilities for file manipulation in golang + +The library is based on docker pkg/archive but does copies instead of handling archive formats. diff --git a/vendor/github.com/mrunalp/fileutils/fileutils.go b/vendor/github.com/mrunalp/fileutils/fileutils.go new file mode 100644 index 000000000..b60cb909c --- /dev/null +++ b/vendor/github.com/mrunalp/fileutils/fileutils.go @@ -0,0 +1,161 @@ +package fileutils + +import ( + "fmt" + "io" + "os" + "path/filepath" + "syscall" +) + +// CopyFile copies the file at source to dest +func CopyFile(source string, dest string) error { + si, err := os.Lstat(source) + if err != nil { + return err + } + + st, ok := si.Sys().(*syscall.Stat_t) + if !ok { + return fmt.Errorf("could not convert to syscall.Stat_t") + } + + uid := int(st.Uid) + gid := int(st.Gid) + + // Handle symlinks + if si.Mode()&os.ModeSymlink != 0 { + target, err := os.Readlink(source) + if err != nil { + return err + } + if err := os.Symlink(target, dest); err != nil { + return err + } + } + + // Handle device files + if st.Mode&syscall.S_IFMT == syscall.S_IFBLK || st.Mode&syscall.S_IFMT == syscall.S_IFCHR { + devMajor := int64(major(uint64(st.Rdev))) + devMinor := int64(minor(uint64(st.Rdev))) + mode := uint32(si.Mode() & 07777) + if st.Mode&syscall.S_IFMT == syscall.S_IFBLK { + mode |= syscall.S_IFBLK + } + if st.Mode&syscall.S_IFMT == syscall.S_IFCHR { + mode |= syscall.S_IFCHR + } + if err := syscall.Mknod(dest, mode, int(mkdev(devMajor, devMinor))); err != nil { + return err + } + } + + // Handle regular files + if si.Mode().IsRegular() { + sf, err := os.Open(source) + if err != nil { + return err + } + defer sf.Close() + + df, err := os.Create(dest) + if err != nil { + return err + } + defer df.Close() + + _, err = io.Copy(df, sf) + if err != nil { + return err + } + } + + // Chown the file + if err := os.Lchown(dest, uid, gid); err != nil { + return err + } + + // Chmod the file + if !(si.Mode()&os.ModeSymlink == os.ModeSymlink) { + if err := os.Chmod(dest, si.Mode()); err != nil { + return err + } + } + + return nil +} + +// CopyDirectory copies the files under the source directory +// to dest directory. The dest directory is created if it +// does not exist. +func CopyDirectory(source string, dest string) error { + fi, err := os.Stat(source) + if err != nil { + return err + } + + // Get owner. + st, ok := fi.Sys().(*syscall.Stat_t) + if !ok { + return fmt.Errorf("could not convert to syscall.Stat_t") + } + + // We have to pick an owner here anyway. + if err := MkdirAllNewAs(dest, fi.Mode(), int(st.Uid), int(st.Gid)); err != nil { + return err + } + + return filepath.Walk(source, func(path string, info os.FileInfo, err error) error { + if err != nil { + return err + } + + // Get the relative path + relPath, err := filepath.Rel(source, path) + if err != nil { + return nil + } + + if info.IsDir() { + // Skip the source directory. + if path != source { + // Get the owner. + st, ok := info.Sys().(*syscall.Stat_t) + if !ok { + return fmt.Errorf("could not convert to syscall.Stat_t") + } + + uid := int(st.Uid) + gid := int(st.Gid) + + if err := os.Mkdir(filepath.Join(dest, relPath), info.Mode()); err != nil { + return err + } + + if err := os.Lchown(filepath.Join(dest, relPath), uid, gid); err != nil { + return err + } + } + return nil + } + + // Copy the file. + if err := CopyFile(path, filepath.Join(dest, relPath)); err != nil { + return err + } + + return nil + }) +} + +func major(device uint64) uint64 { + return (device >> 8) & 0xfff +} + +func minor(device uint64) uint64 { + return (device & 0xff) | ((device >> 12) & 0xfff00) +} + +func mkdev(major int64, minor int64) uint32 { + return uint32(((minor & 0xfff00) << 12) | ((major & 0xfff) << 8) | (minor & 0xff)) +} diff --git a/vendor/github.com/mrunalp/fileutils/idtools.go b/vendor/github.com/mrunalp/fileutils/idtools.go new file mode 100644 index 000000000..161aec8f5 --- /dev/null +++ b/vendor/github.com/mrunalp/fileutils/idtools.go @@ -0,0 +1,49 @@ +package fileutils + +import ( + "os" + "path/filepath" +) + +// MkdirAllNewAs creates a directory (include any along the path) and then modifies +// ownership ONLY of newly created directories to the requested uid/gid. If the +// directories along the path exist, no change of ownership will be performed +func MkdirAllNewAs(path string, mode os.FileMode, ownerUID, ownerGID int) error { + // make an array containing the original path asked for, plus (for mkAll == true) + // all path components leading up to the complete path that don't exist before we MkdirAll + // so that we can chown all of them properly at the end. If chownExisting is false, we won't + // chown the full directory path if it exists + var paths []string + if _, err := os.Stat(path); err != nil && os.IsNotExist(err) { + paths = []string{path} + } else if err == nil { + // nothing to do; directory path fully exists already + return nil + } + + // walk back to "/" looking for directories which do not exist + // and add them to the paths array for chown after creation + dirPath := path + for { + dirPath = filepath.Dir(dirPath) + if dirPath == "/" { + break + } + if _, err := os.Stat(dirPath); err != nil && os.IsNotExist(err) { + paths = append(paths, dirPath) + } + } + + if err := os.MkdirAll(path, mode); err != nil && !os.IsExist(err) { + return err + } + + // even if it existed, we will chown the requested path + any subpaths that + // didn't exist when we called MkdirAll + for _, pathComponent := range paths { + if err := os.Chown(pathComponent, ownerUID, ownerGID); err != nil { + return err + } + } + return nil +} diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/Sirupsen/logrus/LICENSE b/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/Sirupsen/logrus/LICENSE deleted file mode 100644 index f090cb42f..000000000 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/Sirupsen/logrus/LICENSE +++ /dev/null @@ -1,21 +0,0 @@ -The MIT License (MIT) - -Copyright (c) 2014 Simon Eskildsen - -Permission is hereby granted, free of charge, to any person obtaining a copy -of this software and associated documentation files (the "Software"), to deal -in the Software without restriction, including without limitation the rights -to use, copy, modify, merge, publish, distribute, sublicense, and/or sell -copies of the Software, and to permit persons to whom the Software is -furnished to do so, subject to the following conditions: - -The above copyright notice and this permission notice shall be included in -all copies or substantial portions of the Software. - -THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR -IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, -FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE -AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER -LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, -OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN -THE SOFTWARE. diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/codegangsta/cli/LICENSE b/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/codegangsta/cli/LICENSE deleted file mode 100644 index 5515ccfb7..000000000 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/codegangsta/cli/LICENSE +++ /dev/null @@ -1,21 +0,0 @@ -Copyright (C) 2013 Jeremy Saenz -All Rights Reserved. - -MIT LICENSE - -Permission is hereby granted, free of charge, to any person obtaining a copy of -this software and associated documentation files (the "Software"), to deal in -the Software without restriction, including without limitation the rights to -use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of -the Software, and to permit persons to whom the Software is furnished to do so, -subject to the following conditions: - -The above copyright notice and this permission notice shall be included in all -copies or substantial portions of the Software. - -THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR -IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS -FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR -COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER -IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN -CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/contrib/syntax/vim/LICENSE b/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/contrib/syntax/vim/LICENSE deleted file mode 100644 index e67cdabd2..000000000 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/contrib/syntax/vim/LICENSE +++ /dev/null @@ -1,22 +0,0 @@ -Copyright (c) 2013 Honza Pokorny -All rights reserved. - -Redistribution and use in source and binary forms, with or without -modification, are permitted provided that the following conditions are met: - -1. Redistributions of source code must retain the above copyright - notice, this list of conditions and the following disclaimer. -2. Redistributions in binary form must reproduce the above copyright - notice, this list of conditions and the following disclaimer in the - documentation and/or other materials provided with the distribution. - -THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND -ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED -WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE -DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR -ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES -(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; -LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND -ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS -SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/docs/project/images/red_notice.png b/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/docs/project/images/red_notice.png deleted file mode 100644 index d953b32d8..000000000 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/docs/project/images/red_notice.png +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:548caadbe447f53c000549162b602bd80d202458d4b151e71ac80b2c83cd0a33 -size 45387 diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/symlink/LICENSE.BSD b/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/symlink/LICENSE.BSD deleted file mode 100644 index ac74d8f04..000000000 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/docker/docker/pkg/symlink/LICENSE.BSD +++ /dev/null @@ -1,27 +0,0 @@ -Copyright (c) 2014-2015 The Docker & Go Authors. All rights reserved. - -Redistribution and use in source and binary forms, with or without -modification, are permitted provided that the following conditions are -met: - - * Redistributions of source code must retain the above copyright -notice, this list of conditions and the following disclaimer. - * Redistributions in binary form must reproduce the above -copyright notice, this list of conditions and the following disclaimer -in the documentation and/or other materials provided with the -distribution. - * Neither the name of Google Inc. nor the names of its -contributors may be used to endorse or promote products derived from -this software without specific prior written permission. - -THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/golang/protobuf/LICENSE b/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/golang/protobuf/LICENSE deleted file mode 100644 index 1b1b1921e..000000000 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/golang/protobuf/LICENSE +++ /dev/null @@ -1,31 +0,0 @@ -Go support for Protocol Buffers - Google's data interchange format - -Copyright 2010 The Go Authors. All rights reserved. -https://github.com/golang/protobuf - -Redistribution and use in source and binary forms, with or without -modification, are permitted provided that the following conditions are -met: - - * Redistributions of source code must retain the above copyright -notice, this list of conditions and the following disclaimer. - * Redistributions in binary form must reproduce the above -copyright notice, this list of conditions and the following disclaimer -in the documentation and/or other materials provided with the -distribution. - * Neither the name of Google Inc. nor the names of its -contributors may be used to endorse or promote products derived from -this software without specific prior written permission. - -THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/syndtr/gocapability/LICENSE b/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/syndtr/gocapability/LICENSE deleted file mode 100644 index 80dd96de7..000000000 --- a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/syndtr/gocapability/LICENSE +++ /dev/null @@ -1,24 +0,0 @@ -Copyright 2013 Suryandaru Triandana -All rights reserved. - -Redistribution and use in source and binary forms, with or without -modification, are permitted provided that the following conditions are -met: - - * Redistributions of source code must retain the above copyright -notice, this list of conditions and the following disclaimer. - * Redistributions in binary form must reproduce the above copyright -notice, this list of conditions and the following disclaimer in the -documentation and/or other materials provided with the distribution. - -THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. diff --git a/vendor/github.com/opencontainers/runc/libcontainer/README.md b/vendor/github.com/opencontainers/runc/libcontainer/README.md new file mode 100644 index 000000000..640ea9a01 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/README.md @@ -0,0 +1,327 @@ +# libcontainer + +[![GoDoc](https://godoc.org/github.com/opencontainers/runc/libcontainer?status.svg)](https://godoc.org/github.com/opencontainers/runc/libcontainer) + +Libcontainer provides a native Go implementation for creating containers +with namespaces, cgroups, capabilities, and filesystem access controls. +It allows you to manage the lifecycle of the container performing additional operations +after the container is created. + + +#### Container +A container is a self contained execution environment that shares the kernel of the +host system and which is (optionally) isolated from other containers in the system. + +#### Using libcontainer + +Because containers are spawned in a two step process you will need a binary that +will be executed as the init process for the container. In libcontainer, we use +the current binary (/proc/self/exe) to be executed as the init process, and use +arg "init", we call the first step process "bootstrap", so you always need a "init" +function as the entry of "bootstrap". + +In addition to the go init function the early stage bootstrap is handled by importing +[nsenter](https://github.com/opencontainers/runc/blob/master/libcontainer/nsenter/README.md). + +```go +import ( + _ "github.com/opencontainers/runc/libcontainer/nsenter" +) + +func init() { + if len(os.Args) > 1 && os.Args[1] == "init" { + runtime.GOMAXPROCS(1) + runtime.LockOSThread() + factory, _ := libcontainer.New("") + if err := factory.StartInitialization(); err != nil { + logrus.Fatal(err) + } + panic("--this line should have never been executed, congratulations--") + } +} +``` + +Then to create a container you first have to initialize an instance of a factory +that will handle the creation and initialization for a container. + +```go +factory, err := libcontainer.New("/var/lib/container", libcontainer.Cgroupfs, libcontainer.InitArgs(os.Args[0], "init")) +if err != nil { + logrus.Fatal(err) + return +} +``` + +Once you have an instance of the factory created we can create a configuration +struct describing how the container is to be created. A sample would look similar to this: + +```go +defaultMountFlags := unix.MS_NOEXEC | unix.MS_NOSUID | unix.MS_NODEV +config := &configs.Config{ + Rootfs: "/your/path/to/rootfs", + Capabilities: &configs.Capabilities{ + Bounding: []string{ + "CAP_CHOWN", + "CAP_DAC_OVERRIDE", + "CAP_FSETID", + "CAP_FOWNER", + "CAP_MKNOD", + "CAP_NET_RAW", + "CAP_SETGID", + "CAP_SETUID", + "CAP_SETFCAP", + "CAP_SETPCAP", + "CAP_NET_BIND_SERVICE", + "CAP_SYS_CHROOT", + "CAP_KILL", + "CAP_AUDIT_WRITE", + }, + Effective: []string{ + "CAP_CHOWN", + "CAP_DAC_OVERRIDE", + "CAP_FSETID", + "CAP_FOWNER", + "CAP_MKNOD", + "CAP_NET_RAW", + "CAP_SETGID", + "CAP_SETUID", + "CAP_SETFCAP", + "CAP_SETPCAP", + "CAP_NET_BIND_SERVICE", + "CAP_SYS_CHROOT", + "CAP_KILL", + "CAP_AUDIT_WRITE", + }, + Inheritable: []string{ + "CAP_CHOWN", + "CAP_DAC_OVERRIDE", + "CAP_FSETID", + "CAP_FOWNER", + "CAP_MKNOD", + "CAP_NET_RAW", + "CAP_SETGID", + "CAP_SETUID", + "CAP_SETFCAP", + "CAP_SETPCAP", + "CAP_NET_BIND_SERVICE", + "CAP_SYS_CHROOT", + "CAP_KILL", + "CAP_AUDIT_WRITE", + }, + Permitted: []string{ + "CAP_CHOWN", + "CAP_DAC_OVERRIDE", + "CAP_FSETID", + "CAP_FOWNER", + "CAP_MKNOD", + "CAP_NET_RAW", + "CAP_SETGID", + "CAP_SETUID", + "CAP_SETFCAP", + "CAP_SETPCAP", + "CAP_NET_BIND_SERVICE", + "CAP_SYS_CHROOT", + "CAP_KILL", + "CAP_AUDIT_WRITE", + }, + Ambient: []string{ + "CAP_CHOWN", + "CAP_DAC_OVERRIDE", + "CAP_FSETID", + "CAP_FOWNER", + "CAP_MKNOD", + "CAP_NET_RAW", + "CAP_SETGID", + "CAP_SETUID", + "CAP_SETFCAP", + "CAP_SETPCAP", + "CAP_NET_BIND_SERVICE", + "CAP_SYS_CHROOT", + "CAP_KILL", + "CAP_AUDIT_WRITE", + }, + }, + Namespaces: configs.Namespaces([]configs.Namespace{ + {Type: configs.NEWNS}, + {Type: configs.NEWUTS}, + {Type: configs.NEWIPC}, + {Type: configs.NEWPID}, + {Type: configs.NEWUSER}, + {Type: configs.NEWNET}, + }), + Cgroups: &configs.Cgroup{ + Name: "test-container", + Parent: "system", + Resources: &configs.Resources{ + MemorySwappiness: nil, + AllowAllDevices: nil, + AllowedDevices: configs.DefaultAllowedDevices, + }, + }, + MaskPaths: []string{ + "/proc/kcore", + "/sys/firmware", + }, + ReadonlyPaths: []string{ + "/proc/sys", "/proc/sysrq-trigger", "/proc/irq", "/proc/bus", + }, + Devices: configs.DefaultAutoCreatedDevices, + Hostname: "testing", + Mounts: []*configs.Mount{ + { + Source: "proc", + Destination: "/proc", + Device: "proc", + Flags: defaultMountFlags, + }, + { + Source: "tmpfs", + Destination: "/dev", + Device: "tmpfs", + Flags: unix.MS_NOSUID | unix.MS_STRICTATIME, + Data: "mode=755", + }, + { + Source: "devpts", + Destination: "/dev/pts", + Device: "devpts", + Flags: unix.MS_NOSUID | unix.MS_NOEXEC, + Data: "newinstance,ptmxmode=0666,mode=0620,gid=5", + }, + { + Device: "tmpfs", + Source: "shm", + Destination: "/dev/shm", + Data: "mode=1777,size=65536k", + Flags: defaultMountFlags, + }, + { + Source: "mqueue", + Destination: "/dev/mqueue", + Device: "mqueue", + Flags: defaultMountFlags, + }, + { + Source: "sysfs", + Destination: "/sys", + Device: "sysfs", + Flags: defaultMountFlags | unix.MS_RDONLY, + }, + }, + UidMappings: []configs.IDMap{ + { + ContainerID: 0, + HostID: 1000, + Size: 65536, + }, + }, + GidMappings: []configs.IDMap{ + { + ContainerID: 0, + HostID: 1000, + Size: 65536, + }, + }, + Networks: []*configs.Network{ + { + Type: "loopback", + Address: "127.0.0.1/0", + Gateway: "localhost", + }, + }, + Rlimits: []configs.Rlimit{ + { + Type: unix.RLIMIT_NOFILE, + Hard: uint64(1025), + Soft: uint64(1025), + }, + }, +} +``` + +Once you have the configuration populated you can create a container: + +```go +container, err := factory.Create("container-id", config) +if err != nil { + logrus.Fatal(err) + return +} +``` + +To spawn bash as the initial process inside the container and have the +processes pid returned in order to wait, signal, or kill the process: + +```go +process := &libcontainer.Process{ + Args: []string{"/bin/bash"}, + Env: []string{"PATH=/bin"}, + User: "daemon", + Stdin: os.Stdin, + Stdout: os.Stdout, + Stderr: os.Stderr, +} + +err := container.Run(process) +if err != nil { + container.Destroy() + logrus.Fatal(err) + return +} + +// wait for the process to finish. +_, err := process.Wait() +if err != nil { + logrus.Fatal(err) +} + +// destroy the container. +container.Destroy() +``` + +Additional ways to interact with a running container are: + +```go +// return all the pids for all processes running inside the container. +processes, err := container.Processes() + +// get detailed cpu, memory, io, and network statistics for the container and +// it's processes. +stats, err := container.Stats() + +// pause all processes inside the container. +container.Pause() + +// resume all paused processes. +container.Resume() + +// send signal to container's init process. +container.Signal(signal) + +// update container resource constraints. +container.Set(config) + +// get current status of the container. +status, err := container.Status() + +// get current container's state information. +state, err := container.State() +``` + + +#### Checkpoint & Restore + +libcontainer now integrates [CRIU](http://criu.org/) for checkpointing and restoring containers. +This let's you save the state of a process running inside a container to disk, and then restore +that state into a new process, on the same machine or on another machine. + +`criu` version 1.5.2 or higher is required to use checkpoint and restore. +If you don't already have `criu` installed, you can build it from source, following the +[online instructions](http://criu.org/Installation). `criu` is also installed in the docker image +generated when building libcontainer with docker. + + +## Copyright and license + +Code and documentation copyright 2014 Docker, inc. Code released under the Apache 2.0 license. +Docs released under Creative commons. diff --git a/vendor/github.com/opencontainers/runc/libcontainer/SPEC.md b/vendor/github.com/opencontainers/runc/libcontainer/SPEC.md new file mode 100644 index 000000000..5d68bb14c --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/SPEC.md @@ -0,0 +1,418 @@ +## Container Specification - v1 + +This is the standard configuration for version 1 containers. It includes +namespaces, standard filesystem setup, a default Linux capability set, and +information about resource reservations. It also has information about any +populated environment settings for the processes running inside a container. + +Along with the configuration of how a container is created the standard also +discusses actions that can be performed on a container to manage and inspect +information about the processes running inside. + +The v1 profile is meant to be able to accommodate the majority of applications +with a strong security configuration. + +### System Requirements and Compatibility + +Minimum requirements: +* Kernel version - 3.10 recommended 2.6.2x minimum(with backported patches) +* Mounted cgroups with each subsystem in its own hierarchy + + +### Namespaces + +| Flag | Enabled | +| ------------ | ------- | +| CLONE_NEWPID | 1 | +| CLONE_NEWUTS | 1 | +| CLONE_NEWIPC | 1 | +| CLONE_NEWNET | 1 | +| CLONE_NEWNS | 1 | +| CLONE_NEWUSER | 1 | + +Namespaces are created for the container via the `clone` syscall. + + +### Filesystem + +A root filesystem must be provided to a container for execution. The container +will use this root filesystem (rootfs) to jail and spawn processes inside where +the binaries and system libraries are local to that directory. Any binaries +to be executed must be contained within this rootfs. + +Mounts that happen inside the container are automatically cleaned up when the +container exits as the mount namespace is destroyed and the kernel will +unmount all the mounts that were setup within that namespace. + +For a container to execute properly there are certain filesystems that +are required to be mounted within the rootfs that the runtime will setup. + +| Path | Type | Flags | Data | +| ----------- | ------ | -------------------------------------- | ---------------------------------------- | +| /proc | proc | MS_NOEXEC,MS_NOSUID,MS_NODEV | | +| /dev | tmpfs | MS_NOEXEC,MS_STRICTATIME | mode=755 | +| /dev/shm | tmpfs | MS_NOEXEC,MS_NOSUID,MS_NODEV | mode=1777,size=65536k | +| /dev/mqueue | mqueue | MS_NOEXEC,MS_NOSUID,MS_NODEV | | +| /dev/pts | devpts | MS_NOEXEC,MS_NOSUID | newinstance,ptmxmode=0666,mode=620,gid=5 | +| /sys | sysfs | MS_NOEXEC,MS_NOSUID,MS_NODEV,MS_RDONLY | | + + +After a container's filesystems are mounted within the newly created +mount namespace `/dev` will need to be populated with a set of device nodes. +It is expected that a rootfs does not need to have any device nodes specified +for `/dev` within the rootfs as the container will setup the correct devices +that are required for executing a container's process. + +| Path | Mode | Access | +| ------------ | ---- | ---------- | +| /dev/null | 0666 | rwm | +| /dev/zero | 0666 | rwm | +| /dev/full | 0666 | rwm | +| /dev/tty | 0666 | rwm | +| /dev/random | 0666 | rwm | +| /dev/urandom | 0666 | rwm | + + +**ptmx** +`/dev/ptmx` will need to be a symlink to the host's `/dev/ptmx` within +the container. + +The use of a pseudo TTY is optional within a container and it should support both. +If a pseudo is provided to the container `/dev/console` will need to be +setup by binding the console in `/dev/` after it has been populated and mounted +in tmpfs. + +| Source | Destination | UID GID | Mode | Type | +| --------------- | ------------ | ------- | ---- | ---- | +| *pty host path* | /dev/console | 0 0 | 0600 | bind | + + +After `/dev/null` has been setup we check for any external links between +the container's io, STDIN, STDOUT, STDERR. If the container's io is pointing +to `/dev/null` outside the container we close and `dup2` the `/dev/null` +that is local to the container's rootfs. + + +After the container has `/proc` mounted a few standard symlinks are setup +within `/dev/` for the io. + +| Source | Destination | +| --------------- | ----------- | +| /proc/self/fd | /dev/fd | +| /proc/self/fd/0 | /dev/stdin | +| /proc/self/fd/1 | /dev/stdout | +| /proc/self/fd/2 | /dev/stderr | + +A `pivot_root` is used to change the root for the process, effectively +jailing the process inside the rootfs. + +```c +put_old = mkdir(...); +pivot_root(rootfs, put_old); +chdir("/"); +unmount(put_old, MS_DETACH); +rmdir(put_old); +``` + +For container's running with a rootfs inside `ramfs` a `MS_MOVE` combined +with a `chroot` is required as `pivot_root` is not supported in `ramfs`. + +```c +mount(rootfs, "/", NULL, MS_MOVE, NULL); +chroot("."); +chdir("/"); +``` + +The `umask` is set back to `0022` after the filesystem setup has been completed. + +### Resources + +Cgroups are used to handle resource allocation for containers. This includes +system resources like cpu, memory, and device access. + +| Subsystem | Enabled | +| ---------- | ------- | +| devices | 1 | +| memory | 1 | +| cpu | 1 | +| cpuacct | 1 | +| cpuset | 1 | +| blkio | 1 | +| perf_event | 1 | +| freezer | 1 | +| hugetlb | 1 | +| pids | 1 | + + +All cgroup subsystem are joined so that statistics can be collected from +each of the subsystems. Freezer does not expose any stats but is joined +so that containers can be paused and resumed. + +The parent process of the container's init must place the init pid inside +the correct cgroups before the initialization begins. This is done so +that no processes or threads escape the cgroups. This sync is +done via a pipe ( specified in the runtime section below ) that the container's +init process will block waiting for the parent to finish setup. + +### IntelRdt + +Intel platforms with new Xeon CPU support Intel Resource Director Technology +(RDT). Cache Allocation Technology (CAT) is a sub-feature of RDT, which +currently supports L3 cache resource allocation. + +This feature provides a way for the software to restrict cache allocation to a +defined 'subset' of L3 cache which may be overlapping with other 'subsets'. +The different subsets are identified by class of service (CLOS) and each CLOS +has a capacity bitmask (CBM). + +It can be used to handle L3 cache resource allocation for containers if +hardware and kernel support Intel RDT/CAT. + +In Linux 4.10 kernel or newer, the interface is defined and exposed via +"resource control" filesystem, which is a "cgroup-like" interface. + +Comparing with cgroups, it has similar process management lifecycle and +interfaces in a container. But unlike cgroups' hierarchy, it has single level +filesystem layout. + +Intel RDT "resource control" filesystem hierarchy: +``` +mount -t resctrl resctrl /sys/fs/resctrl +tree /sys/fs/resctrl +/sys/fs/resctrl/ +|-- info +| |-- L3 +| |-- cbm_mask +| |-- min_cbm_bits +| |-- num_closids +|-- cpus +|-- schemata +|-- tasks +|-- + |-- cpus + |-- schemata + |-- tasks + +``` + +For runc, we can make use of `tasks` and `schemata` configuration for L3 cache +resource constraints. + +The file `tasks` has a list of tasks that belongs to this group (e.g., +" group). Tasks can be added to a group by writing the task ID +to the "tasks" file (which will automatically remove them from the previous +group to which they belonged). New tasks created by fork(2) and clone(2) are +added to the same group as their parent. If a pid is not in any sub group, it +is in root group. + +The file `schemata` has allocation masks/values for L3 cache on each socket, +which contains L3 cache id and capacity bitmask (CBM). +``` + Format: "L3:=;=;..." +``` +For example, on a two-socket machine, L3's schema line could be `L3:0=ff;1=c0` +Which means L3 cache id 0's CBM is 0xff, and L3 cache id 1's CBM is 0xc0. + +The valid L3 cache CBM is a *contiguous bits set* and number of bits that can +be set is less than the max bit. The max bits in the CBM is varied among +supported Intel Xeon platforms. In Intel RDT "resource control" filesystem +layout, the CBM in a group should be a subset of the CBM in root. Kernel will +check if it is valid when writing. e.g., 0xfffff in root indicates the max bits +of CBM is 20 bits, which mapping to entire L3 cache capacity. Some valid CBM +values to set in a group: 0xf, 0xf0, 0x3ff, 0x1f00 and etc. + +For more information about Intel RDT/CAT kernel interface: +https://www.kernel.org/doc/Documentation/x86/intel_rdt_ui.txt + +An example for runc: +``` +Consider a two-socket machine with two L3 caches where the default CBM is +0xfffff and the max CBM length is 20 bits. With this configuration, tasks +inside the container only have access to the "upper" 80% of L3 cache id 0 and +the "lower" 50% L3 cache id 1: + +"linux": { + "intelRdt": { + "l3CacheSchema": "L3:0=ffff0;1=3ff" + } +} +``` + +### Security + +The standard set of Linux capabilities that are set in a container +provide a good default for security and flexibility for the applications. + + +| Capability | Enabled | +| -------------------- | ------- | +| CAP_NET_RAW | 1 | +| CAP_NET_BIND_SERVICE | 1 | +| CAP_AUDIT_READ | 1 | +| CAP_AUDIT_WRITE | 1 | +| CAP_DAC_OVERRIDE | 1 | +| CAP_SETFCAP | 1 | +| CAP_SETPCAP | 1 | +| CAP_SETGID | 1 | +| CAP_SETUID | 1 | +| CAP_MKNOD | 1 | +| CAP_CHOWN | 1 | +| CAP_FOWNER | 1 | +| CAP_FSETID | 1 | +| CAP_KILL | 1 | +| CAP_SYS_CHROOT | 1 | +| CAP_NET_BROADCAST | 0 | +| CAP_SYS_MODULE | 0 | +| CAP_SYS_RAWIO | 0 | +| CAP_SYS_PACCT | 0 | +| CAP_SYS_ADMIN | 0 | +| CAP_SYS_NICE | 0 | +| CAP_SYS_RESOURCE | 0 | +| CAP_SYS_TIME | 0 | +| CAP_SYS_TTY_CONFIG | 0 | +| CAP_AUDIT_CONTROL | 0 | +| CAP_MAC_OVERRIDE | 0 | +| CAP_MAC_ADMIN | 0 | +| CAP_NET_ADMIN | 0 | +| CAP_SYSLOG | 0 | +| CAP_DAC_READ_SEARCH | 0 | +| CAP_LINUX_IMMUTABLE | 0 | +| CAP_IPC_LOCK | 0 | +| CAP_IPC_OWNER | 0 | +| CAP_SYS_PTRACE | 0 | +| CAP_SYS_BOOT | 0 | +| CAP_LEASE | 0 | +| CAP_WAKE_ALARM | 0 | +| CAP_BLOCK_SUSPEND | 0 | + + +Additional security layers like [apparmor](https://wiki.ubuntu.com/AppArmor) +and [selinux](http://selinuxproject.org/page/Main_Page) can be used with +the containers. A container should support setting an apparmor profile or +selinux process and mount labels if provided in the configuration. + +Standard apparmor profile: +```c +#include +profile flags=(attach_disconnected,mediate_deleted) { + #include + network, + capability, + file, + umount, + + deny @{PROC}/sys/fs/** wklx, + deny @{PROC}/sysrq-trigger rwklx, + deny @{PROC}/mem rwklx, + deny @{PROC}/kmem rwklx, + deny @{PROC}/sys/kernel/[^s][^h][^m]* wklx, + deny @{PROC}/sys/kernel/*/** wklx, + + deny mount, + + deny /sys/[^f]*/** wklx, + deny /sys/f[^s]*/** wklx, + deny /sys/fs/[^c]*/** wklx, + deny /sys/fs/c[^g]*/** wklx, + deny /sys/fs/cg[^r]*/** wklx, + deny /sys/firmware/efi/efivars/** rwklx, + deny /sys/kernel/security/** rwklx, +} +``` + +*TODO: seccomp work is being done to find a good default config* + +### Runtime and Init Process + +During container creation the parent process needs to talk to the container's init +process and have a form of synchronization. This is accomplished by creating +a pipe that is passed to the container's init. When the init process first spawns +it will block on its side of the pipe until the parent closes its side. This +allows the parent to have time to set the new process inside a cgroup hierarchy +and/or write any uid/gid mappings required for user namespaces. +The pipe is passed to the init process via FD 3. + +The application consuming libcontainer should be compiled statically. libcontainer +does not define any init process and the arguments provided are used to `exec` the +process inside the application. There should be no long running init within the +container spec. + +If a pseudo tty is provided to a container it will open and `dup2` the console +as the container's STDIN, STDOUT, STDERR as well as mounting the console +as `/dev/console`. + +An extra set of mounts are provided to a container and setup for use. A container's +rootfs can contain some non portable files inside that can cause side effects during +execution of a process. These files are usually created and populated with the container +specific information via the runtime. + +**Extra runtime files:** +* /etc/hosts +* /etc/resolv.conf +* /etc/hostname +* /etc/localtime + + +#### Defaults + +There are a few defaults that can be overridden by users, but in their omission +these apply to processes within a container. + +| Type | Value | +| ------------------- | ------------------------------ | +| Parent Death Signal | SIGKILL | +| UID | 0 | +| GID | 0 | +| GROUPS | 0, NULL | +| CWD | "/" | +| $HOME | Current user's home dir or "/" | +| Readonly rootfs | false | +| Pseudo TTY | false | + + +## Actions + +After a container is created there is a standard set of actions that can +be done to the container. These actions are part of the public API for +a container. + +| Action | Description | +| -------------- | ------------------------------------------------------------------ | +| Get processes | Return all the pids for processes running inside a container | +| Get Stats | Return resource statistics for the container as a whole | +| Wait | Waits on the container's init process ( pid 1 ) | +| Wait Process | Wait on any of the container's processes returning the exit status | +| Destroy | Kill the container's init process and remove any filesystem state | +| Signal | Send a signal to the container's init process | +| Signal Process | Send a signal to any of the container's processes | +| Pause | Pause all processes inside the container | +| Resume | Resume all processes inside the container if paused | +| Exec | Execute a new process inside of the container ( requires setns ) | +| Set | Setup configs of the container after it's created | + +### Execute a new process inside of a running container + +User can execute a new process inside of a running container. Any binaries to be +executed must be accessible within the container's rootfs. + +The started process will run inside the container's rootfs. Any changes +made by the process to the container's filesystem will persist after the +process finished executing. + +The started process will join all the container's existing namespaces. When the +container is paused, the process will also be paused and will resume when +the container is unpaused. The started process will only run when the container's +primary process (PID 1) is running, and will not be restarted when the container +is restarted. + +#### Planned additions + +The started process will have its own cgroups nested inside the container's +cgroups. This is used for process tracking and optionally resource allocation +handling for the new process. Freezer cgroup is required, the rest of the cgroups +are optional. The process executor must place its pid inside the correct +cgroups before starting the process. This is done so that no child processes or +threads can escape the cgroups. + +When the process is stopped, the process executor will try (in a best-effort way) +to stop all its children and remove the sub-cgroups. diff --git a/vendor/github.com/opencontainers/runc/libcontainer/apparmor/apparmor.go b/vendor/github.com/opencontainers/runc/libcontainer/apparmor/apparmor.go new file mode 100644 index 000000000..7fff0627f --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/apparmor/apparmor.go @@ -0,0 +1,54 @@ +// +build apparmor,linux + +package apparmor + +import ( + "fmt" + "io/ioutil" + "os" +) + +// IsEnabled returns true if apparmor is enabled for the host. +func IsEnabled() bool { + if _, err := os.Stat("/sys/kernel/security/apparmor"); err == nil && os.Getenv("container") == "" { + if _, err = os.Stat("/sbin/apparmor_parser"); err == nil { + buf, err := ioutil.ReadFile("/sys/module/apparmor/parameters/enabled") + return err == nil && len(buf) > 1 && buf[0] == 'Y' + } + } + return false +} + +func setprocattr(attr, value string) error { + // Under AppArmor you can only change your own attr, so use /proc/self/ + // instead of /proc// like libapparmor does + path := fmt.Sprintf("/proc/self/attr/%s", attr) + + f, err := os.OpenFile(path, os.O_WRONLY, 0) + if err != nil { + return err + } + defer f.Close() + + _, err = fmt.Fprintf(f, "%s", value) + return err +} + +// changeOnExec reimplements aa_change_onexec from libapparmor in Go +func changeOnExec(name string) error { + value := "exec " + name + if err := setprocattr("exec", value); err != nil { + return fmt.Errorf("apparmor failed to apply profile: %s", err) + } + return nil +} + +// ApplyProfile will apply the profile with the specified name to the process after +// the next exec. +func ApplyProfile(name string) error { + if name == "" { + return nil + } + + return changeOnExec(name) +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/apparmor/apparmor_disabled.go b/vendor/github.com/opencontainers/runc/libcontainer/apparmor/apparmor_disabled.go new file mode 100644 index 000000000..d4110cf0b --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/apparmor/apparmor_disabled.go @@ -0,0 +1,20 @@ +// +build !apparmor !linux + +package apparmor + +import ( + "errors" +) + +var ErrApparmorNotEnabled = errors.New("apparmor: config provided but apparmor not supported") + +func IsEnabled() bool { + return false +} + +func ApplyProfile(name string) error { + if name != "" { + return ErrApparmorNotEnabled + } + return nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/capabilities_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/capabilities_linux.go new file mode 100644 index 000000000..7c66f5725 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/capabilities_linux.go @@ -0,0 +1,113 @@ +// +build linux + +package libcontainer + +import ( + "fmt" + "strings" + + "github.com/opencontainers/runc/libcontainer/configs" + "github.com/syndtr/gocapability/capability" +) + +const allCapabilityTypes = capability.CAPS | capability.BOUNDS | capability.AMBS + +var capabilityMap map[string]capability.Cap + +func init() { + capabilityMap = make(map[string]capability.Cap) + last := capability.CAP_LAST_CAP + // workaround for RHEL6 which has no /proc/sys/kernel/cap_last_cap + if last == capability.Cap(63) { + last = capability.CAP_BLOCK_SUSPEND + } + for _, cap := range capability.List() { + if cap > last { + continue + } + capKey := fmt.Sprintf("CAP_%s", strings.ToUpper(cap.String())) + capabilityMap[capKey] = cap + } +} + +func newContainerCapList(capConfig *configs.Capabilities) (*containerCapabilities, error) { + bounding := []capability.Cap{} + for _, c := range capConfig.Bounding { + v, ok := capabilityMap[c] + if !ok { + return nil, fmt.Errorf("unknown capability %q", c) + } + bounding = append(bounding, v) + } + effective := []capability.Cap{} + for _, c := range capConfig.Effective { + v, ok := capabilityMap[c] + if !ok { + return nil, fmt.Errorf("unknown capability %q", c) + } + effective = append(effective, v) + } + inheritable := []capability.Cap{} + for _, c := range capConfig.Inheritable { + v, ok := capabilityMap[c] + if !ok { + return nil, fmt.Errorf("unknown capability %q", c) + } + inheritable = append(inheritable, v) + } + permitted := []capability.Cap{} + for _, c := range capConfig.Permitted { + v, ok := capabilityMap[c] + if !ok { + return nil, fmt.Errorf("unknown capability %q", c) + } + permitted = append(permitted, v) + } + ambient := []capability.Cap{} + for _, c := range capConfig.Ambient { + v, ok := capabilityMap[c] + if !ok { + return nil, fmt.Errorf("unknown capability %q", c) + } + ambient = append(ambient, v) + } + pid, err := capability.NewPid(0) + if err != nil { + return nil, err + } + return &containerCapabilities{ + bounding: bounding, + effective: effective, + inheritable: inheritable, + permitted: permitted, + ambient: ambient, + pid: pid, + }, nil +} + +type containerCapabilities struct { + pid capability.Capabilities + bounding []capability.Cap + effective []capability.Cap + inheritable []capability.Cap + permitted []capability.Cap + ambient []capability.Cap +} + +// ApplyBoundingSet sets the capability bounding set to those specified in the whitelist. +func (c *containerCapabilities) ApplyBoundingSet() error { + c.pid.Clear(capability.BOUNDS) + c.pid.Set(capability.BOUNDS, c.bounding...) + return c.pid.Apply(capability.BOUNDS) +} + +// Apply sets all the capabilities for the current process in the config. +func (c *containerCapabilities) ApplyCaps() error { + c.pid.Clear(allCapabilityTypes) + c.pid.Set(capability.BOUNDS, c.bounding...) + c.pid.Set(capability.PERMITTED, c.permitted...) + c.pid.Set(capability.INHERITABLE, c.inheritable...) + c.pid.Set(capability.EFFECTIVE, c.effective...) + c.pid.Set(capability.AMBIENT, c.ambient...) + return c.pid.Apply(allCapabilityTypes) +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/cgroups.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/cgroups.go index 274ab47dd..25ff51589 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/cgroups.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/cgroups.go @@ -27,9 +27,9 @@ type Manager interface { // Destroys the cgroup set Destroy() error - // NewCgroupManager() and LoadCgroupManager() require following attributes: + // The option func SystemdCgroups() and Cgroupfs() require following attributes: // Paths map[string]string - // Cgroups *cgroups.Cgroup + // Cgroups *configs.Cgroup // Paths maps cgroup subsystem to path at which it is mounted. // Cgroups specifies specific cgroup settings for the various subsystems @@ -37,7 +37,7 @@ type Manager interface { // restore the object later. GetPaths() map[string]string - // Set the cgroup as configured. + // Sets the cgroup as configured. Set(container *configs.Config) error } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/apply_raw.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/apply_raw.go index 114f002ec..09b96de5d 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/apply_raw.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/apply_raw.go @@ -3,18 +3,18 @@ package fs import ( - "errors" "fmt" "io" "io/ioutil" "os" "path/filepath" - "strconv" "sync" "github.com/opencontainers/runc/libcontainer/cgroups" "github.com/opencontainers/runc/libcontainer/configs" libcontainerUtils "github.com/opencontainers/runc/libcontainer/utils" + "github.com/pkg/errors" + "golang.org/x/sys/unix" ) var ( @@ -33,11 +33,10 @@ var ( &FreezerGroup{}, &NameGroup{GroupName: "name=systemd", Join: true}, } - CgroupProcesses = "cgroup.procs" HugePageSizes, _ = cgroups.GetHugePageSize() ) -var errSubsystemDoesNotExist = errors.New("cgroup: subsystem does not exist") +var errSubsystemDoesNotExist = fmt.Errorf("cgroup: subsystem does not exist") type subsystemSet []subsystem @@ -64,9 +63,10 @@ type subsystem interface { } type Manager struct { - mu sync.Mutex - Cgroups *configs.Cgroup - Paths map[string]string + mu sync.Mutex + Cgroups *configs.Cgroup + Rootless bool + Paths map[string]string } // The absolute path to the root of the cgroup hierarchies. @@ -102,10 +102,39 @@ type cgroupData struct { pid int } +// isIgnorableError returns whether err is a permission error (in the loose +// sense of the word). This includes EROFS (which for an unprivileged user is +// basically a permission error) and EACCES (for similar reasons) as well as +// the normal EPERM. +func isIgnorableError(rootless bool, err error) bool { + // We do not ignore errors if we are root. + if !rootless { + return false + } + // Is it an ordinary EPERM? + if os.IsPermission(errors.Cause(err)) { + return true + } + + // Try to handle other errnos. + var errno error + switch err := errors.Cause(err).(type) { + case *os.PathError: + errno = err.Err + case *os.LinkError: + errno = err.Err + case *os.SyscallError: + errno = err.Err + } + return errno == unix.EROFS || errno == unix.EPERM || errno == unix.EACCES +} + func (m *Manager) Apply(pid int) (err error) { if m.Cgroups == nil { return nil } + m.mu.Lock() + defer m.mu.Unlock() var c = m.Cgroups @@ -114,8 +143,8 @@ func (m *Manager) Apply(pid int) (err error) { return err } + m.Paths = make(map[string]string) if c.Paths != nil { - paths := make(map[string]string) for name, path := range c.Paths { _, err := d.path(name) if err != nil { @@ -124,37 +153,44 @@ func (m *Manager) Apply(pid int) (err error) { } return err } - paths[name] = path + m.Paths[name] = path } - m.Paths = paths return cgroups.EnterPid(m.Paths, pid) } - m.mu.Lock() - defer m.mu.Unlock() - paths := make(map[string]string) for _, sys := range subsystems { - if err := sys.Apply(d); err != nil { - return err - } // TODO: Apply should, ideally, be reentrant or be broken up into a separate // create and join phase so that the cgroup hierarchy for a container can be // created then join consists of writing the process pids to cgroup.procs p, err := d.path(sys.Name()) if err != nil { - if cgroups.IsNotFound(err) { + // The non-presence of the devices subsystem is + // considered fatal for security reasons. + if cgroups.IsNotFound(err) && sys.Name() != "devices" { continue } return err } - paths[sys.Name()] = p + m.Paths[sys.Name()] = p + + if err := sys.Apply(d); err != nil { + // In the case of rootless, where an explicit cgroup path hasn't + // been set, we don't bail on error in case of permission problems. + // Cases where limits have been set (and we couldn't create our own + // cgroup) are handled by Set. + if isIgnorableError(m.Rootless, err) && m.Cgroups.Path == "" { + delete(m.Paths, sys.Name()) + continue + } + return err + } + } - m.Paths = paths return nil } func (m *Manager) Destroy() error { - if m.Cgroups.Paths != nil { + if m.Cgroups == nil || m.Cgroups.Paths != nil { return nil } m.mu.Lock() @@ -190,19 +226,21 @@ func (m *Manager) GetStats() (*cgroups.Stats, error) { } func (m *Manager) Set(container *configs.Config) error { - for _, sys := range subsystems { - // Generate fake cgroup data. - d, err := getCgroupData(container.Cgroups, -1) - if err != nil { - return err - } - // Get the path, but don't error out if the cgroup wasn't found. - path, err := d.path(sys.Name()) - if err != nil && !cgroups.IsNotFound(err) { - return err - } + // If Paths are set, then we are just joining cgroups paths + // and there is no need to set any values. + if m.Cgroups.Paths != nil { + return nil + } + paths := m.GetPaths() + for _, sys := range subsystems { + path := paths[sys.Name()] if err := sys.Set(path, container.Cgroups); err != nil { + if path == "" { + // We never created a path for this cgroup, so we cannot set + // limits for it (though we have already tried at this point). + return fmt.Errorf("cannot set %s limit: container could not join or create cgroup", sys.Name()) + } return err } } @@ -218,14 +256,8 @@ func (m *Manager) Set(container *configs.Config) error { // Freeze toggles the container's freezer cgroup depending on the state // provided func (m *Manager) Freeze(state configs.FreezerState) error { - d, err := getCgroupData(m.Cgroups, 0) - if err != nil { - return err - } - dir, err := d.path("freezer") - if err != nil { - return err - } + paths := m.GetPaths() + dir := paths["freezer"] prevState := m.Cgroups.Resources.Freezer m.Cgroups.Resources.Freezer = state freezer, err := subsystems.Get("freezer") @@ -241,28 +273,13 @@ func (m *Manager) Freeze(state configs.FreezerState) error { } func (m *Manager) GetPids() ([]int, error) { - dir, err := getCgroupPath(m.Cgroups) - if err != nil { - return nil, err - } - return cgroups.GetPids(dir) + paths := m.GetPaths() + return cgroups.GetPids(paths["devices"]) } func (m *Manager) GetAllPids() ([]int, error) { - dir, err := getCgroupPath(m.Cgroups) - if err != nil { - return nil, err - } - return cgroups.GetAllPids(dir) -} - -func getCgroupPath(c *configs.Cgroup) (string, error) { - d, err := getCgroupData(c, 0) - if err != nil { - return "", err - } - - return d.path("devices") + paths := m.GetPaths() + return cgroups.GetAllPids(paths["devices"]) } func getCgroupData(c *configs.Cgroup, pid int) (*cgroupData, error) { @@ -293,25 +310,8 @@ func getCgroupData(c *configs.Cgroup, pid int) (*cgroupData, error) { }, nil } -func (raw *cgroupData) parentPath(subsystem, mountpoint, root string) (string, error) { - // Use GetThisCgroupDir instead of GetInitCgroupDir, because the creating - // process could in container and shared pid namespace with host, and - // /proc/1/cgroup could point to whole other world of cgroups. - initPath, err := cgroups.GetThisCgroupDir(subsystem) - if err != nil { - return "", err - } - // This is needed for nested containers, because in /proc/self/cgroup we - // see pathes from host, which don't exist in container. - relDir, err := filepath.Rel(root, initPath) - if err != nil { - return "", err - } - return filepath.Join(mountpoint, relDir), nil -} - func (raw *cgroupData) path(subsystem string) (string, error) { - mnt, root, err := cgroups.FindCgroupMountpointAndRoot(subsystem) + mnt, err := cgroups.FindCgroupMountpoint(subsystem) // If we didn't mount the subsystem, there is no point we make the path. if err != nil { return "", err @@ -319,11 +319,14 @@ func (raw *cgroupData) path(subsystem string) (string, error) { // If the cgroup name/path is absolute do not look relative to the cgroup of the init process. if filepath.IsAbs(raw.innerPath) { - // Sometimes subsystems can be mounted togethger as 'cpu,cpuacct'. + // Sometimes subsystems can be mounted together as 'cpu,cpuacct'. return filepath.Join(raw.root, filepath.Base(mnt), raw.innerPath), nil } - parentPath, err := raw.parentPath(subsystem, mnt, root) + // Use GetOwnCgroupPath instead of GetInitCgroupPath, because the creating + // process could in container and shared pid namespace with host, and + // /proc/1/cgroup could point to whole other world of cgroups. + parentPath, err := cgroups.GetOwnCgroupPath(subsystem) if err != nil { return "", err } @@ -339,7 +342,7 @@ func (raw *cgroupData) join(subsystem string) (string, error) { if err := os.MkdirAll(path, 0755); err != nil { return "", err } - if err := writeFile(path, CgroupProcesses, strconv.Itoa(raw.pid)); err != nil { + if err := cgroups.WriteCgroupProc(path, raw.pid); err != nil { return "", err } return path, nil @@ -349,7 +352,7 @@ func writeFile(dir, file, data string) error { // Normally dir should not be empty, one case is that cgroup subsystem // is not mounted, we will get empty dir, and we want it fail here. if dir == "" { - return fmt.Errorf("no such directory for %s.", file) + return fmt.Errorf("no such directory for %s", file) } if err := ioutil.WriteFile(filepath.Join(dir, file), []byte(data), 0700); err != nil { return fmt.Errorf("failed to write %v to %v: %v", data, file, err) @@ -372,8 +375,8 @@ func removePath(p string, err error) error { return nil } -func CheckCpushares(path string, c int64) error { - var cpuShares int64 +func CheckCpushares(path string, c uint64) error { + var cpuShares uint64 if c == 0 { return nil diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/cpu.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/cpu.go index a4ef28a60..b712bd0b1 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/cpu.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/cpu.go @@ -22,21 +22,59 @@ func (s *CpuGroup) Name() string { func (s *CpuGroup) Apply(d *cgroupData) error { // We always want to join the cpu group, to allow fair cpu scheduling // on a container basis - _, err := d.join("cpu") + path, err := d.path("cpu") if err != nil && !cgroups.IsNotFound(err) { return err } + return s.ApplyDir(path, d.config, d.pid) +} + +func (s *CpuGroup) ApplyDir(path string, cgroup *configs.Cgroup, pid int) error { + // This might happen if we have no cpu cgroup mounted. + // Just do nothing and don't fail. + if path == "" { + return nil + } + if err := os.MkdirAll(path, 0755); err != nil { + return err + } + // We should set the real-Time group scheduling settings before moving + // in the process because if the process is already in SCHED_RR mode + // and no RT bandwidth is set, adding it will fail. + if err := s.SetRtSched(path, cgroup); err != nil { + return err + } + // because we are not using d.join we need to place the pid into the procs file + // unlike the other subsystems + if err := cgroups.WriteCgroupProc(path, pid); err != nil { + return err + } + + return nil +} + +func (s *CpuGroup) SetRtSched(path string, cgroup *configs.Cgroup) error { + if cgroup.Resources.CpuRtPeriod != 0 { + if err := writeFile(path, "cpu.rt_period_us", strconv.FormatUint(cgroup.Resources.CpuRtPeriod, 10)); err != nil { + return err + } + } + if cgroup.Resources.CpuRtRuntime != 0 { + if err := writeFile(path, "cpu.rt_runtime_us", strconv.FormatInt(cgroup.Resources.CpuRtRuntime, 10)); err != nil { + return err + } + } return nil } func (s *CpuGroup) Set(path string, cgroup *configs.Cgroup) error { if cgroup.Resources.CpuShares != 0 { - if err := writeFile(path, "cpu.shares", strconv.FormatInt(cgroup.Resources.CpuShares, 10)); err != nil { + if err := writeFile(path, "cpu.shares", strconv.FormatUint(cgroup.Resources.CpuShares, 10)); err != nil { return err } } if cgroup.Resources.CpuPeriod != 0 { - if err := writeFile(path, "cpu.cfs_period_us", strconv.FormatInt(cgroup.Resources.CpuPeriod, 10)); err != nil { + if err := writeFile(path, "cpu.cfs_period_us", strconv.FormatUint(cgroup.Resources.CpuPeriod, 10)); err != nil { return err } } @@ -45,15 +83,8 @@ func (s *CpuGroup) Set(path string, cgroup *configs.Cgroup) error { return err } } - if cgroup.Resources.CpuRtPeriod != 0 { - if err := writeFile(path, "cpu.rt_period_us", strconv.FormatInt(cgroup.Resources.CpuRtPeriod, 10)); err != nil { - return err - } - } - if cgroup.Resources.CpuRtRuntime != 0 { - if err := writeFile(path, "cpu.rt_runtime_us", strconv.FormatInt(cgroup.Resources.CpuRtRuntime, 10)); err != nil { - return err - } + if err := s.SetRtSched(path, cgroup); err != nil { + return err } return nil diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/cpuset.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/cpuset.go index cbe62bd98..20c9eafac 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/cpuset.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/cpuset.go @@ -8,7 +8,6 @@ import ( "io/ioutil" "os" "path/filepath" - "strconv" "github.com/opencontainers/runc/libcontainer/cgroups" "github.com/opencontainers/runc/libcontainer/configs" @@ -58,16 +57,34 @@ func (s *CpusetGroup) ApplyDir(dir string, cgroup *configs.Cgroup, pid int) erro if dir == "" { return nil } - root, err := getCgroupRoot() + mountInfo, err := ioutil.ReadFile("/proc/self/mountinfo") if err != nil { return err } - if err := s.ensureParent(dir, root); err != nil { + root := filepath.Dir(cgroups.GetClosestMountpointAncestor(dir, string(mountInfo))) + // 'ensureParent' start with parent because we don't want to + // explicitly inherit from parent, it could conflict with + // 'cpuset.cpu_exclusive'. + if err := s.ensureParent(filepath.Dir(dir), root); err != nil { return err } + if err := os.MkdirAll(dir, 0755); err != nil { + return err + } + // We didn't inherit cpuset configs from parent, but we have + // to ensure cpuset configs are set before moving task into the + // cgroup. + // The logic is, if user specified cpuset configs, use these + // specified configs, otherwise, inherit from parent. This makes + // cpuset configs work correctly with 'cpuset.cpu_exclusive', and + // keep backward compatbility. + if err := s.ensureCpusAndMems(dir, cgroup); err != nil { + return err + } + // because we are not using d.join we need to place the pid into the procs file // unlike the other subsystems - if err := writeFile(dir, "cgroup.procs", strconv.Itoa(pid)); err != nil { + if err := cgroups.WriteCgroupProc(dir, pid); err != nil { return err } @@ -137,3 +154,10 @@ func (s *CpusetGroup) copyIfNeeded(current, parent string) error { func (s *CpusetGroup) isEmpty(b []byte) bool { return len(bytes.Trim(b, "\n")) == 0 } + +func (s *CpusetGroup) ensureCpusAndMems(path string, cgroup *configs.Cgroup) error { + if err := s.Set(path, cgroup); err != nil { + return err + } + return s.copyIfNeeded(path, filepath.Dir(path)) +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/devices.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/devices.go index 5f7833109..0ac5b4ed7 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/devices.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/devices.go @@ -43,21 +43,23 @@ func (s *DevicesGroup) Set(path string, cgroup *configs.Cgroup) error { } return nil } - if !cgroup.Resources.AllowAllDevices { - if err := writeFile(path, "devices.deny", "a"); err != nil { - return err - } - - for _, dev := range cgroup.Resources.AllowedDevices { - if err := writeFile(path, "devices.allow", dev.CgroupString()); err != nil { + if cgroup.Resources.AllowAllDevices != nil { + if *cgroup.Resources.AllowAllDevices == false { + if err := writeFile(path, "devices.deny", "a"); err != nil { return err } - } - return nil - } - if err := writeFile(path, "devices.allow", "a"); err != nil { - return err + for _, dev := range cgroup.Resources.AllowedDevices { + if err := writeFile(path, "devices.allow", dev.CgroupString()); err != nil { + return err + } + } + return nil + } + + if err := writeFile(path, "devices.allow", "a"); err != nil { + return err + } } for _, dev := range cgroup.Resources.DeniedDevices { diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/freezer.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/freezer.go index e70dfe3b9..4b19f8a97 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/freezer.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/freezer.go @@ -29,11 +29,15 @@ func (s *FreezerGroup) Apply(d *cgroupData) error { func (s *FreezerGroup) Set(path string, cgroup *configs.Cgroup) error { switch cgroup.Resources.Freezer { case configs.Frozen, configs.Thawed: - if err := writeFile(path, "freezer.state", string(cgroup.Resources.Freezer)); err != nil { - return err - } - for { + // In case this loop does not exit because it doesn't get the expected + // state, let's write again this state, hoping it's going to be properly + // set this time. Otherwise, this loop could run infinitely, waiting for + // a state change that would never happen. + if err := writeFile(path, "freezer.state", string(cgroup.Resources.Freezer)); err != nil { + return err + } + state, err := readFile(path, "freezer.state") if err != nil { return err @@ -41,6 +45,7 @@ func (s *FreezerGroup) Set(path string, cgroup *configs.Cgroup) error { if strings.TrimSpace(state) == string(cgroup.Resources.Freezer) { break } + time.Sleep(1 * time.Millisecond) } case configs.Undefined: diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/memory.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/memory.go index 3b8ff21ab..ad395a5d6 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/memory.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/memory.go @@ -5,13 +5,23 @@ package fs import ( "bufio" "fmt" + "io/ioutil" "os" "path/filepath" "strconv" "strings" + "syscall" // only for Errno "github.com/opencontainers/runc/libcontainer/cgroups" "github.com/opencontainers/runc/libcontainer/configs" + + "golang.org/x/sys/unix" +) + +const ( + cgroupKernelMemoryLimit = "memory.kmem.limit_in_bytes" + cgroupMemorySwapLimit = "memory.memsw.limit_in_bytes" + cgroupMemoryLimit = "memory.limit_in_bytes" ) type MemoryGroup struct { @@ -25,20 +35,23 @@ func (s *MemoryGroup) Apply(d *cgroupData) (err error) { path, err := d.path("memory") if err != nil && !cgroups.IsNotFound(err) { return err + } else if path == "" { + return nil } if memoryAssigned(d.config) { - if path != "" { + if _, err := os.Stat(path); os.IsNotExist(err) { if err := os.MkdirAll(path, 0755); err != nil { return err } - } - // We have to set kernel memory here, as we can't change it once - // processes have been attached. - if err := s.SetKernelMemory(path, d.config); err != nil { - return err + // Only enable kernel memory accouting when this cgroup + // is created by libcontainer, otherwise we might get + // error when people use `cgroupsPath` to join an existed + // cgroup whose kernel memory is not initialized. + if err := EnableKernelMemoryAccounting(path); err != nil { + return err + } } } - defer func() { if err != nil { os.RemoveAll(path) @@ -54,33 +67,113 @@ func (s *MemoryGroup) Apply(d *cgroupData) (err error) { return nil } -func (s *MemoryGroup) SetKernelMemory(path string, cgroup *configs.Cgroup) error { - // This has to be done separately because it has special constraints (it - // can't be done after there are processes attached to the cgroup). - if cgroup.Resources.KernelMemory > 0 { - if err := writeFile(path, "memory.kmem.limit_in_bytes", strconv.FormatInt(cgroup.Resources.KernelMemory, 10)); err != nil { +func EnableKernelMemoryAccounting(path string) error { + // Check if kernel memory is enabled + // We have to limit the kernel memory here as it won't be accounted at all + // until a limit is set on the cgroup and limit cannot be set once the + // cgroup has children, or if there are already tasks in the cgroup. + for _, i := range []int64{1, -1} { + if err := setKernelMemory(path, i); err != nil { return err } } return nil } +func setKernelMemory(path string, kernelMemoryLimit int64) error { + if path == "" { + return fmt.Errorf("no such directory for %s", cgroupKernelMemoryLimit) + } + if !cgroups.PathExists(filepath.Join(path, cgroupKernelMemoryLimit)) { + // kernel memory is not enabled on the system so we should do nothing + return nil + } + if err := ioutil.WriteFile(filepath.Join(path, cgroupKernelMemoryLimit), []byte(strconv.FormatInt(kernelMemoryLimit, 10)), 0700); err != nil { + // Check if the error number returned by the syscall is "EBUSY" + // The EBUSY signal is returned on attempts to write to the + // memory.kmem.limit_in_bytes file if the cgroup has children or + // once tasks have been attached to the cgroup + if pathErr, ok := err.(*os.PathError); ok { + if errNo, ok := pathErr.Err.(syscall.Errno); ok { + if errNo == unix.EBUSY { + return fmt.Errorf("failed to set %s, because either tasks have already joined this cgroup or it has children", cgroupKernelMemoryLimit) + } + } + } + return fmt.Errorf("failed to write %v to %v: %v", kernelMemoryLimit, cgroupKernelMemoryLimit, err) + } + return nil +} + +func setMemoryAndSwap(path string, cgroup *configs.Cgroup) error { + // If the memory update is set to -1 we should also + // set swap to -1, it means unlimited memory. + if cgroup.Resources.Memory == -1 { + // Only set swap if it's enabled in kernel + if cgroups.PathExists(filepath.Join(path, cgroupMemorySwapLimit)) { + cgroup.Resources.MemorySwap = -1 + } + } + + // When memory and swap memory are both set, we need to handle the cases + // for updating container. + if cgroup.Resources.Memory != 0 && cgroup.Resources.MemorySwap != 0 { + memoryUsage, err := getMemoryData(path, "") + if err != nil { + return err + } + + // When update memory limit, we should adapt the write sequence + // for memory and swap memory, so it won't fail because the new + // value and the old value don't fit kernel's validation. + if cgroup.Resources.MemorySwap == -1 || memoryUsage.Limit < uint64(cgroup.Resources.MemorySwap) { + if err := writeFile(path, cgroupMemorySwapLimit, strconv.FormatInt(cgroup.Resources.MemorySwap, 10)); err != nil { + return err + } + if err := writeFile(path, cgroupMemoryLimit, strconv.FormatInt(cgroup.Resources.Memory, 10)); err != nil { + return err + } + } else { + if err := writeFile(path, cgroupMemoryLimit, strconv.FormatInt(cgroup.Resources.Memory, 10)); err != nil { + return err + } + if err := writeFile(path, cgroupMemorySwapLimit, strconv.FormatInt(cgroup.Resources.MemorySwap, 10)); err != nil { + return err + } + } + } else { + if cgroup.Resources.Memory != 0 { + if err := writeFile(path, cgroupMemoryLimit, strconv.FormatInt(cgroup.Resources.Memory, 10)); err != nil { + return err + } + } + if cgroup.Resources.MemorySwap != 0 { + if err := writeFile(path, cgroupMemorySwapLimit, strconv.FormatInt(cgroup.Resources.MemorySwap, 10)); err != nil { + return err + } + } + } + + return nil +} + func (s *MemoryGroup) Set(path string, cgroup *configs.Cgroup) error { - if cgroup.Resources.Memory != 0 { - if err := writeFile(path, "memory.limit_in_bytes", strconv.FormatInt(cgroup.Resources.Memory, 10)); err != nil { + if err := setMemoryAndSwap(path, cgroup); err != nil { + return err + } + + if cgroup.Resources.KernelMemory != 0 { + if err := setKernelMemory(path, cgroup.Resources.KernelMemory); err != nil { return err } } + if cgroup.Resources.MemoryReservation != 0 { if err := writeFile(path, "memory.soft_limit_in_bytes", strconv.FormatInt(cgroup.Resources.MemoryReservation, 10)); err != nil { return err } } - if cgroup.Resources.MemorySwap > 0 { - if err := writeFile(path, "memory.memsw.limit_in_bytes", strconv.FormatInt(cgroup.Resources.MemorySwap, 10)); err != nil { - return err - } - } + if cgroup.Resources.KernelMemoryTCP != 0 { if err := writeFile(path, "memory.kmem.tcp.limit_in_bytes", strconv.FormatInt(cgroup.Resources.KernelMemoryTCP, 10)); err != nil { return err @@ -93,12 +186,12 @@ func (s *MemoryGroup) Set(path string, cgroup *configs.Cgroup) error { } if cgroup.Resources.MemorySwappiness == nil || int64(*cgroup.Resources.MemorySwappiness) == -1 { return nil - } else if int64(*cgroup.Resources.MemorySwappiness) >= 0 && int64(*cgroup.Resources.MemorySwappiness) <= 100 { - if err := writeFile(path, "memory.swappiness", strconv.FormatInt(*cgroup.Resources.MemorySwappiness, 10)); err != nil { + } else if *cgroup.Resources.MemorySwappiness <= 100 { + if err := writeFile(path, "memory.swappiness", strconv.FormatUint(*cgroup.Resources.MemorySwappiness, 10)); err != nil { return err } } else { - return fmt.Errorf("invalid value:%d. valid memory swappiness range is 0-100", int64(*cgroup.Resources.MemorySwappiness)) + return fmt.Errorf("invalid value:%d. valid memory swappiness range is 0-100", *cgroup.Resources.MemorySwappiness) } return nil @@ -150,6 +243,14 @@ func (s *MemoryGroup) GetStats(path string, stats *cgroups.Stats) error { } stats.MemoryStats.KernelTCPUsage = kernelTCPUsage + useHierarchy := strings.Join([]string{"memory", "use_hierarchy"}, ".") + value, err := getCgroupParamUint(path, useHierarchy) + if err != nil { + return err + } + if value == 1 { + stats.MemoryStats.UseHierarchy = true + } return nil } @@ -160,7 +261,7 @@ func memoryAssigned(cgroup *configs.Cgroup) bool { cgroup.Resources.KernelMemory > 0 || cgroup.Resources.KernelMemoryTCP > 0 || cgroup.Resources.OomKillDisable || - (cgroup.Resources.MemorySwappiness != nil && *cgroup.Resources.MemorySwappiness != -1) + (cgroup.Resources.MemorySwappiness != nil && int64(*cgroup.Resources.MemorySwappiness) != -1) } func getMemoryData(path, name string) (cgroups.MemoryData, error) { diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/net_cls.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/net_cls.go index 8a4054ba8..8e74b645e 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/net_cls.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/net_cls.go @@ -3,6 +3,8 @@ package fs import ( + "strconv" + "github.com/opencontainers/runc/libcontainer/cgroups" "github.com/opencontainers/runc/libcontainer/configs" ) @@ -23,8 +25,8 @@ func (s *NetClsGroup) Apply(d *cgroupData) error { } func (s *NetClsGroup) Set(path string, cgroup *configs.Cgroup) error { - if cgroup.Resources.NetClsClassid != "" { - if err := writeFile(path, "net_cls.classid", cgroup.Resources.NetClsClassid); err != nil { + if cgroup.Resources.NetClsClassid != 0 { + if err := writeFile(path, "net_cls.classid", strconv.FormatUint(uint64(cgroup.Resources.NetClsClassid), 10)); err != nil { return err } } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/utils.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/utils.go index 852b18391..5ff0a1615 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/utils.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/fs/utils.go @@ -12,7 +12,6 @@ import ( ) var ( - ErrNotSupportStat = errors.New("stats are not supported for subsystem") ErrNotValidFormat = errors.New("line is not a valid key value format") ) diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/stats.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/stats.go index 797a923c3..8eeedc55b 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/stats.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/stats.go @@ -11,6 +11,7 @@ type ThrottlingData struct { ThrottledTime uint64 `json:"throttled_time,omitempty"` } +// CpuUsage denotes the usage of a CPU. // All CPU stats are aggregate since container inception. type CpuUsage struct { // Total CPU time consumed. @@ -50,6 +51,8 @@ type MemoryStats struct { KernelUsage MemoryData `json:"kernel_usage,omitempty"` // usage of kernel TCP memory KernelTCPUsage MemoryData `json:"kernel_tcp_usage,omitempty"` + // if true, memory usage is accounted for throughout a hierarchy of cgroups. + UseHierarchy bool `json:"use_hierarchy"` Stats map[string]uint64 `json:"stats,omitempty"` } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/systemd/apply_nosystemd.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/systemd/apply_nosystemd.go new file mode 100644 index 000000000..a65d8e443 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/systemd/apply_nosystemd.go @@ -0,0 +1,55 @@ +// +build !linux static_build + +package systemd + +import ( + "fmt" + + "github.com/opencontainers/runc/libcontainer/cgroups" + "github.com/opencontainers/runc/libcontainer/configs" +) + +type Manager struct { + Cgroups *configs.Cgroup + Paths map[string]string +} + +func UseSystemd() bool { + return false +} + +func (m *Manager) Apply(pid int) error { + return fmt.Errorf("Systemd not supported") +} + +func (m *Manager) GetPids() ([]int, error) { + return nil, fmt.Errorf("Systemd not supported") +} + +func (m *Manager) GetAllPids() ([]int, error) { + return nil, fmt.Errorf("Systemd not supported") +} + +func (m *Manager) Destroy() error { + return fmt.Errorf("Systemd not supported") +} + +func (m *Manager) GetPaths() map[string]string { + return nil +} + +func (m *Manager) GetStats() (*cgroups.Stats, error) { + return nil, fmt.Errorf("Systemd not supported") +} + +func (m *Manager) Set(container *configs.Config) error { + return fmt.Errorf("Systemd not supported") +} + +func (m *Manager) Freeze(state configs.FreezerState) error { + return fmt.Errorf("Systemd not supported") +} + +func Freeze(c *configs.Cgroup, state configs.FreezerState) error { + return fmt.Errorf("Systemd not supported") +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/systemd/apply_systemd.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/systemd/apply_systemd.go new file mode 100644 index 000000000..0c625ba34 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/systemd/apply_systemd.go @@ -0,0 +1,598 @@ +// +build linux,!static_build + +package systemd + +import ( + "errors" + "fmt" + "math" + "os" + "path/filepath" + "strings" + "sync" + "time" + + systemdDbus "github.com/coreos/go-systemd/dbus" + systemdUtil "github.com/coreos/go-systemd/util" + "github.com/godbus/dbus" + "github.com/opencontainers/runc/libcontainer/cgroups" + "github.com/opencontainers/runc/libcontainer/cgroups/fs" + "github.com/opencontainers/runc/libcontainer/configs" + "github.com/sirupsen/logrus" +) + +type Manager struct { + mu sync.Mutex + Cgroups *configs.Cgroup + Paths map[string]string +} + +type subsystem interface { + // Name returns the name of the subsystem. + Name() string + // Returns the stats, as 'stats', corresponding to the cgroup under 'path'. + GetStats(path string, stats *cgroups.Stats) error + // Set the cgroup represented by cgroup. + Set(path string, cgroup *configs.Cgroup) error +} + +var errSubsystemDoesNotExist = errors.New("cgroup: subsystem does not exist") + +type subsystemSet []subsystem + +func (s subsystemSet) Get(name string) (subsystem, error) { + for _, ss := range s { + if ss.Name() == name { + return ss, nil + } + } + return nil, errSubsystemDoesNotExist +} + +var subsystems = subsystemSet{ + &fs.CpusetGroup{}, + &fs.DevicesGroup{}, + &fs.MemoryGroup{}, + &fs.CpuGroup{}, + &fs.CpuacctGroup{}, + &fs.PidsGroup{}, + &fs.BlkioGroup{}, + &fs.HugetlbGroup{}, + &fs.PerfEventGroup{}, + &fs.FreezerGroup{}, + &fs.NetPrioGroup{}, + &fs.NetClsGroup{}, + &fs.NameGroup{GroupName: "name=systemd"}, +} + +const ( + testScopeWait = 4 + testSliceWait = 4 +) + +var ( + connLock sync.Mutex + theConn *systemdDbus.Conn + hasStartTransientUnit bool + hasStartTransientSliceUnit bool + hasTransientDefaultDependencies bool + hasDelegateScope bool + hasDelegateSlice bool +) + +func newProp(name string, units interface{}) systemdDbus.Property { + return systemdDbus.Property{ + Name: name, + Value: dbus.MakeVariant(units), + } +} + +func UseSystemd() bool { + if !systemdUtil.IsRunningSystemd() { + return false + } + + connLock.Lock() + defer connLock.Unlock() + + if theConn == nil { + var err error + theConn, err = systemdDbus.New() + if err != nil { + return false + } + + // Assume we have StartTransientUnit + hasStartTransientUnit = true + + // But if we get UnknownMethod error we don't + if _, err := theConn.StartTransientUnit("test.scope", "invalid", nil, nil); err != nil { + if dbusError, ok := err.(dbus.Error); ok { + if dbusError.Name == "org.freedesktop.DBus.Error.UnknownMethod" { + hasStartTransientUnit = false + return hasStartTransientUnit + } + } + } + + // Ensure the scope name we use doesn't exist. Use the Pid to + // avoid collisions between multiple libcontainer users on a + // single host. + scope := fmt.Sprintf("libcontainer-%d-systemd-test-default-dependencies.scope", os.Getpid()) + testScopeExists := true + for i := 0; i <= testScopeWait; i++ { + if _, err := theConn.StopUnit(scope, "replace", nil); err != nil { + if dbusError, ok := err.(dbus.Error); ok { + if strings.Contains(dbusError.Name, "org.freedesktop.systemd1.NoSuchUnit") { + testScopeExists = false + break + } + } + } + time.Sleep(time.Millisecond) + } + + // Bail out if we can't kill this scope without testing for DefaultDependencies + if testScopeExists { + return hasStartTransientUnit + } + + // Assume StartTransientUnit on a scope allows DefaultDependencies + hasTransientDefaultDependencies = true + ddf := newProp("DefaultDependencies", false) + if _, err := theConn.StartTransientUnit(scope, "replace", []systemdDbus.Property{ddf}, nil); err != nil { + if dbusError, ok := err.(dbus.Error); ok { + if strings.Contains(dbusError.Name, "org.freedesktop.DBus.Error.PropertyReadOnly") { + hasTransientDefaultDependencies = false + } + } + } + + // Not critical because of the stop unit logic above. + theConn.StopUnit(scope, "replace", nil) + + // Assume StartTransientUnit on a scope allows Delegate + hasDelegateScope = true + dlScope := newProp("Delegate", true) + if _, err := theConn.StartTransientUnit(scope, "replace", []systemdDbus.Property{dlScope}, nil); err != nil { + if dbusError, ok := err.(dbus.Error); ok { + if strings.Contains(dbusError.Name, "org.freedesktop.DBus.Error.PropertyReadOnly") { + hasDelegateScope = false + } + } + } + + // Assume we have the ability to start a transient unit as a slice + // This was broken until systemd v229, but has been back-ported on RHEL environments >= 219 + // For details, see: https://bugzilla.redhat.com/show_bug.cgi?id=1370299 + hasStartTransientSliceUnit = true + + // To ensure simple clean-up, we create a slice off the root with no hierarchy + slice := fmt.Sprintf("libcontainer_%d_systemd_test_default.slice", os.Getpid()) + if _, err := theConn.StartTransientUnit(slice, "replace", nil, nil); err != nil { + if _, ok := err.(dbus.Error); ok { + hasStartTransientSliceUnit = false + } + } + + for i := 0; i <= testSliceWait; i++ { + if _, err := theConn.StopUnit(slice, "replace", nil); err != nil { + if dbusError, ok := err.(dbus.Error); ok { + if strings.Contains(dbusError.Name, "org.freedesktop.systemd1.NoSuchUnit") { + hasStartTransientSliceUnit = false + break + } + } + } else { + break + } + time.Sleep(time.Millisecond) + } + + // Not critical because of the stop unit logic above. + theConn.StopUnit(slice, "replace", nil) + + // Assume StartTransientUnit on a slice allows Delegate + hasDelegateSlice = true + dlSlice := newProp("Delegate", true) + if _, err := theConn.StartTransientUnit(slice, "replace", []systemdDbus.Property{dlSlice}, nil); err != nil { + if dbusError, ok := err.(dbus.Error); ok { + // Starting with systemd v237, Delegate is not even a property of slices anymore, + // so the D-Bus call fails with "InvalidArgs" error. + if strings.Contains(dbusError.Name, "org.freedesktop.DBus.Error.PropertyReadOnly") || strings.Contains(dbusError.Name, "org.freedesktop.DBus.Error.InvalidArgs") { + hasDelegateSlice = false + } + } + } + + // Not critical because of the stop unit logic above. + theConn.StopUnit(scope, "replace", nil) + theConn.StopUnit(slice, "replace", nil) + } + return hasStartTransientUnit +} + +func (m *Manager) Apply(pid int) error { + var ( + c = m.Cgroups + unitName = getUnitName(c) + slice = "system.slice" + properties []systemdDbus.Property + ) + + if c.Paths != nil { + paths := make(map[string]string) + for name, path := range c.Paths { + _, err := getSubsystemPath(m.Cgroups, name) + if err != nil { + // Don't fail if a cgroup hierarchy was not found, just skip this subsystem + if cgroups.IsNotFound(err) { + continue + } + return err + } + paths[name] = path + } + m.Paths = paths + return cgroups.EnterPid(m.Paths, pid) + } + + if c.Parent != "" { + slice = c.Parent + } + + properties = append(properties, systemdDbus.PropDescription("libcontainer container "+c.Name)) + + // if we create a slice, the parent is defined via a Wants= + if strings.HasSuffix(unitName, ".slice") { + // This was broken until systemd v229, but has been back-ported on RHEL environments >= 219 + if !hasStartTransientSliceUnit { + return fmt.Errorf("systemd version does not support ability to start a slice as transient unit") + } + properties = append(properties, systemdDbus.PropWants(slice)) + } else { + // otherwise, we use Slice= + properties = append(properties, systemdDbus.PropSlice(slice)) + } + + // only add pid if its valid, -1 is used w/ general slice creation. + if pid != -1 { + properties = append(properties, newProp("PIDs", []uint32{uint32(pid)})) + } + + // Check if we can delegate. This is only supported on systemd versions 218 and above. + if strings.HasSuffix(unitName, ".slice") { + if hasDelegateSlice { + // systemd 237 and above no longer allows delegation on a slice + properties = append(properties, newProp("Delegate", true)) + } + } else { + if hasDelegateScope { + properties = append(properties, newProp("Delegate", true)) + } + } + + // Always enable accounting, this gets us the same behaviour as the fs implementation, + // plus the kernel has some problems with joining the memory cgroup at a later time. + properties = append(properties, + newProp("MemoryAccounting", true), + newProp("CPUAccounting", true), + newProp("BlockIOAccounting", true)) + + if hasTransientDefaultDependencies { + properties = append(properties, + newProp("DefaultDependencies", false)) + } + + if c.Resources.Memory != 0 { + properties = append(properties, + newProp("MemoryLimit", uint64(c.Resources.Memory))) + } + + if c.Resources.CpuShares != 0 { + properties = append(properties, + newProp("CPUShares", c.Resources.CpuShares)) + } + + // cpu.cfs_quota_us and cpu.cfs_period_us are controlled by systemd. + if c.Resources.CpuQuota != 0 && c.Resources.CpuPeriod != 0 { + // corresponds to USEC_INFINITY in systemd + // if USEC_INFINITY is provided, CPUQuota is left unbound by systemd + // always setting a property value ensures we can apply a quota and remove it later + cpuQuotaPerSecUSec := uint64(math.MaxUint64) + if c.Resources.CpuQuota > 0 { + // systemd converts CPUQuotaPerSecUSec (microseconds per CPU second) to CPUQuota + // (integer percentage of CPU) internally. This means that if a fractional percent of + // CPU is indicated by Resources.CpuQuota, we need to round up to the nearest + // 10ms (1% of a second) such that child cgroups can set the cpu.cfs_quota_us they expect. + cpuQuotaPerSecUSec = uint64(c.Resources.CpuQuota*1000000) / c.Resources.CpuPeriod + if cpuQuotaPerSecUSec%10000 != 0 { + cpuQuotaPerSecUSec = ((cpuQuotaPerSecUSec / 10000) + 1) * 10000 + } + } + properties = append(properties, + newProp("CPUQuotaPerSecUSec", cpuQuotaPerSecUSec)) + } + + if c.Resources.BlkioWeight != 0 { + properties = append(properties, + newProp("BlockIOWeight", uint64(c.Resources.BlkioWeight))) + } + + // We have to set kernel memory here, as we can't change it once + // processes have been attached to the cgroup. + if c.Resources.KernelMemory != 0 { + if err := setKernelMemory(c); err != nil { + return err + } + } + + statusChan := make(chan string, 1) + if _, err := theConn.StartTransientUnit(unitName, "replace", properties, statusChan); err == nil { + select { + case <-statusChan: + case <-time.After(time.Second): + logrus.Warnf("Timed out while waiting for StartTransientUnit(%s) completion signal from dbus. Continuing...", unitName) + } + } else if !isUnitExists(err) { + return err + } + + if err := joinCgroups(c, pid); err != nil { + return err + } + + paths := make(map[string]string) + for _, s := range subsystems { + subsystemPath, err := getSubsystemPath(m.Cgroups, s.Name()) + if err != nil { + // Don't fail if a cgroup hierarchy was not found, just skip this subsystem + if cgroups.IsNotFound(err) { + continue + } + return err + } + paths[s.Name()] = subsystemPath + } + m.Paths = paths + return nil +} + +func (m *Manager) Destroy() error { + if m.Cgroups.Paths != nil { + return nil + } + m.mu.Lock() + defer m.mu.Unlock() + theConn.StopUnit(getUnitName(m.Cgroups), "replace", nil) + if err := cgroups.RemovePaths(m.Paths); err != nil { + return err + } + m.Paths = make(map[string]string) + return nil +} + +func (m *Manager) GetPaths() map[string]string { + m.mu.Lock() + paths := m.Paths + m.mu.Unlock() + return paths +} + +func join(c *configs.Cgroup, subsystem string, pid int) (string, error) { + path, err := getSubsystemPath(c, subsystem) + if err != nil { + return "", err + } + if err := os.MkdirAll(path, 0755); err != nil { + return "", err + } + if err := cgroups.WriteCgroupProc(path, pid); err != nil { + return "", err + } + return path, nil +} + +func joinCgroups(c *configs.Cgroup, pid int) error { + for _, sys := range subsystems { + name := sys.Name() + switch name { + case "name=systemd": + // let systemd handle this + case "cpuset": + path, err := getSubsystemPath(c, name) + if err != nil && !cgroups.IsNotFound(err) { + return err + } + s := &fs.CpusetGroup{} + if err := s.ApplyDir(path, c, pid); err != nil { + return err + } + default: + _, err := join(c, name, pid) + if err != nil { + // Even if it's `not found` error, we'll return err + // because devices cgroup is hard requirement for + // container security. + if name == "devices" { + return err + } + // For other subsystems, omit the `not found` error + // because they are optional. + if !cgroups.IsNotFound(err) { + return err + } + } + } + } + + return nil +} + +// systemd represents slice hierarchy using `-`, so we need to follow suit when +// generating the path of slice. Essentially, test-a-b.slice becomes +// /test.slice/test-a.slice/test-a-b.slice. +func ExpandSlice(slice string) (string, error) { + suffix := ".slice" + // Name has to end with ".slice", but can't be just ".slice". + if len(slice) < len(suffix) || !strings.HasSuffix(slice, suffix) { + return "", fmt.Errorf("invalid slice name: %s", slice) + } + + // Path-separators are not allowed. + if strings.Contains(slice, "/") { + return "", fmt.Errorf("invalid slice name: %s", slice) + } + + var path, prefix string + sliceName := strings.TrimSuffix(slice, suffix) + // if input was -.slice, we should just return root now + if sliceName == "-" { + return "/", nil + } + for _, component := range strings.Split(sliceName, "-") { + // test--a.slice isn't permitted, nor is -test.slice. + if component == "" { + return "", fmt.Errorf("invalid slice name: %s", slice) + } + + // Append the component to the path and to the prefix. + path += "/" + prefix + component + suffix + prefix += component + "-" + } + return path, nil +} + +func getSubsystemPath(c *configs.Cgroup, subsystem string) (string, error) { + mountpoint, err := cgroups.FindCgroupMountpoint(subsystem) + if err != nil { + return "", err + } + + initPath, err := cgroups.GetInitCgroup(subsystem) + if err != nil { + return "", err + } + // if pid 1 is systemd 226 or later, it will be in init.scope, not the root + initPath = strings.TrimSuffix(filepath.Clean(initPath), "init.scope") + + slice := "system.slice" + if c.Parent != "" { + slice = c.Parent + } + + slice, err = ExpandSlice(slice) + if err != nil { + return "", err + } + + return filepath.Join(mountpoint, initPath, slice, getUnitName(c)), nil +} + +func (m *Manager) Freeze(state configs.FreezerState) error { + path, err := getSubsystemPath(m.Cgroups, "freezer") + if err != nil { + return err + } + prevState := m.Cgroups.Resources.Freezer + m.Cgroups.Resources.Freezer = state + freezer, err := subsystems.Get("freezer") + if err != nil { + return err + } + err = freezer.Set(path, m.Cgroups) + if err != nil { + m.Cgroups.Resources.Freezer = prevState + return err + } + return nil +} + +func (m *Manager) GetPids() ([]int, error) { + path, err := getSubsystemPath(m.Cgroups, "devices") + if err != nil { + return nil, err + } + return cgroups.GetPids(path) +} + +func (m *Manager) GetAllPids() ([]int, error) { + path, err := getSubsystemPath(m.Cgroups, "devices") + if err != nil { + return nil, err + } + return cgroups.GetAllPids(path) +} + +func (m *Manager) GetStats() (*cgroups.Stats, error) { + m.mu.Lock() + defer m.mu.Unlock() + stats := cgroups.NewStats() + for name, path := range m.Paths { + sys, err := subsystems.Get(name) + if err == errSubsystemDoesNotExist || !cgroups.PathExists(path) { + continue + } + if err := sys.GetStats(path, stats); err != nil { + return nil, err + } + } + + return stats, nil +} + +func (m *Manager) Set(container *configs.Config) error { + // If Paths are set, then we are just joining cgroups paths + // and there is no need to set any values. + if m.Cgroups.Paths != nil { + return nil + } + for _, sys := range subsystems { + // Get the subsystem path, but don't error out for not found cgroups. + path, err := getSubsystemPath(container.Cgroups, sys.Name()) + if err != nil && !cgroups.IsNotFound(err) { + return err + } + + if err := sys.Set(path, container.Cgroups); err != nil { + return err + } + } + + if m.Paths["cpu"] != "" { + if err := fs.CheckCpushares(m.Paths["cpu"], container.Cgroups.Resources.CpuShares); err != nil { + return err + } + } + return nil +} + +func getUnitName(c *configs.Cgroup) string { + // by default, we create a scope unless the user explicitly asks for a slice. + if !strings.HasSuffix(c.Name, ".slice") { + return fmt.Sprintf("%s-%s.scope", c.ScopePrefix, c.Name) + } + return c.Name +} + +func setKernelMemory(c *configs.Cgroup) error { + path, err := getSubsystemPath(c, "memory") + if err != nil && !cgroups.IsNotFound(err) { + return err + } + + if err := os.MkdirAll(path, 0755); err != nil { + return err + } + return fs.EnableKernelMemoryAccounting(path) +} + +// isUnitExists returns true if the error is that a systemd unit already exists. +func isUnitExists(err error) bool { + if err != nil { + if dbusError, ok := err.(dbus.Error); ok { + return strings.Contains(dbusError.Name, "org.freedesktop.systemd1.UnitExists") + } + } + return false +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/utils.go b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/utils.go index 235273299..7c995efee 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/cgroups/utils.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/cgroups/utils.go @@ -16,37 +16,24 @@ import ( "github.com/docker/go-units" ) -const cgroupNamePrefix = "name=" +const ( + cgroupNamePrefix = "name=" + CgroupProcesses = "cgroup.procs" +) -// https://www.kernel.org/doc/Documentation/cgroups/cgroups.txt +// https://www.kernel.org/doc/Documentation/cgroup-v1/cgroups.txt func FindCgroupMountpoint(subsystem string) (string, error) { - // We are not using mount.GetMounts() because it's super-inefficient, - // parsing it directly sped up x10 times because of not using Sscanf. - // It was one of two major performance drawbacks in container start. - f, err := os.Open("/proc/self/mountinfo") - if err != nil { - return "", err - } - defer f.Close() - - scanner := bufio.NewScanner(f) - for scanner.Scan() { - txt := scanner.Text() - fields := strings.Split(txt, " ") - for _, opt := range strings.Split(fields[len(fields)-1], ",") { - if opt == subsystem { - return fields[4], nil - } - } - } - if err := scanner.Err(); err != nil { - return "", err - } - - return "", NewNotFoundError(subsystem) + mnt, _, err := FindCgroupMountpointAndRoot(subsystem) + return mnt, err } func FindCgroupMountpointAndRoot(subsystem string) (string, string, error) { + // We are not using mount.GetMounts() because it's super-inefficient, + // parsing it directly sped up x10 times because of not using Sscanf. + // It was one of two major performance drawbacks in container start. + if !isSubsystemAvailable(subsystem) { + return "", "", NewNotFoundError(subsystem) + } f, err := os.Open("/proc/self/mountinfo") if err != nil { return "", "", err @@ -70,6 +57,30 @@ func FindCgroupMountpointAndRoot(subsystem string) (string, string, error) { return "", "", NewNotFoundError(subsystem) } +func isSubsystemAvailable(subsystem string) bool { + cgroups, err := ParseCgroupFile("/proc/self/cgroup") + if err != nil { + return false + } + _, avail := cgroups[subsystem] + return avail +} + +func GetClosestMountpointAncestor(dir, mountinfo string) string { + deepestMountPoint := "" + for _, mountInfoEntry := range strings.Split(mountinfo, "\n") { + mountInfoParts := strings.Fields(mountInfoEntry) + if len(mountInfoParts) < 5 { + continue + } + mountPoint := mountInfoParts[4] + if strings.HasPrefix(mountPoint, deepestMountPoint) && strings.HasPrefix(dir, mountPoint) { + deepestMountPoint = mountPoint + } + } + return deepestMountPoint +} + func FindCgroupMountpointDir() (string, error) { f, err := os.Open("/proc/self/mountinfo") if err != nil { @@ -113,7 +124,7 @@ type Mount struct { Subsystems []string } -func (m Mount) GetThisCgroupDir(cgroups map[string]string) (string, error) { +func (m Mount) GetOwnCgroup(cgroups map[string]string) (string, error) { if len(m.Subsystems) == 0 { return "", fmt.Errorf("no subsystem for mount") } @@ -121,16 +132,17 @@ func (m Mount) GetThisCgroupDir(cgroups map[string]string) (string, error) { return getControllerPath(m.Subsystems[0], cgroups) } -func getCgroupMountsHelper(ss map[string]bool, mi io.Reader) ([]Mount, error) { +func getCgroupMountsHelper(ss map[string]bool, mi io.Reader, all bool) ([]Mount, error) { res := make([]Mount, 0, len(ss)) scanner := bufio.NewScanner(mi) - for scanner.Scan() { + numFound := 0 + for scanner.Scan() && numFound < len(ss) { txt := scanner.Text() sepIdx := strings.Index(txt, " - ") if sepIdx == -1 { return nil, fmt.Errorf("invalid mountinfo format") } - if txt[sepIdx+3:sepIdx+9] != "cgroup" { + if txt[sepIdx+3:sepIdx+10] == "cgroup2" || txt[sepIdx+3:sepIdx+9] != "cgroup" { continue } fields := strings.Split(txt, " ") @@ -139,12 +151,17 @@ func getCgroupMountsHelper(ss map[string]bool, mi io.Reader) ([]Mount, error) { Root: fields[3], } for _, opt := range strings.Split(fields[len(fields)-1], ",") { + if !ss[opt] { + continue + } if strings.HasPrefix(opt, cgroupNamePrefix) { m.Subsystems = append(m.Subsystems, opt[len(cgroupNamePrefix):]) - } - if ss[opt] { + } else { m.Subsystems = append(m.Subsystems, opt) } + if !all { + numFound++ + } } res = append(res, m) } @@ -154,26 +171,28 @@ func getCgroupMountsHelper(ss map[string]bool, mi io.Reader) ([]Mount, error) { return res, nil } -func GetCgroupMounts() ([]Mount, error) { +// GetCgroupMounts returns the mounts for the cgroup subsystems. +// all indicates whether to return just the first instance or all the mounts. +func GetCgroupMounts(all bool) ([]Mount, error) { f, err := os.Open("/proc/self/mountinfo") if err != nil { return nil, err } defer f.Close() - all, err := GetAllSubsystems() + allSubsystems, err := ParseCgroupFile("/proc/self/cgroup") if err != nil { return nil, err } allMap := make(map[string]bool) - for _, s := range all { + for s := range allSubsystems { allMap[s] = true } - return getCgroupMountsHelper(allMap, f) + return getCgroupMountsHelper(allMap, f, all) } -// Returns all the cgroup subsystems supported by the kernel +// GetAllSubsystems returns all the cgroup subsystems supported by the kernel func GetAllSubsystems() ([]string, error) { f, err := os.Open("/proc/cgroups") if err != nil { @@ -185,9 +204,6 @@ func GetAllSubsystems() ([]string, error) { s := bufio.NewScanner(f) for s.Scan() { - if err := s.Err(); err != nil { - return nil, err - } text := s.Text() if text[0] != '#' { parts := strings.Fields(text) @@ -196,11 +212,14 @@ func GetAllSubsystems() ([]string, error) { } } } + if err := s.Err(); err != nil { + return nil, err + } return subsystems, nil } -// Returns the relative path to the cgroup docker is running in. -func GetThisCgroupDir(subsystem string) (string, error) { +// GetOwnCgroup returns the relative path to the cgroup docker is running in. +func GetOwnCgroup(subsystem string) (string, error) { cgroups, err := ParseCgroupFile("/proc/self/cgroup") if err != nil { return "", err @@ -209,8 +228,16 @@ func GetThisCgroupDir(subsystem string) (string, error) { return getControllerPath(subsystem, cgroups) } -func GetInitCgroupDir(subsystem string) (string, error) { +func GetOwnCgroupPath(subsystem string) (string, error) { + cgroup, err := GetOwnCgroup(subsystem) + if err != nil { + return "", err + } + return getCgroupPathHelper(subsystem, cgroup) +} + +func GetInitCgroup(subsystem string) (string, error) { cgroups, err := ParseCgroupFile("/proc/1/cgroup") if err != nil { return "", err @@ -219,8 +246,33 @@ func GetInitCgroupDir(subsystem string) (string, error) { return getControllerPath(subsystem, cgroups) } +func GetInitCgroupPath(subsystem string) (string, error) { + cgroup, err := GetInitCgroup(subsystem) + if err != nil { + return "", err + } + + return getCgroupPathHelper(subsystem, cgroup) +} + +func getCgroupPathHelper(subsystem, cgroup string) (string, error) { + mnt, root, err := FindCgroupMountpointAndRoot(subsystem) + if err != nil { + return "", err + } + + // This is needed for nested containers, because in /proc/self/cgroup we + // see pathes from host, which don't exist in container. + relCgroup, err := filepath.Rel(root, cgroup) + if err != nil { + return "", err + } + + return filepath.Join(mnt, relCgroup), nil +} + func readProcsFile(dir string) ([]int, error) { - f, err := os.Open(filepath.Join(dir, "cgroup.procs")) + f, err := os.Open(filepath.Join(dir, CgroupProcesses)) if err != nil { return nil, err } @@ -243,6 +295,8 @@ func readProcsFile(dir string) ([]int, error) { return out, nil } +// ParseCgroupFile parses the given cgroup file, typically from +// /proc//cgroup, into a map of subgroups to cgroup names. func ParseCgroupFile(path string) (map[string]string, error) { f, err := os.Open(path) if err != nil { @@ -250,21 +304,35 @@ func ParseCgroupFile(path string) (map[string]string, error) { } defer f.Close() - s := bufio.NewScanner(f) + return parseCgroupFromReader(f) +} + +// helper function for ParseCgroupFile to make testing easier +func parseCgroupFromReader(r io.Reader) (map[string]string, error) { + s := bufio.NewScanner(r) cgroups := make(map[string]string) for s.Scan() { - if err := s.Err(); err != nil { - return nil, err - } - text := s.Text() - parts := strings.Split(text, ":") + // from cgroups(7): + // /proc/[pid]/cgroup + // ... + // For each cgroup hierarchy ... there is one entry + // containing three colon-separated fields of the form: + // hierarchy-ID:subsystem-list:cgroup-path + parts := strings.SplitN(text, ":", 3) + if len(parts) < 3 { + return nil, fmt.Errorf("invalid cgroup entry: must contain at least two colons: %v", text) + } for _, subs := range strings.Split(parts[1], ",") { cgroups[subs] = parts[2] } } + if err := s.Err(); err != nil { + return nil, err + } + return cgroups, nil } @@ -291,8 +359,7 @@ func PathExists(path string) bool { func EnterPid(cgroupPaths map[string]string, pid int) error { for _, path := range cgroupPaths { if PathExists(path) { - if err := ioutil.WriteFile(filepath.Join(path, "cgroup.procs"), - []byte(strconv.Itoa(pid)), 0700); err != nil { + if err := WriteCgroupProc(path, pid); err != nil { return err } } @@ -361,7 +428,7 @@ func GetAllPids(path string) ([]int, error) { // collect pids from all sub-cgroups err := filepath.Walk(path, func(p string, info os.FileInfo, iErr error) error { dir, file := filepath.Split(p) - if file != "cgroup.procs" { + if file != CgroupProcesses { return nil } if iErr != nil { @@ -376,3 +443,20 @@ func GetAllPids(path string) ([]int, error) { }) return pids, err } + +// WriteCgroupProc writes the specified pid into the cgroup's cgroup.procs file +func WriteCgroupProc(dir string, pid int) error { + // Normally dir should not be empty, one case is that cgroup subsystem + // is not mounted, we will get empty dir, and we want it fail here. + if dir == "" { + return fmt.Errorf("no such directory for %s", CgroupProcesses) + } + + // Dont attach any pid to the cgroup if -1 is specified as a pid + if pid != -1 { + if err := ioutil.WriteFile(filepath.Join(dir, CgroupProcesses), []byte(strconv.Itoa(pid)), 0700); err != nil { + return fmt.Errorf("failed to write %v to %v: %v", pid, CgroupProcesses, err) + } + } + return nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/cgroup_unix.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/cgroup_linux.go similarity index 88% rename from vendor/github.com/opencontainers/runc/libcontainer/configs/cgroup_unix.go rename to vendor/github.com/opencontainers/runc/libcontainer/configs/cgroup_linux.go index f2eff91cf..e15a662f5 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/configs/cgroup_unix.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/cgroup_linux.go @@ -1,5 +1,3 @@ -// +build linux freebsd - package configs type FreezerState string @@ -22,7 +20,7 @@ type Cgroup struct { // The path is assumed to be relative to the host system cgroup mountpoint. Path string `json:"path"` - // ScopePrefix decribes prefix for the scope name + // ScopePrefix describes prefix for the scope name ScopePrefix string `json:"scope_prefix"` // Paths represent the absolute cgroups paths to join. @@ -36,7 +34,7 @@ type Cgroup struct { type Resources struct { // If this is true allow access to any kind of device within the container. If false, allow access only to devices explicitly listed in the allowed_devices list. // Deprecated - AllowAllDevices bool `json:"allow_all_devices,omitempty"` + AllowAllDevices *bool `json:"allow_all_devices,omitempty"` // Deprecated AllowedDevices []*Device `json:"allowed_devices,omitempty"` // Deprecated @@ -60,19 +58,19 @@ type Resources struct { KernelMemoryTCP int64 `json:"kernel_memory_tcp"` // CPU shares (relative weight vs. other containers) - CpuShares int64 `json:"cpu_shares"` + CpuShares uint64 `json:"cpu_shares"` // CPU hardcap limit (in usecs). Allowed cpu time in a given period. CpuQuota int64 `json:"cpu_quota"` // CPU period to be used for hardcapping (in usecs). 0 to use system default. - CpuPeriod int64 `json:"cpu_period"` + CpuPeriod uint64 `json:"cpu_period"` // How many time CPU will use in realtime scheduling (in usecs). - CpuRtRuntime int64 `json:"cpu_quota"` + CpuRtRuntime int64 `json:"cpu_rt_quota"` // CPU period to be used for realtime scheduling (in usecs). - CpuRtPeriod int64 `json:"cpu_period"` + CpuRtPeriod uint64 `json:"cpu_rt_period"` // CPU to use CpusetCpus string `json:"cpuset_cpus"` @@ -95,7 +93,7 @@ type Resources struct { // IO read rate limit per cgroup per device, bytes per second. BlkioThrottleReadBpsDevice []*ThrottleDevice `json:"blkio_throttle_read_bps_device"` - // IO write rate limit per cgroup per divice, bytes per second. + // IO write rate limit per cgroup per device, bytes per second. BlkioThrottleWriteBpsDevice []*ThrottleDevice `json:"blkio_throttle_write_bps_device"` // IO read rate limit per cgroup per device, IO per second. @@ -114,11 +112,11 @@ type Resources struct { OomKillDisable bool `json:"oom_kill_disable"` // Tuning swappiness behaviour per cgroup - MemorySwappiness *int64 `json:"memory_swappiness"` + MemorySwappiness *uint64 `json:"memory_swappiness"` // Set priority of network traffic for container NetPrioIfpriomap []*IfPrioMap `json:"net_prio_ifpriomap"` // Set class identifier for container's network packets - NetClsClassid string `json:"net_cls_classid"` + NetClsClassid uint32 `json:"net_cls_classid_u"` } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/cgroup_unsupported.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/cgroup_unsupported.go deleted file mode 100644 index 95e2830a4..000000000 --- a/vendor/github.com/opencontainers/runc/libcontainer/configs/cgroup_unsupported.go +++ /dev/null @@ -1,6 +0,0 @@ -// +build !windows,!linux,!freebsd - -package configs - -type Cgroup struct { -} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/config.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/config.go index f83f638e3..b1c4762fe 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/configs/config.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/config.go @@ -7,7 +7,9 @@ import ( "os/exec" "time" - "github.com/Sirupsen/logrus" + "github.com/opencontainers/runtime-spec/specs-go" + + "github.com/sirupsen/logrus" ) type Rlimit struct { @@ -33,7 +35,7 @@ type Seccomp struct { Syscalls []*Syscall `json:"syscalls"` } -// An action to be taken upon rule match in Seccomp +// Action is taken upon rule match in Seccomp type Action int const ( @@ -44,7 +46,7 @@ const ( Trace ) -// A comparison operator to be used when matching syscall arguments in Seccomp +// Operator is a comparison operator to be used when matching syscall arguments in Seccomp type Operator int const ( @@ -57,7 +59,7 @@ const ( MaskEqualTo ) -// A rule to match a specific syscall argument in Seccomp +// Arg is a rule to match a specific syscall argument in Seccomp type Arg struct { Index uint `json:"index"` Value uint64 `json:"value"` @@ -65,7 +67,7 @@ type Arg struct { Op Operator `json:"op"` } -// An rule to match a syscall in Seccomp +// Syscall is a rule to match a syscall in Seccomp type Syscall struct { Name string `json:"name"` Action Action `json:"action"` @@ -85,11 +87,6 @@ type Config struct { // that the parent process dies. ParentDeathSignal int `json:"parent_death_signal"` - // PivotDir allows a custom directory inside the container's root filesystem to be used as pivot, when NoPivotRoot is not set. - // When a custom PivotDir not set, a temporary dir inside the root filesystem will be used. The pivot dir needs to be writeable. - // This is required when using read only root filesystems. In these cases, a read/writeable path can be (bind) mounted somewhere inside the root filesystem to act as pivot. - PivotDir string `json:"pivot_dir"` - // Path to a directory containing the container's root filesystem. Rootfs string `json:"rootfs"` @@ -117,8 +114,8 @@ type Config struct { Namespaces Namespaces `json:"namespaces"` // Capabilities specify the capabilities to keep when executing the process inside the container - // All capbilities not specified will be dropped from the processes capability mask - Capabilities []string `json:"capabilities"` + // All capabilities not specified will be dropped from the processes capability mask + Capabilities *Capabilities `json:"capabilities"` // Networks specifies the container's network setup to be created Networks []*Network `json:"networks"` @@ -144,13 +141,10 @@ type Config struct { // OomScoreAdj specifies the adjustment to be made by the kernel when calculating oom scores // for a process. Valid values are between the range [-1000, '1000'], where processes with - // higher scores are preferred for being killed. + // higher scores are preferred for being killed. If it is unset then we don't touch the current + // value. // More information about kernel oom score calculation here: https://lwn.net/Articles/317814/ - OomScoreAdj int `json:"oom_score_adj"` - - // AdditionalGroups specifies the gids that should be added to supplementary groups - // in addition to those that the user belongs to. - AdditionalGroups []string `json:"additional_groups"` + OomScoreAdj *int `json:"oom_score_adj,omitempty"` // UidMappings is an array of User ID mappings for User Namespaces UidMappings []IDMap `json:"uid_mappings"` @@ -187,6 +181,17 @@ type Config struct { // Labels are user defined metadata that is stored in the config and populated on the state Labels []string `json:"labels"` + + // NoNewKeyring will not allocated a new session keyring for the container. It will use the + // callers keyring in this case. + NoNewKeyring bool `json:"no_new_keyring"` + + // Rootless specifies whether the container is a rootless container. + Rootless bool `json:"rootless"` + + // IntelRdt specifies settings for Intel RDT/CAT group that the container is placed into + // to limit the resources (e.g., L3 cache) the container has available + IntelRdt *IntelRdt `json:"intel_rdt,omitempty"` } type Hooks struct { @@ -201,6 +206,19 @@ type Hooks struct { Poststop []Hook } +type Capabilities struct { + // Bounding is the set of capabilities checked by the kernel. + Bounding []string + // Effective is the set of capabilities checked by the kernel. + Effective []string + // Inheritable is the capabilities preserved across execve. + Inheritable []string + // Permitted is the limiting superset for effective capabilities. + Permitted []string + // Ambient is the ambient set of capabilities that are kept. + Ambient []string +} + func (hooks *Hooks) UnmarshalJSON(b []byte) error { var state struct { Prestart []CommandHook @@ -248,19 +266,14 @@ func (hooks Hooks) MarshalJSON() ([]byte, error) { } // HookState is the payload provided to a hook on execution. -type HookState struct { - Version string `json:"version"` - ID string `json:"id"` - Pid int `json:"pid"` - Root string `json:"root"` -} +type HookState specs.State type Hook interface { // Run executes the hook with the provided state. Run(HookState) error } -// NewFunctionHooks will call the provided function when the hook is run. +// NewFunctionHook will call the provided function when the hook is run. func NewFunctionHook(f func(HookState) error) FuncHook { return FuncHook{ run: f, @@ -283,7 +296,7 @@ type Command struct { Timeout *time.Duration `json:"timeout"` } -// NewCommandHooks will execute the provided command when the hook is run. +// NewCommandHook will execute the provided command when the hook is run. func NewCommandHook(cmd Command) CommandHook { return CommandHook{ Command: cmd, @@ -299,25 +312,38 @@ func (c Command) Run(s HookState) error { if err != nil { return err } + var stdout, stderr bytes.Buffer cmd := exec.Cmd{ - Path: c.Path, - Args: c.Args, - Env: c.Env, - Stdin: bytes.NewReader(b), + Path: c.Path, + Args: c.Args, + Env: c.Env, + Stdin: bytes.NewReader(b), + Stdout: &stdout, + Stderr: &stderr, + } + if err := cmd.Start(); err != nil { + return err } errC := make(chan error, 1) go func() { - errC <- cmd.Run() - }() - if c.Timeout != nil { - select { - case err := <-errC: - return err - case <-time.After(*c.Timeout): - cmd.Process.Kill() - cmd.Wait() - return fmt.Errorf("hook ran past specified timeout of %.1fs", c.Timeout.Seconds()) + err := cmd.Wait() + if err != nil { + err = fmt.Errorf("error running hook: %v, stdout: %s, stderr: %s", err, stdout.String(), stderr.String()) } + errC <- err + }() + var timerCh <-chan time.Time + if c.Timeout != nil { + timer := time.NewTimer(*c.Timeout) + defer timer.Stop() + timerCh = timer.C + } + select { + case err := <-errC: + return err + case <-timerCh: + cmd.Process.Kill() + cmd.Wait() + return fmt.Errorf("hook ran past specified timeout of %.1fs", c.Timeout.Seconds()) } - return <-errC } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/config_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/config_linux.go new file mode 100644 index 000000000..07da10804 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/config_linux.go @@ -0,0 +1,61 @@ +package configs + +import "fmt" + +// HostUID gets the translated uid for the process on host which could be +// different when user namespaces are enabled. +func (c Config) HostUID(containerId int) (int, error) { + if c.Namespaces.Contains(NEWUSER) { + if c.UidMappings == nil { + return -1, fmt.Errorf("User namespaces enabled, but no uid mappings found.") + } + id, found := c.hostIDFromMapping(containerId, c.UidMappings) + if !found { + return -1, fmt.Errorf("User namespaces enabled, but no user mapping found.") + } + return id, nil + } + // Return unchanged id. + return containerId, nil +} + +// HostRootUID gets the root uid for the process on host which could be non-zero +// when user namespaces are enabled. +func (c Config) HostRootUID() (int, error) { + return c.HostUID(0) +} + +// HostGID gets the translated gid for the process on host which could be +// different when user namespaces are enabled. +func (c Config) HostGID(containerId int) (int, error) { + if c.Namespaces.Contains(NEWUSER) { + if c.GidMappings == nil { + return -1, fmt.Errorf("User namespaces enabled, but no gid mappings found.") + } + id, found := c.hostIDFromMapping(containerId, c.GidMappings) + if !found { + return -1, fmt.Errorf("User namespaces enabled, but no group mapping found.") + } + return id, nil + } + // Return unchanged id. + return containerId, nil +} + +// HostRootGID gets the root gid for the process on host which could be non-zero +// when user namespaces are enabled. +func (c Config) HostRootGID() (int, error) { + return c.HostGID(0) +} + +// Utility function that gets a host ID for a container ID from user namespace map +// if that ID is present in the map. +func (c Config) hostIDFromMapping(containerID int, uMap []IDMap) (int, bool) { + for _, m := range uMap { + if (containerID >= m.ContainerID) && (containerID <= (m.ContainerID + m.Size - 1)) { + hostID := m.HostID + (containerID - m.ContainerID) + return hostID, true + } + } + return -1, false +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/config_unix.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/config_unix.go deleted file mode 100644 index c447f3ef2..000000000 --- a/vendor/github.com/opencontainers/runc/libcontainer/configs/config_unix.go +++ /dev/null @@ -1,51 +0,0 @@ -// +build freebsd linux - -package configs - -import "fmt" - -// Gets the root uid for the process on host which could be non-zero -// when user namespaces are enabled. -func (c Config) HostUID() (int, error) { - if c.Namespaces.Contains(NEWUSER) { - if c.UidMappings == nil { - return -1, fmt.Errorf("User namespaces enabled, but no user mappings found.") - } - id, found := c.hostIDFromMapping(0, c.UidMappings) - if !found { - return -1, fmt.Errorf("User namespaces enabled, but no root user mapping found.") - } - return id, nil - } - // Return default root uid 0 - return 0, nil -} - -// Gets the root gid for the process on host which could be non-zero -// when user namespaces are enabled. -func (c Config) HostGID() (int, error) { - if c.Namespaces.Contains(NEWUSER) { - if c.GidMappings == nil { - return -1, fmt.Errorf("User namespaces enabled, but no gid mappings found.") - } - id, found := c.hostIDFromMapping(0, c.GidMappings) - if !found { - return -1, fmt.Errorf("User namespaces enabled, but no root group mapping found.") - } - return id, nil - } - // Return default root gid 0 - return 0, nil -} - -// Utility function that gets a host ID for a container ID from user namespace map -// if that ID is present in the map. -func (c Config) hostIDFromMapping(containerID int, uMap []IDMap) (int, bool) { - for _, m := range uMap { - if (containerID >= m.ContainerID) && (containerID <= (m.ContainerID + m.Size - 1)) { - hostID := m.HostID + (containerID - m.ContainerID) - return hostID, true - } - } - return -1, false -} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/device_defaults.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/device_defaults.go index e45299264..e4f423c52 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/configs/device_defaults.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/device_defaults.go @@ -1,9 +1,9 @@ -// +build linux freebsd +// +build linux package configs var ( - // These are devices that are to be both allowed and created. + // DefaultSimpleDevices are devices that are to be both allowed and created. DefaultSimpleDevices = []*Device{ // /dev/null and zero { @@ -107,19 +107,5 @@ var ( Permissions: "rwm", }, }, DefaultSimpleDevices...) - DefaultAutoCreatedDevices = append([]*Device{ - { - // /dev/fuse is created but not allowed. - // This is to allow java to work. Because java - // Insists on there being a /dev/fuse - // https://github.com/docker/docker/issues/514 - // https://github.com/docker/docker/issues/2393 - // - Path: "/dev/fuse", - Type: 'c', - Major: 10, - Minor: 229, - Permissions: "rwm", - }, - }, DefaultSimpleDevices...) + DefaultAutoCreatedDevices = append([]*Device{}, DefaultSimpleDevices...) ) diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/intelrdt.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/intelrdt.go new file mode 100644 index 000000000..36bd5f96a --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/intelrdt.go @@ -0,0 +1,7 @@ +package configs + +type IntelRdt struct { + // The schema for L3 cache id and capacity bitmask (CBM) + // Format: "L3:=;=;..." + L3CacheSchema string `json:"l3_cache_schema,omitempty"` +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/mount.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/mount.go index cc770c916..670757ddb 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/configs/mount.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/mount.go @@ -1,5 +1,11 @@ package configs +const ( + // EXT_COPYUP is a directive to copy up the contents of a directory when + // a tmpfs is mounted over it. + EXT_COPYUP = 1 << iota +) + type Mount struct { // Source path for the mount. Source string `json:"source"` @@ -22,6 +28,9 @@ type Mount struct { // Relabel source if set, "z" indicates shared, "Z" indicates unshared. Relabel string `json:"relabel"` + // Extensions are additional flags that are specific to runc. + Extensions int `json:"extensions"` + // Optional Command to be run before Source is mounted. PremountCmds []Command `json:"premount_cmds"` diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_unix.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_linux.go similarity index 89% rename from vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_unix.go rename to vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_linux.go index b9c820d06..5fc171a57 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_unix.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_linux.go @@ -1,5 +1,3 @@ -// +build linux freebsd - package configs import ( @@ -22,8 +20,8 @@ var ( supportedNamespaces = make(map[NamespaceType]bool) ) -// nsToFile converts the namespace type to its filename -func nsToFile(ns NamespaceType) string { +// NsName converts the namespace type to its filename +func NsName(ns NamespaceType) string { switch ns { case NEWNET: return "net" @@ -50,7 +48,7 @@ func IsNamespaceSupported(ns NamespaceType) bool { if ok { return supported } - nsFile := nsToFile(ns) + nsFile := NsName(ns) // if the namespace type is unknown, just return false if nsFile == "" { return false @@ -64,12 +62,12 @@ func IsNamespaceSupported(ns NamespaceType) bool { func NamespaceTypes() []NamespaceType { return []NamespaceType{ + NEWUSER, // Keep user NS always first, don't move it. + NEWIPC, + NEWUTS, NEWNET, NEWPID, NEWNS, - NEWUTS, - NEWIPC, - NEWUSER, } } @@ -81,10 +79,7 @@ type Namespace struct { } func (n *Namespace) GetPath(pid int) string { - if n.Path != "" { - return n.Path - } - return fmt.Sprintf("/proc/%d/ns/%s", pid, nsToFile(n.Type)) + return fmt.Sprintf("/proc/%d/ns/%s", pid, NsName(n.Type)) } func (n *Namespaces) Remove(t NamespaceType) bool { diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_syscall.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_syscall.go index fb4b85222..4ce6813d2 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_syscall.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_syscall.go @@ -2,19 +2,19 @@ package configs -import "syscall" +import "golang.org/x/sys/unix" func (n *Namespace) Syscall() int { return namespaceInfo[n.Type] } var namespaceInfo = map[NamespaceType]int{ - NEWNET: syscall.CLONE_NEWNET, - NEWNS: syscall.CLONE_NEWNS, - NEWUSER: syscall.CLONE_NEWUSER, - NEWIPC: syscall.CLONE_NEWIPC, - NEWUTS: syscall.CLONE_NEWUTS, - NEWPID: syscall.CLONE_NEWPID, + NEWNET: unix.CLONE_NEWNET, + NEWNS: unix.CLONE_NEWNS, + NEWUSER: unix.CLONE_NEWUSER, + NEWIPC: unix.CLONE_NEWIPC, + NEWUTS: unix.CLONE_NEWUTS, + NEWPID: unix.CLONE_NEWPID, } // CloneFlags parses the container's Namespaces options to set the correct diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_syscall_unsupported.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_syscall_unsupported.go index 0547223a9..5d9a5c81f 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_syscall_unsupported.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_syscall_unsupported.go @@ -4,12 +4,10 @@ package configs func (n *Namespace) Syscall() int { panic("No namespace syscall support") - return 0 } // CloneFlags parses the container's Namespaces options to set the correct // flags on clone, unshare. This function returns flags only for new namespaces. func (n *Namespaces) CloneFlags() uintptr { panic("No namespace syscall support") - return uintptr(0) } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_unsupported.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_unsupported.go index 9a74033ce..19bf713de 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_unsupported.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/namespaces_unsupported.go @@ -1,4 +1,4 @@ -// +build !linux,!freebsd +// +build !linux package configs diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/validate/rootless.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/validate/rootless.go new file mode 100644 index 000000000..8c3954ce2 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/validate/rootless.go @@ -0,0 +1,116 @@ +package validate + +import ( + "fmt" + "os" + "reflect" + "strings" + + "github.com/opencontainers/runc/libcontainer/configs" +) + +var ( + geteuid = os.Geteuid + getegid = os.Getegid +) + +func (v *ConfigValidator) rootless(config *configs.Config) error { + if err := rootlessMappings(config); err != nil { + return err + } + if err := rootlessMount(config); err != nil { + return err + } + + // XXX: We currently can't verify the user config at all, because + // configs.Config doesn't store the user-related configs. So this + // has to be verified by setupUser() in init_linux.go. + + return nil +} + +func hasIDMapping(id int, mappings []configs.IDMap) bool { + for _, m := range mappings { + if id >= m.ContainerID && id < m.ContainerID+m.Size { + return true + } + } + return false +} + +func rootlessMappings(config *configs.Config) error { + if euid := geteuid(); euid != 0 { + if !config.Namespaces.Contains(configs.NEWUSER) { + return fmt.Errorf("rootless containers require user namespaces") + } + if len(config.UidMappings) == 0 { + return fmt.Errorf("rootless containers requires at least one UID mapping") + } + if len(config.GidMappings) == 0 { + return fmt.Errorf("rootless containers requires at least one GID mapping") + } + } + + return nil +} + +// cgroup verifies that the user isn't trying to set any cgroup limits or paths. +func rootlessCgroup(config *configs.Config) error { + // Nothing set at all. + if config.Cgroups == nil || config.Cgroups.Resources == nil { + return nil + } + + // Used for comparing to the zero value. + left := reflect.ValueOf(*config.Cgroups.Resources) + right := reflect.Zero(left.Type()) + + // This is all we need to do, since specconv won't add cgroup options in + // rootless mode. + if !reflect.DeepEqual(left.Interface(), right.Interface()) { + return fmt.Errorf("cannot specify resource limits in rootless container") + } + + return nil +} + +// mount verifies that the user isn't trying to set up any mounts they don't have +// the rights to do. In addition, it makes sure that no mount has a `uid=` or +// `gid=` option that doesn't resolve to root. +func rootlessMount(config *configs.Config) error { + // XXX: We could whitelist allowed devices at this point, but I'm not + // convinced that's a good idea. The kernel is the best arbiter of + // access control. + + for _, mount := range config.Mounts { + // Check that the options list doesn't contain any uid= or gid= entries + // that don't resolve to root. + for _, opt := range strings.Split(mount.Data, ",") { + if strings.HasPrefix(opt, "uid=") { + var uid int + n, err := fmt.Sscanf(opt, "uid=%d", &uid) + if n != 1 || err != nil { + // Ignore unknown mount options. + continue + } + if !hasIDMapping(uid, config.UidMappings) { + return fmt.Errorf("cannot specify uid= mount options for unmapped uid in rootless containers") + } + } + + if strings.HasPrefix(opt, "gid=") { + var gid int + n, err := fmt.Sscanf(opt, "gid=%d", &gid) + if n != 1 || err != nil { + // Ignore unknown mount options. + continue + } + if !hasIDMapping(gid, config.GidMappings) { + return fmt.Errorf("cannot specify gid= mount options for unmapped gid in rootless containers") + } + } + } + } + + return nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/configs/validate/validator.go b/vendor/github.com/opencontainers/runc/libcontainer/configs/validate/validator.go new file mode 100644 index 000000000..b36e553d2 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/configs/validate/validator.go @@ -0,0 +1,222 @@ +package validate + +import ( + "fmt" + "os" + "path/filepath" + "strings" + + "github.com/opencontainers/runc/libcontainer/configs" + "github.com/opencontainers/runc/libcontainer/intelrdt" + selinux "github.com/opencontainers/selinux/go-selinux" +) + +type Validator interface { + Validate(*configs.Config) error +} + +func New() Validator { + return &ConfigValidator{} +} + +type ConfigValidator struct { +} + +func (v *ConfigValidator) Validate(config *configs.Config) error { + if err := v.rootfs(config); err != nil { + return err + } + if err := v.network(config); err != nil { + return err + } + if err := v.hostname(config); err != nil { + return err + } + if err := v.security(config); err != nil { + return err + } + if err := v.usernamespace(config); err != nil { + return err + } + if err := v.sysctl(config); err != nil { + return err + } + if err := v.intelrdt(config); err != nil { + return err + } + if config.Rootless { + if err := v.rootless(config); err != nil { + return err + } + } + return nil +} + +// rootfs validates if the rootfs is an absolute path and is not a symlink +// to the container's root filesystem. +func (v *ConfigValidator) rootfs(config *configs.Config) error { + if _, err := os.Stat(config.Rootfs); err != nil { + if os.IsNotExist(err) { + return fmt.Errorf("rootfs (%s) does not exist", config.Rootfs) + } + return err + } + cleaned, err := filepath.Abs(config.Rootfs) + if err != nil { + return err + } + if cleaned, err = filepath.EvalSymlinks(cleaned); err != nil { + return err + } + if filepath.Clean(config.Rootfs) != cleaned { + return fmt.Errorf("%s is not an absolute path or is a symlink", config.Rootfs) + } + return nil +} + +func (v *ConfigValidator) network(config *configs.Config) error { + if !config.Namespaces.Contains(configs.NEWNET) { + if len(config.Networks) > 0 || len(config.Routes) > 0 { + return fmt.Errorf("unable to apply network settings without a private NET namespace") + } + } + return nil +} + +func (v *ConfigValidator) hostname(config *configs.Config) error { + if config.Hostname != "" && !config.Namespaces.Contains(configs.NEWUTS) { + return fmt.Errorf("unable to set hostname without a private UTS namespace") + } + return nil +} + +func (v *ConfigValidator) security(config *configs.Config) error { + // restrict sys without mount namespace + if (len(config.MaskPaths) > 0 || len(config.ReadonlyPaths) > 0) && + !config.Namespaces.Contains(configs.NEWNS) { + return fmt.Errorf("unable to restrict sys entries without a private MNT namespace") + } + if config.ProcessLabel != "" && !selinux.GetEnabled() { + return fmt.Errorf("selinux label is specified in config, but selinux is disabled or not supported") + } + + return nil +} + +func (v *ConfigValidator) usernamespace(config *configs.Config) error { + if config.Namespaces.Contains(configs.NEWUSER) { + if _, err := os.Stat("/proc/self/ns/user"); os.IsNotExist(err) { + return fmt.Errorf("USER namespaces aren't enabled in the kernel") + } + } else { + if config.UidMappings != nil || config.GidMappings != nil { + return fmt.Errorf("User namespace mappings specified, but USER namespace isn't enabled in the config") + } + } + return nil +} + +// sysctl validates that the specified sysctl keys are valid or not. +// /proc/sys isn't completely namespaced and depending on which namespaces +// are specified, a subset of sysctls are permitted. +func (v *ConfigValidator) sysctl(config *configs.Config) error { + validSysctlMap := map[string]bool{ + "kernel.msgmax": true, + "kernel.msgmnb": true, + "kernel.msgmni": true, + "kernel.sem": true, + "kernel.shmall": true, + "kernel.shmmax": true, + "kernel.shmmni": true, + "kernel.shm_rmid_forced": true, + } + + for s := range config.Sysctl { + if validSysctlMap[s] || strings.HasPrefix(s, "fs.mqueue.") { + if config.Namespaces.Contains(configs.NEWIPC) { + continue + } else { + return fmt.Errorf("sysctl %q is not allowed in the hosts ipc namespace", s) + } + } + if strings.HasPrefix(s, "net.") { + if config.Namespaces.Contains(configs.NEWNET) { + if path := config.Namespaces.PathOf(configs.NEWNET); path != "" { + if err := checkHostNs(s, path); err != nil { + return err + } + } + continue + } else { + return fmt.Errorf("sysctl %q is not allowed in the hosts network namespace", s) + } + } + if config.Namespaces.Contains(configs.NEWUTS) { + switch s { + case "kernel.domainname": + // This is namespaced and there's no explicit OCI field for it. + continue + case "kernel.hostname": + // This is namespaced but there's a conflicting (dedicated) OCI field for it. + return fmt.Errorf("sysctl %q is not allowed as it conflicts with the OCI %q field", s, "hostname") + } + } + return fmt.Errorf("sysctl %q is not in a separate kernel namespace", s) + } + + return nil +} + +func (v *ConfigValidator) intelrdt(config *configs.Config) error { + if config.IntelRdt != nil { + if !intelrdt.IsEnabled() { + return fmt.Errorf("intelRdt is specified in config, but Intel RDT feature is not supported or enabled") + } + if config.IntelRdt.L3CacheSchema == "" { + return fmt.Errorf("intelRdt is specified in config, but intelRdt.l3CacheSchema is empty") + } + } + + return nil +} + +func isSymbolicLink(path string) (bool, error) { + fi, err := os.Lstat(path) + if err != nil { + return false, err + } + + return fi.Mode()&os.ModeSymlink == os.ModeSymlink, nil +} + +// checkHostNs checks whether network sysctl is used in host namespace. +func checkHostNs(sysctlConfig string, path string) error { + var currentProcessNetns = "/proc/self/ns/net" + // readlink on the current processes network namespace + destOfCurrentProcess, err := os.Readlink(currentProcessNetns) + if err != nil { + return fmt.Errorf("read soft link %q error", currentProcessNetns) + } + + // First check if the provided path is a symbolic link + symLink, err := isSymbolicLink(path) + if err != nil { + return fmt.Errorf("could not check that %q is a symlink: %v", path, err) + } + + if symLink == false { + // The provided namespace is not a symbolic link, + // it is not the host namespace. + return nil + } + + // readlink on the path provided in the struct + destOfContainer, err := os.Readlink(path) + if err != nil { + return fmt.Errorf("read soft link %q error", path) + } + if destOfContainer == destOfCurrentProcess { + return fmt.Errorf("sysctl %q is not allowed in the hosts network namespace", sysctlConfig) + } + return nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/console_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/console_linux.go new file mode 100644 index 000000000..9997e93ed --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/console_linux.go @@ -0,0 +1,41 @@ +package libcontainer + +import ( + "os" + + "golang.org/x/sys/unix" +) + +// mount initializes the console inside the rootfs mounting with the specified mount label +// and applying the correct ownership of the console. +func mountConsole(slavePath string) error { + oldMask := unix.Umask(0000) + defer unix.Umask(oldMask) + f, err := os.Create("/dev/console") + if err != nil && !os.IsExist(err) { + return err + } + if f != nil { + f.Close() + } + return unix.Mount(slavePath, "/dev/console", "bind", unix.MS_BIND, "") +} + +// dupStdio opens the slavePath for the console and dups the fds to the current +// processes stdio, fd 0,1,2. +func dupStdio(slavePath string) error { + fd, err := unix.Open(slavePath, unix.O_RDWR, 0) + if err != nil { + return &os.PathError{ + Op: "open", + Path: slavePath, + Err: err, + } + } + for _, i := range []int{0, 1, 2} { + if err := unix.Dup3(fd, i, 0); err != nil { + return err + } + } + return nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/container.go b/vendor/github.com/opencontainers/runc/libcontainer/container.go new file mode 100644 index 000000000..2e31b4d4f --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/container.go @@ -0,0 +1,166 @@ +// Package libcontainer provides a native Go implementation for creating containers +// with namespaces, cgroups, capabilities, and filesystem access controls. +// It allows you to manage the lifecycle of the container performing additional operations +// after the container is created. +package libcontainer + +import ( + "os" + "time" + + "github.com/opencontainers/runc/libcontainer/configs" +) + +// Status is the status of a container. +type Status int + +const ( + // Created is the status that denotes the container exists but has not been run yet. + Created Status = iota + // Running is the status that denotes the container exists and is running. + Running + // Pausing is the status that denotes the container exists, it is in the process of being paused. + Pausing + // Paused is the status that denotes the container exists, but all its processes are paused. + Paused + // Stopped is the status that denotes the container does not have a created or running process. + Stopped +) + +func (s Status) String() string { + switch s { + case Created: + return "created" + case Running: + return "running" + case Pausing: + return "pausing" + case Paused: + return "paused" + case Stopped: + return "stopped" + default: + return "unknown" + } +} + +// BaseState represents the platform agnostic pieces relating to a +// running container's state +type BaseState struct { + // ID is the container ID. + ID string `json:"id"` + + // InitProcessPid is the init process id in the parent namespace. + InitProcessPid int `json:"init_process_pid"` + + // InitProcessStartTime is the init process start time in clock cycles since boot time. + InitProcessStartTime uint64 `json:"init_process_start"` + + // Created is the unix timestamp for the creation time of the container in UTC + Created time.Time `json:"created"` + + // Config is the container's configuration. + Config configs.Config `json:"config"` +} + +// BaseContainer is a libcontainer container object. +// +// Each container is thread-safe within the same process. Since a container can +// be destroyed by a separate process, any function may return that the container +// was not found. BaseContainer includes methods that are platform agnostic. +type BaseContainer interface { + // Returns the ID of the container + ID() string + + // Returns the current status of the container. + // + // errors: + // ContainerNotExists - Container no longer exists, + // Systemerror - System error. + Status() (Status, error) + + // State returns the current container's state information. + // + // errors: + // SystemError - System error. + State() (*State, error) + + // Returns the current config of the container. + Config() configs.Config + + // Returns the PIDs inside this container. The PIDs are in the namespace of the calling process. + // + // errors: + // ContainerNotExists - Container no longer exists, + // Systemerror - System error. + // + // Some of the returned PIDs may no longer refer to processes in the Container, unless + // the Container state is PAUSED in which case every PID in the slice is valid. + Processes() ([]int, error) + + // Returns statistics for the container. + // + // errors: + // ContainerNotExists - Container no longer exists, + // Systemerror - System error. + Stats() (*Stats, error) + + // Set resources of container as configured + // + // We can use this to change resources when containers are running. + // + // errors: + // SystemError - System error. + Set(config configs.Config) error + + // Start a process inside the container. Returns error if process fails to + // start. You can track process lifecycle with passed Process structure. + // + // errors: + // ContainerNotExists - Container no longer exists, + // ConfigInvalid - config is invalid, + // ContainerPaused - Container is paused, + // SystemError - System error. + Start(process *Process) (err error) + + // Run immediately starts the process inside the container. Returns error if process + // fails to start. It does not block waiting for the exec fifo after start returns but + // opens the fifo after start returns. + // + // errors: + // ContainerNotExists - Container no longer exists, + // ConfigInvalid - config is invalid, + // ContainerPaused - Container is paused, + // SystemError - System error. + Run(process *Process) (err error) + + // Destroys the container, if its in a valid state, after killing any + // remaining running processes. + // + // Any event registrations are removed before the container is destroyed. + // No error is returned if the container is already destroyed. + // + // Running containers must first be stopped using Signal(..). + // Paused containers must first be resumed using Resume(..). + // + // errors: + // ContainerNotStopped - Container is still running, + // ContainerPaused - Container is paused, + // SystemError - System error. + Destroy() error + + // Signal sends the provided signal code to the container's initial process. + // + // If all is specified the signal is sent to all processes in the container + // including the initial process. + // + // errors: + // SystemError - System error. + Signal(s os.Signal, all bool) error + + // Exec signals the container to exec the users process at the end of the init. + // + // errors: + // SystemError - System error. + Exec() error +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/container_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/container_linux.go new file mode 100644 index 000000000..9f551fd09 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/container_linux.go @@ -0,0 +1,1899 @@ +// +build linux + +package libcontainer + +import ( + "bytes" + "encoding/json" + "errors" + "fmt" + "io" + "io/ioutil" + "net" + "os" + "os/exec" + "path/filepath" + "reflect" + "strings" + "sync" + "syscall" // only for SysProcAttr and Signal + "time" + + "github.com/opencontainers/runc/libcontainer/cgroups" + "github.com/opencontainers/runc/libcontainer/configs" + "github.com/opencontainers/runc/libcontainer/criurpc" + "github.com/opencontainers/runc/libcontainer/intelrdt" + "github.com/opencontainers/runc/libcontainer/system" + "github.com/opencontainers/runc/libcontainer/utils" + + "github.com/golang/protobuf/proto" + "github.com/sirupsen/logrus" + "github.com/vishvananda/netlink/nl" + "golang.org/x/sys/unix" +) + +const stdioFdCount = 3 + +type linuxContainer struct { + id string + root string + config *configs.Config + cgroupManager cgroups.Manager + intelRdtManager intelrdt.Manager + initPath string + initArgs []string + initProcess parentProcess + initProcessStartTime uint64 + criuPath string + newuidmapPath string + newgidmapPath string + m sync.Mutex + criuVersion int + state containerState + created time.Time +} + +// State represents a running container's state +type State struct { + BaseState + + // Platform specific fields below here + + // Specifies if the container was started under the rootless mode. + Rootless bool `json:"rootless"` + + // Path to all the cgroups setup for a container. Key is cgroup subsystem name + // with the value as the path. + CgroupPaths map[string]string `json:"cgroup_paths"` + + // NamespacePaths are filepaths to the container's namespaces. Key is the namespace type + // with the value as the path. + NamespacePaths map[configs.NamespaceType]string `json:"namespace_paths"` + + // Container's standard descriptors (std{in,out,err}), needed for checkpoint and restore + ExternalDescriptors []string `json:"external_descriptors,omitempty"` + + // Intel RDT "resource control" filesystem path + IntelRdtPath string `json:"intel_rdt_path"` +} + +// Container is a libcontainer container object. +// +// Each container is thread-safe within the same process. Since a container can +// be destroyed by a separate process, any function may return that the container +// was not found. +type Container interface { + BaseContainer + + // Methods below here are platform specific + + // Checkpoint checkpoints the running container's state to disk using the criu(8) utility. + // + // errors: + // Systemerror - System error. + Checkpoint(criuOpts *CriuOpts) error + + // Restore restores the checkpointed container to a running state using the criu(8) utility. + // + // errors: + // Systemerror - System error. + Restore(process *Process, criuOpts *CriuOpts) error + + // If the Container state is RUNNING or CREATED, sets the Container state to PAUSING and pauses + // the execution of any user processes. Asynchronously, when the container finished being paused the + // state is changed to PAUSED. + // If the Container state is PAUSED, do nothing. + // + // errors: + // ContainerNotExists - Container no longer exists, + // ContainerNotRunning - Container not running or created, + // Systemerror - System error. + Pause() error + + // If the Container state is PAUSED, resumes the execution of any user processes in the + // Container before setting the Container state to RUNNING. + // If the Container state is RUNNING, do nothing. + // + // errors: + // ContainerNotExists - Container no longer exists, + // ContainerNotPaused - Container is not paused, + // Systemerror - System error. + Resume() error + + // NotifyOOM returns a read-only channel signaling when the container receives an OOM notification. + // + // errors: + // Systemerror - System error. + NotifyOOM() (<-chan struct{}, error) + + // NotifyMemoryPressure returns a read-only channel signaling when the container reaches a given pressure level + // + // errors: + // Systemerror - System error. + NotifyMemoryPressure(level PressureLevel) (<-chan struct{}, error) +} + +// ID returns the container's unique ID +func (c *linuxContainer) ID() string { + return c.id +} + +// Config returns the container's configuration +func (c *linuxContainer) Config() configs.Config { + return *c.config +} + +func (c *linuxContainer) Status() (Status, error) { + c.m.Lock() + defer c.m.Unlock() + return c.currentStatus() +} + +func (c *linuxContainer) State() (*State, error) { + c.m.Lock() + defer c.m.Unlock() + return c.currentState() +} + +func (c *linuxContainer) Processes() ([]int, error) { + pids, err := c.cgroupManager.GetAllPids() + if err != nil { + return nil, newSystemErrorWithCause(err, "getting all container pids from cgroups") + } + return pids, nil +} + +func (c *linuxContainer) Stats() (*Stats, error) { + var ( + err error + stats = &Stats{} + ) + if stats.CgroupStats, err = c.cgroupManager.GetStats(); err != nil { + return stats, newSystemErrorWithCause(err, "getting container stats from cgroups") + } + if c.intelRdtManager != nil { + if stats.IntelRdtStats, err = c.intelRdtManager.GetStats(); err != nil { + return stats, newSystemErrorWithCause(err, "getting container's Intel RDT stats") + } + } + for _, iface := range c.config.Networks { + switch iface.Type { + case "veth": + istats, err := getNetworkInterfaceStats(iface.HostInterfaceName) + if err != nil { + return stats, newSystemErrorWithCausef(err, "getting network stats for interface %q", iface.HostInterfaceName) + } + stats.Interfaces = append(stats.Interfaces, istats) + } + } + return stats, nil +} + +func (c *linuxContainer) Set(config configs.Config) error { + c.m.Lock() + defer c.m.Unlock() + status, err := c.currentStatus() + if err != nil { + return err + } + if status == Stopped { + return newGenericError(fmt.Errorf("container not running"), ContainerNotRunning) + } + if err := c.cgroupManager.Set(&config); err != nil { + // Set configs back + if err2 := c.cgroupManager.Set(c.config); err2 != nil { + logrus.Warnf("Setting back cgroup configs failed due to error: %v, your state.json and actual configs might be inconsistent.", err2) + } + return err + } + if c.intelRdtManager != nil { + if err := c.intelRdtManager.Set(&config); err != nil { + // Set configs back + if err2 := c.intelRdtManager.Set(c.config); err2 != nil { + logrus.Warnf("Setting back intelrdt configs failed due to error: %v, your state.json and actual configs might be inconsistent.", err2) + } + return err + } + } + // After config setting succeed, update config and states + c.config = &config + _, err = c.updateState(nil) + return err +} + +func (c *linuxContainer) Start(process *Process) error { + c.m.Lock() + defer c.m.Unlock() + if process.Init { + if err := c.createExecFifo(); err != nil { + return err + } + } + if err := c.start(process); err != nil { + if process.Init { + c.deleteExecFifo() + } + return err + } + return nil +} + +func (c *linuxContainer) Run(process *Process) error { + if err := c.Start(process); err != nil { + return err + } + if process.Init { + return c.exec() + } + return nil +} + +func (c *linuxContainer) Exec() error { + c.m.Lock() + defer c.m.Unlock() + return c.exec() +} + +func (c *linuxContainer) exec() error { + path := filepath.Join(c.root, execFifoFilename) + + fifoOpen := make(chan struct{}) + select { + case <-awaitProcessExit(c.initProcess.pid(), fifoOpen): + return errors.New("container process is already dead") + case result := <-awaitFifoOpen(path): + close(fifoOpen) + if result.err != nil { + return result.err + } + f := result.file + defer f.Close() + if err := readFromExecFifo(f); err != nil { + return err + } + return os.Remove(path) + } +} + +func readFromExecFifo(execFifo io.Reader) error { + data, err := ioutil.ReadAll(execFifo) + if err != nil { + return err + } + if len(data) <= 0 { + return fmt.Errorf("cannot start an already running container") + } + return nil +} + +func awaitProcessExit(pid int, exit <-chan struct{}) <-chan struct{} { + isDead := make(chan struct{}) + go func() { + for { + select { + case <-exit: + return + case <-time.After(time.Millisecond * 100): + stat, err := system.Stat(pid) + if err != nil || stat.State == system.Zombie { + close(isDead) + return + } + } + } + }() + return isDead +} + +func awaitFifoOpen(path string) <-chan openResult { + fifoOpened := make(chan openResult) + go func() { + f, err := os.OpenFile(path, os.O_RDONLY, 0) + if err != nil { + fifoOpened <- openResult{err: newSystemErrorWithCause(err, "open exec fifo for reading")} + return + } + fifoOpened <- openResult{file: f} + }() + return fifoOpened +} + +type openResult struct { + file *os.File + err error +} + +func (c *linuxContainer) start(process *Process) error { + parent, err := c.newParentProcess(process) + if err != nil { + return newSystemErrorWithCause(err, "creating new parent process") + } + if err := parent.start(); err != nil { + // terminate the process to ensure that it properly is reaped. + if err := ignoreTerminateErrors(parent.terminate()); err != nil { + logrus.Warn(err) + } + return newSystemErrorWithCause(err, "starting container process") + } + // generate a timestamp indicating when the container was started + c.created = time.Now().UTC() + if process.Init { + c.state = &createdState{ + c: c, + } + state, err := c.updateState(parent) + if err != nil { + return err + } + c.initProcessStartTime = state.InitProcessStartTime + + if c.config.Hooks != nil { + bundle, annotations := utils.Annotations(c.config.Labels) + s := configs.HookState{ + Version: c.config.Version, + ID: c.id, + Pid: parent.pid(), + Bundle: bundle, + Annotations: annotations, + } + for i, hook := range c.config.Hooks.Poststart { + if err := hook.Run(s); err != nil { + if err := ignoreTerminateErrors(parent.terminate()); err != nil { + logrus.Warn(err) + } + return newSystemErrorWithCausef(err, "running poststart hook %d", i) + } + } + } + } + return nil +} + +func (c *linuxContainer) Signal(s os.Signal, all bool) error { + if all { + return signalAllProcesses(c.cgroupManager, s) + } + if err := c.initProcess.signal(s); err != nil { + return newSystemErrorWithCause(err, "signaling init process") + } + return nil +} + +func (c *linuxContainer) createExecFifo() error { + rootuid, err := c.Config().HostRootUID() + if err != nil { + return err + } + rootgid, err := c.Config().HostRootGID() + if err != nil { + return err + } + + fifoName := filepath.Join(c.root, execFifoFilename) + if _, err := os.Stat(fifoName); err == nil { + return fmt.Errorf("exec fifo %s already exists", fifoName) + } + oldMask := unix.Umask(0000) + if err := unix.Mkfifo(fifoName, 0622); err != nil { + unix.Umask(oldMask) + return err + } + unix.Umask(oldMask) + if err := os.Chown(fifoName, rootuid, rootgid); err != nil { + return err + } + return nil +} + +func (c *linuxContainer) deleteExecFifo() { + fifoName := filepath.Join(c.root, execFifoFilename) + os.Remove(fifoName) +} + +// includeExecFifo opens the container's execfifo as a pathfd, so that the +// container cannot access the statedir (and the FIFO itself remains +// un-opened). It then adds the FifoFd to the given exec.Cmd as an inherited +// fd, with _LIBCONTAINER_FIFOFD set to its fd number. +func (c *linuxContainer) includeExecFifo(cmd *exec.Cmd) error { + fifoName := filepath.Join(c.root, execFifoFilename) + fifoFd, err := unix.Open(fifoName, unix.O_PATH|unix.O_CLOEXEC, 0) + if err != nil { + return err + } + + cmd.ExtraFiles = append(cmd.ExtraFiles, os.NewFile(uintptr(fifoFd), fifoName)) + cmd.Env = append(cmd.Env, + fmt.Sprintf("_LIBCONTAINER_FIFOFD=%d", stdioFdCount+len(cmd.ExtraFiles)-1)) + return nil +} + +func (c *linuxContainer) newParentProcess(p *Process) (parentProcess, error) { + parentPipe, childPipe, err := utils.NewSockPair("init") + if err != nil { + return nil, newSystemErrorWithCause(err, "creating new init pipe") + } + cmd, err := c.commandTemplate(p, childPipe) + if err != nil { + return nil, newSystemErrorWithCause(err, "creating new command template") + } + if !p.Init { + return c.newSetnsProcess(p, cmd, parentPipe, childPipe) + } + + // We only set up fifoFd if we're not doing a `runc exec`. The historic + // reason for this is that previously we would pass a dirfd that allowed + // for container rootfs escape (and not doing it in `runc exec` avoided + // that problem), but we no longer do that. However, there's no need to do + // this for `runc exec` so we just keep it this way to be safe. + if err := c.includeExecFifo(cmd); err != nil { + return nil, newSystemErrorWithCause(err, "including execfifo in cmd.Exec setup") + } + return c.newInitProcess(p, cmd, parentPipe, childPipe) +} + +func (c *linuxContainer) commandTemplate(p *Process, childPipe *os.File) (*exec.Cmd, error) { + cmd := exec.Command(c.initPath, c.initArgs[1:]...) + cmd.Args[0] = c.initArgs[0] + cmd.Stdin = p.Stdin + cmd.Stdout = p.Stdout + cmd.Stderr = p.Stderr + cmd.Dir = c.config.Rootfs + if cmd.SysProcAttr == nil { + cmd.SysProcAttr = &syscall.SysProcAttr{} + } + cmd.Env = append(cmd.Env, fmt.Sprintf("GOMAXPROCS=%s", os.Getenv("GOMAXPROCS"))) + cmd.ExtraFiles = append(cmd.ExtraFiles, p.ExtraFiles...) + if p.ConsoleSocket != nil { + cmd.ExtraFiles = append(cmd.ExtraFiles, p.ConsoleSocket) + cmd.Env = append(cmd.Env, + fmt.Sprintf("_LIBCONTAINER_CONSOLE=%d", stdioFdCount+len(cmd.ExtraFiles)-1), + ) + } + cmd.ExtraFiles = append(cmd.ExtraFiles, childPipe) + cmd.Env = append(cmd.Env, + fmt.Sprintf("_LIBCONTAINER_INITPIPE=%d", stdioFdCount+len(cmd.ExtraFiles)-1), + ) + // NOTE: when running a container with no PID namespace and the parent process spawning the container is + // PID1 the pdeathsig is being delivered to the container's init process by the kernel for some reason + // even with the parent still running. + if c.config.ParentDeathSignal > 0 { + cmd.SysProcAttr.Pdeathsig = syscall.Signal(c.config.ParentDeathSignal) + } + return cmd, nil +} + +func (c *linuxContainer) newInitProcess(p *Process, cmd *exec.Cmd, parentPipe, childPipe *os.File) (*initProcess, error) { + cmd.Env = append(cmd.Env, "_LIBCONTAINER_INITTYPE="+string(initStandard)) + nsMaps := make(map[configs.NamespaceType]string) + for _, ns := range c.config.Namespaces { + if ns.Path != "" { + nsMaps[ns.Type] = ns.Path + } + } + _, sharePidns := nsMaps[configs.NEWPID] + data, err := c.bootstrapData(c.config.Namespaces.CloneFlags(), nsMaps) + if err != nil { + return nil, err + } + return &initProcess{ + cmd: cmd, + childPipe: childPipe, + parentPipe: parentPipe, + manager: c.cgroupManager, + intelRdtManager: c.intelRdtManager, + config: c.newInitConfig(p), + container: c, + process: p, + bootstrapData: data, + sharePidns: sharePidns, + }, nil +} + +func (c *linuxContainer) newSetnsProcess(p *Process, cmd *exec.Cmd, parentPipe, childPipe *os.File) (*setnsProcess, error) { + cmd.Env = append(cmd.Env, "_LIBCONTAINER_INITTYPE="+string(initSetns)) + state, err := c.currentState() + if err != nil { + return nil, newSystemErrorWithCause(err, "getting container's current state") + } + // for setns process, we don't have to set cloneflags as the process namespaces + // will only be set via setns syscall + data, err := c.bootstrapData(0, state.NamespacePaths) + if err != nil { + return nil, err + } + return &setnsProcess{ + cmd: cmd, + cgroupPaths: c.cgroupManager.GetPaths(), + intelRdtPath: state.IntelRdtPath, + childPipe: childPipe, + parentPipe: parentPipe, + config: c.newInitConfig(p), + process: p, + bootstrapData: data, + }, nil +} + +func (c *linuxContainer) newInitConfig(process *Process) *initConfig { + cfg := &initConfig{ + Config: c.config, + Args: process.Args, + Env: process.Env, + User: process.User, + AdditionalGroups: process.AdditionalGroups, + Cwd: process.Cwd, + Capabilities: process.Capabilities, + PassedFilesCount: len(process.ExtraFiles), + ContainerId: c.ID(), + NoNewPrivileges: c.config.NoNewPrivileges, + Rootless: c.config.Rootless, + AppArmorProfile: c.config.AppArmorProfile, + ProcessLabel: c.config.ProcessLabel, + Rlimits: c.config.Rlimits, + } + if process.NoNewPrivileges != nil { + cfg.NoNewPrivileges = *process.NoNewPrivileges + } + if process.AppArmorProfile != "" { + cfg.AppArmorProfile = process.AppArmorProfile + } + if process.Label != "" { + cfg.ProcessLabel = process.Label + } + if len(process.Rlimits) > 0 { + cfg.Rlimits = process.Rlimits + } + cfg.CreateConsole = process.ConsoleSocket != nil + cfg.ConsoleWidth = process.ConsoleWidth + cfg.ConsoleHeight = process.ConsoleHeight + return cfg +} + +func (c *linuxContainer) Destroy() error { + c.m.Lock() + defer c.m.Unlock() + return c.state.destroy() +} + +func (c *linuxContainer) Pause() error { + c.m.Lock() + defer c.m.Unlock() + status, err := c.currentStatus() + if err != nil { + return err + } + switch status { + case Running, Created: + if err := c.cgroupManager.Freeze(configs.Frozen); err != nil { + return err + } + return c.state.transition(&pausedState{ + c: c, + }) + } + return newGenericError(fmt.Errorf("container not running or created: %s", status), ContainerNotRunning) +} + +func (c *linuxContainer) Resume() error { + c.m.Lock() + defer c.m.Unlock() + status, err := c.currentStatus() + if err != nil { + return err + } + if status != Paused { + return newGenericError(fmt.Errorf("container not paused"), ContainerNotPaused) + } + if err := c.cgroupManager.Freeze(configs.Thawed); err != nil { + return err + } + return c.state.transition(&runningState{ + c: c, + }) +} + +func (c *linuxContainer) NotifyOOM() (<-chan struct{}, error) { + // XXX(cyphar): This requires cgroups. + if c.config.Rootless { + return nil, fmt.Errorf("cannot get OOM notifications from rootless container") + } + return notifyOnOOM(c.cgroupManager.GetPaths()) +} + +func (c *linuxContainer) NotifyMemoryPressure(level PressureLevel) (<-chan struct{}, error) { + // XXX(cyphar): This requires cgroups. + if c.config.Rootless { + return nil, fmt.Errorf("cannot get memory pressure notifications from rootless container") + } + return notifyMemoryPressure(c.cgroupManager.GetPaths(), level) +} + +var criuFeatures *criurpc.CriuFeatures + +func (c *linuxContainer) checkCriuFeatures(criuOpts *CriuOpts, rpcOpts *criurpc.CriuOpts, criuFeat *criurpc.CriuFeatures) error { + + var t criurpc.CriuReqType + t = criurpc.CriuReqType_FEATURE_CHECK + + // criu 1.8 => 10800 + if err := c.checkCriuVersion(10800); err != nil { + // Feature checking was introduced with CRIU 1.8. + // Ignore the feature check if an older CRIU version is used + // and just act as before. + // As all automated PR testing is done using CRIU 1.7 this + // code will not be tested by automated PR testing. + return nil + } + + // make sure the features we are looking for are really not from + // some previous check + criuFeatures = nil + + req := &criurpc.CriuReq{ + Type: &t, + // Theoretically this should not be necessary but CRIU + // segfaults if Opts is empty. + // Fixed in CRIU 2.12 + Opts: rpcOpts, + Features: criuFeat, + } + + err := c.criuSwrk(nil, req, criuOpts, false, nil) + if err != nil { + logrus.Debugf("%s", err) + return fmt.Errorf("CRIU feature check failed") + } + + logrus.Debugf("Feature check says: %s", criuFeatures) + missingFeatures := false + + // The outer if checks if the fields actually exist + if (criuFeat.MemTrack != nil) && + (criuFeatures.MemTrack != nil) { + // The inner if checks if they are set to true + if *criuFeat.MemTrack && !*criuFeatures.MemTrack { + missingFeatures = true + logrus.Debugf("CRIU does not support MemTrack") + } + } + + // This needs to be repeated for every new feature check. + // Is there a way to put this in a function. Reflection? + if (criuFeat.LazyPages != nil) && + (criuFeatures.LazyPages != nil) { + if *criuFeat.LazyPages && !*criuFeatures.LazyPages { + missingFeatures = true + logrus.Debugf("CRIU does not support LazyPages") + } + } + + if missingFeatures { + return fmt.Errorf("CRIU is missing features") + } + + return nil +} + +func parseCriuVersion(path string) (int, error) { + var x, y, z int + + out, err := exec.Command(path, "-V").Output() + if err != nil { + return 0, fmt.Errorf("Unable to execute CRIU command: %s", path) + } + + x = 0 + y = 0 + z = 0 + if ep := strings.Index(string(out), "-"); ep >= 0 { + // criu Git version format + var version string + if sp := strings.Index(string(out), "GitID"); sp > 0 { + version = string(out)[sp:ep] + } else { + return 0, fmt.Errorf("Unable to parse the CRIU version: %s", path) + } + + n, err := fmt.Sscanf(version, "GitID: v%d.%d.%d", &x, &y, &z) // 1.5.2 + if err != nil { + n, err = fmt.Sscanf(version, "GitID: v%d.%d", &x, &y) // 1.6 + y++ + } else { + z++ + } + if n < 2 || err != nil { + return 0, fmt.Errorf("Unable to parse the CRIU version: %s %d %s", version, n, err) + } + } else { + // criu release version format + n, err := fmt.Sscanf(string(out), "Version: %d.%d.%d\n", &x, &y, &z) // 1.5.2 + if err != nil { + n, err = fmt.Sscanf(string(out), "Version: %d.%d\n", &x, &y) // 1.6 + } + if n < 2 || err != nil { + return 0, fmt.Errorf("Unable to parse the CRIU version: %s %d %s", out, n, err) + } + } + + return x*10000 + y*100 + z, nil +} + +func compareCriuVersion(criuVersion int, minVersion int) error { + // simple function to perform the actual version compare + if criuVersion < minVersion { + return fmt.Errorf("CRIU version %d must be %d or higher", criuVersion, minVersion) + } + + return nil +} + +// This is used to store the result of criu version RPC +var criuVersionRPC *criurpc.CriuVersion + +// checkCriuVersion checks Criu version greater than or equal to minVersion +func (c *linuxContainer) checkCriuVersion(minVersion int) error { + + // If the version of criu has already been determined there is no need + // to ask criu for the version again. Use the value from c.criuVersion. + if c.criuVersion != 0 { + return compareCriuVersion(c.criuVersion, minVersion) + } + + // First try if this version of CRIU support the version RPC. + // The CRIU version RPC was introduced with CRIU 3.0. + + // First, reset the variable for the RPC answer to nil + criuVersionRPC = nil + + var t criurpc.CriuReqType + t = criurpc.CriuReqType_VERSION + req := &criurpc.CriuReq{ + Type: &t, + } + + err := c.criuSwrk(nil, req, nil, false, nil) + if err != nil { + return fmt.Errorf("CRIU version check failed: %s", err) + } + + if criuVersionRPC != nil { + logrus.Debugf("CRIU version: %s", criuVersionRPC) + // major and minor are always set + c.criuVersion = int(*criuVersionRPC.Major) * 10000 + c.criuVersion += int(*criuVersionRPC.Minor) * 100 + if criuVersionRPC.Sublevel != nil { + c.criuVersion += int(*criuVersionRPC.Sublevel) + } + if criuVersionRPC.Gitid != nil { + // runc's convention is that a CRIU git release is + // always the same as increasing the minor by 1 + c.criuVersion -= (c.criuVersion % 100) + c.criuVersion += 100 + } + return compareCriuVersion(c.criuVersion, minVersion) + } + + // This is CRIU without the version RPC and therefore + // older than 3.0. Parsing the output is required. + + // This can be remove once runc does not work with criu older than 3.0 + + c.criuVersion, err = parseCriuVersion(c.criuPath) + if err != nil { + return err + } + + return compareCriuVersion(c.criuVersion, minVersion) +} + +const descriptorsFilename = "descriptors.json" + +func (c *linuxContainer) addCriuDumpMount(req *criurpc.CriuReq, m *configs.Mount) { + mountDest := m.Destination + if strings.HasPrefix(mountDest, c.config.Rootfs) { + mountDest = mountDest[len(c.config.Rootfs):] + } + + extMnt := &criurpc.ExtMountMap{ + Key: proto.String(mountDest), + Val: proto.String(mountDest), + } + req.Opts.ExtMnt = append(req.Opts.ExtMnt, extMnt) +} + +func (c *linuxContainer) addMaskPaths(req *criurpc.CriuReq) error { + for _, path := range c.config.MaskPaths { + fi, err := os.Stat(fmt.Sprintf("/proc/%d/root/%s", c.initProcess.pid(), path)) + if err != nil { + if os.IsNotExist(err) { + continue + } + return err + } + if fi.IsDir() { + continue + } + + extMnt := &criurpc.ExtMountMap{ + Key: proto.String(path), + Val: proto.String("/dev/null"), + } + req.Opts.ExtMnt = append(req.Opts.ExtMnt, extMnt) + } + return nil +} + +func waitForCriuLazyServer(r *os.File, status string) error { + + data := make([]byte, 1) + _, err := r.Read(data) + if err != nil { + return err + } + fd, err := os.OpenFile(status, os.O_TRUNC|os.O_WRONLY, os.ModeAppend) + if err != nil { + return err + } + _, err = fd.Write(data) + if err != nil { + return err + } + fd.Close() + + return nil +} + +func (c *linuxContainer) Checkpoint(criuOpts *CriuOpts) error { + c.m.Lock() + defer c.m.Unlock() + + // TODO(avagin): Figure out how to make this work nicely. CRIU 2.0 has + // support for doing unprivileged dumps, but the setup of + // rootless containers might make this complicated. + if c.config.Rootless { + return fmt.Errorf("cannot checkpoint a rootless container") + } + + // criu 1.5.2 => 10502 + if err := c.checkCriuVersion(10502); err != nil { + return err + } + + if criuOpts.ImagesDirectory == "" { + return fmt.Errorf("invalid directory to save checkpoint") + } + + // Since a container can be C/R'ed multiple times, + // the checkpoint directory may already exist. + if err := os.Mkdir(criuOpts.ImagesDirectory, 0755); err != nil && !os.IsExist(err) { + return err + } + + if criuOpts.WorkDirectory == "" { + criuOpts.WorkDirectory = filepath.Join(c.root, "criu.work") + } + + if err := os.Mkdir(criuOpts.WorkDirectory, 0755); err != nil && !os.IsExist(err) { + return err + } + + workDir, err := os.Open(criuOpts.WorkDirectory) + if err != nil { + return err + } + defer workDir.Close() + + imageDir, err := os.Open(criuOpts.ImagesDirectory) + if err != nil { + return err + } + defer imageDir.Close() + + rpcOpts := criurpc.CriuOpts{ + ImagesDirFd: proto.Int32(int32(imageDir.Fd())), + WorkDirFd: proto.Int32(int32(workDir.Fd())), + LogLevel: proto.Int32(4), + LogFile: proto.String("dump.log"), + Root: proto.String(c.config.Rootfs), + ManageCgroups: proto.Bool(true), + NotifyScripts: proto.Bool(true), + Pid: proto.Int32(int32(c.initProcess.pid())), + ShellJob: proto.Bool(criuOpts.ShellJob), + LeaveRunning: proto.Bool(criuOpts.LeaveRunning), + TcpEstablished: proto.Bool(criuOpts.TcpEstablished), + ExtUnixSk: proto.Bool(criuOpts.ExternalUnixConnections), + FileLocks: proto.Bool(criuOpts.FileLocks), + EmptyNs: proto.Uint32(criuOpts.EmptyNs), + OrphanPtsMaster: proto.Bool(true), + AutoDedup: proto.Bool(criuOpts.AutoDedup), + LazyPages: proto.Bool(criuOpts.LazyPages), + } + + // If the container is running in a network namespace and has + // a path to the network namespace configured, we will dump + // that network namespace as an external namespace and we + // will expect that the namespace exists during restore. + // This basically means that CRIU will ignore the namespace + // and expect to be setup correctly. + nsPath := c.config.Namespaces.PathOf(configs.NEWNET) + if nsPath != "" { + // For this to work we need at least criu 3.11.0 => 31100. + // As there was already a successful version check we will + // not error out if it fails. runc will just behave as it used + // to do and ignore external network namespaces. + err := c.checkCriuVersion(31100) + if err == nil { + // CRIU expects the information about an external namespace + // like this: --external net[]: + // This is always 'extRootNetNS'. + var netns syscall.Stat_t + err = syscall.Stat(nsPath, &netns) + if err != nil { + return err + } + criuExternal := fmt.Sprintf("net[%d]:extRootNetNS", netns.Ino) + rpcOpts.External = append(rpcOpts.External, criuExternal) + } + } + + fcg := c.cgroupManager.GetPaths()["freezer"] + if fcg != "" { + rpcOpts.FreezeCgroup = proto.String(fcg) + } + + // append optional criu opts, e.g., page-server and port + if criuOpts.PageServer.Address != "" && criuOpts.PageServer.Port != 0 { + rpcOpts.Ps = &criurpc.CriuPageServerInfo{ + Address: proto.String(criuOpts.PageServer.Address), + Port: proto.Int32(criuOpts.PageServer.Port), + } + } + + //pre-dump may need parentImage param to complete iterative migration + if criuOpts.ParentImage != "" { + rpcOpts.ParentImg = proto.String(criuOpts.ParentImage) + rpcOpts.TrackMem = proto.Bool(true) + } + + // append optional manage cgroups mode + if criuOpts.ManageCgroupsMode != 0 { + // criu 1.7 => 10700 + if err := c.checkCriuVersion(10700); err != nil { + return err + } + mode := criurpc.CriuCgMode(criuOpts.ManageCgroupsMode) + rpcOpts.ManageCgroupsMode = &mode + } + + var t criurpc.CriuReqType + if criuOpts.PreDump { + feat := criurpc.CriuFeatures{ + MemTrack: proto.Bool(true), + } + + if err := c.checkCriuFeatures(criuOpts, &rpcOpts, &feat); err != nil { + return err + } + + t = criurpc.CriuReqType_PRE_DUMP + } else { + t = criurpc.CriuReqType_DUMP + } + req := &criurpc.CriuReq{ + Type: &t, + Opts: &rpcOpts, + } + + if criuOpts.LazyPages { + // lazy migration requested; check if criu supports it + feat := criurpc.CriuFeatures{ + LazyPages: proto.Bool(true), + } + + if err := c.checkCriuFeatures(criuOpts, &rpcOpts, &feat); err != nil { + return err + } + + statusRead, statusWrite, err := os.Pipe() + if err != nil { + return err + } + rpcOpts.StatusFd = proto.Int32(int32(statusWrite.Fd())) + go waitForCriuLazyServer(statusRead, criuOpts.StatusFd) + } + + //no need to dump these information in pre-dump + if !criuOpts.PreDump { + for _, m := range c.config.Mounts { + switch m.Device { + case "bind": + c.addCriuDumpMount(req, m) + case "cgroup": + binds, err := getCgroupMounts(m) + if err != nil { + return err + } + for _, b := range binds { + c.addCriuDumpMount(req, b) + } + } + } + + if err := c.addMaskPaths(req); err != nil { + return err + } + + for _, node := range c.config.Devices { + m := &configs.Mount{Destination: node.Path, Source: node.Path} + c.addCriuDumpMount(req, m) + } + + // Write the FD info to a file in the image directory + fdsJSON, err := json.Marshal(c.initProcess.externalDescriptors()) + if err != nil { + return err + } + + err = ioutil.WriteFile(filepath.Join(criuOpts.ImagesDirectory, descriptorsFilename), fdsJSON, 0655) + if err != nil { + return err + } + } + + err = c.criuSwrk(nil, req, criuOpts, false, nil) + if err != nil { + return err + } + return nil +} + +func (c *linuxContainer) addCriuRestoreMount(req *criurpc.CriuReq, m *configs.Mount) { + mountDest := m.Destination + if strings.HasPrefix(mountDest, c.config.Rootfs) { + mountDest = mountDest[len(c.config.Rootfs):] + } + + extMnt := &criurpc.ExtMountMap{ + Key: proto.String(mountDest), + Val: proto.String(m.Source), + } + req.Opts.ExtMnt = append(req.Opts.ExtMnt, extMnt) +} + +func (c *linuxContainer) restoreNetwork(req *criurpc.CriuReq, criuOpts *CriuOpts) { + for _, iface := range c.config.Networks { + switch iface.Type { + case "veth": + veth := new(criurpc.CriuVethPair) + veth.IfOut = proto.String(iface.HostInterfaceName) + veth.IfIn = proto.String(iface.Name) + req.Opts.Veths = append(req.Opts.Veths, veth) + case "loopback": + // Do nothing + } + } + for _, i := range criuOpts.VethPairs { + veth := new(criurpc.CriuVethPair) + veth.IfOut = proto.String(i.HostInterfaceName) + veth.IfIn = proto.String(i.ContainerInterfaceName) + req.Opts.Veths = append(req.Opts.Veths, veth) + } +} + +func (c *linuxContainer) Restore(process *Process, criuOpts *CriuOpts) error { + c.m.Lock() + defer c.m.Unlock() + + var extraFiles []*os.File + + // TODO(avagin): Figure out how to make this work nicely. CRIU doesn't have + // support for unprivileged restore at the moment. + if c.config.Rootless { + return fmt.Errorf("cannot restore a rootless container") + } + + // criu 1.5.2 => 10502 + if err := c.checkCriuVersion(10502); err != nil { + return err + } + if criuOpts.WorkDirectory == "" { + criuOpts.WorkDirectory = filepath.Join(c.root, "criu.work") + } + // Since a container can be C/R'ed multiple times, + // the work directory may already exist. + if err := os.Mkdir(criuOpts.WorkDirectory, 0655); err != nil && !os.IsExist(err) { + return err + } + workDir, err := os.Open(criuOpts.WorkDirectory) + if err != nil { + return err + } + defer workDir.Close() + if criuOpts.ImagesDirectory == "" { + return fmt.Errorf("invalid directory to restore checkpoint") + } + imageDir, err := os.Open(criuOpts.ImagesDirectory) + if err != nil { + return err + } + defer imageDir.Close() + // CRIU has a few requirements for a root directory: + // * it must be a mount point + // * its parent must not be overmounted + // c.config.Rootfs is bind-mounted to a temporary directory + // to satisfy these requirements. + root := filepath.Join(c.root, "criu-root") + if err := os.Mkdir(root, 0755); err != nil { + return err + } + defer os.Remove(root) + root, err = filepath.EvalSymlinks(root) + if err != nil { + return err + } + err = unix.Mount(c.config.Rootfs, root, "", unix.MS_BIND|unix.MS_REC, "") + if err != nil { + return err + } + defer unix.Unmount(root, unix.MNT_DETACH) + t := criurpc.CriuReqType_RESTORE + req := &criurpc.CriuReq{ + Type: &t, + Opts: &criurpc.CriuOpts{ + ImagesDirFd: proto.Int32(int32(imageDir.Fd())), + WorkDirFd: proto.Int32(int32(workDir.Fd())), + EvasiveDevices: proto.Bool(true), + LogLevel: proto.Int32(4), + LogFile: proto.String("restore.log"), + RstSibling: proto.Bool(true), + Root: proto.String(root), + ManageCgroups: proto.Bool(true), + NotifyScripts: proto.Bool(true), + ShellJob: proto.Bool(criuOpts.ShellJob), + ExtUnixSk: proto.Bool(criuOpts.ExternalUnixConnections), + TcpEstablished: proto.Bool(criuOpts.TcpEstablished), + FileLocks: proto.Bool(criuOpts.FileLocks), + EmptyNs: proto.Uint32(criuOpts.EmptyNs), + OrphanPtsMaster: proto.Bool(true), + AutoDedup: proto.Bool(criuOpts.AutoDedup), + LazyPages: proto.Bool(criuOpts.LazyPages), + }, + } + + // Same as during checkpointing. If the container has a specific network namespace + // assigned to it, this now expects that the checkpoint will be restored in a + // already created network namespace. + nsPath := c.config.Namespaces.PathOf(configs.NEWNET) + if nsPath != "" { + // For this to work we need at least criu 3.11.0 => 31100. + // As there was already a successful version check we will + // not error out if it fails. runc will just behave as it used + // to do and ignore external network namespaces. + err := c.checkCriuVersion(31100) + if err == nil { + // CRIU wants the information about an existing network namespace + // like this: --inherit-fd fd[]: + // The needs to be the same as during checkpointing. + // We are always using 'extRootNetNS' as the key in this. + netns, err := os.Open(nsPath) + defer netns.Close() + if err != nil { + logrus.Error("If a specific network namespace is defined it must exist: %s", err) + return fmt.Errorf("Requested network namespace %v does not exist", nsPath) + } + inheritFd := new(criurpc.InheritFd) + inheritFd.Key = proto.String("extRootNetNS") + // The offset of four is necessary because 0, 1, 2 and 3 is already + // used by stdin, stdout, stderr, 'criu swrk' socket. + inheritFd.Fd = proto.Int32(int32(4 + len(extraFiles))) + req.Opts.InheritFd = append(req.Opts.InheritFd, inheritFd) + // All open FDs need to be transferred to CRIU via extraFiles + extraFiles = append(extraFiles, netns) + } + } + + for _, m := range c.config.Mounts { + switch m.Device { + case "bind": + c.addCriuRestoreMount(req, m) + case "cgroup": + binds, err := getCgroupMounts(m) + if err != nil { + return err + } + for _, b := range binds { + c.addCriuRestoreMount(req, b) + } + } + } + + if len(c.config.MaskPaths) > 0 { + m := &configs.Mount{Destination: "/dev/null", Source: "/dev/null"} + c.addCriuRestoreMount(req, m) + } + + for _, node := range c.config.Devices { + m := &configs.Mount{Destination: node.Path, Source: node.Path} + c.addCriuRestoreMount(req, m) + } + + if criuOpts.EmptyNs&unix.CLONE_NEWNET == 0 { + c.restoreNetwork(req, criuOpts) + } + + // append optional manage cgroups mode + if criuOpts.ManageCgroupsMode != 0 { + // criu 1.7 => 10700 + if err := c.checkCriuVersion(10700); err != nil { + return err + } + mode := criurpc.CriuCgMode(criuOpts.ManageCgroupsMode) + req.Opts.ManageCgroupsMode = &mode + } + + var ( + fds []string + fdJSON []byte + ) + if fdJSON, err = ioutil.ReadFile(filepath.Join(criuOpts.ImagesDirectory, descriptorsFilename)); err != nil { + return err + } + + if err := json.Unmarshal(fdJSON, &fds); err != nil { + return err + } + for i := range fds { + if s := fds[i]; strings.Contains(s, "pipe:") { + inheritFd := new(criurpc.InheritFd) + inheritFd.Key = proto.String(s) + inheritFd.Fd = proto.Int32(int32(i)) + req.Opts.InheritFd = append(req.Opts.InheritFd, inheritFd) + } + } + return c.criuSwrk(process, req, criuOpts, true, extraFiles) +} + +func (c *linuxContainer) criuApplyCgroups(pid int, req *criurpc.CriuReq) error { + // XXX: Do we need to deal with this case? AFAIK criu still requires root. + if err := c.cgroupManager.Apply(pid); err != nil { + return err + } + + if err := c.cgroupManager.Set(c.config); err != nil { + return newSystemError(err) + } + + path := fmt.Sprintf("/proc/%d/cgroup", pid) + cgroupsPaths, err := cgroups.ParseCgroupFile(path) + if err != nil { + return err + } + + for c, p := range cgroupsPaths { + cgroupRoot := &criurpc.CgroupRoot{ + Ctrl: proto.String(c), + Path: proto.String(p), + } + req.Opts.CgRoot = append(req.Opts.CgRoot, cgroupRoot) + } + + return nil +} + +func (c *linuxContainer) criuSwrk(process *Process, req *criurpc.CriuReq, opts *CriuOpts, applyCgroups bool, extraFiles []*os.File) error { + fds, err := unix.Socketpair(unix.AF_LOCAL, unix.SOCK_SEQPACKET|unix.SOCK_CLOEXEC, 0) + if err != nil { + return err + } + + var logPath string + if opts != nil { + logPath = filepath.Join(opts.WorkDirectory, req.GetOpts().GetLogFile()) + } else { + // For the VERSION RPC 'opts' is set to 'nil' and therefore + // opts.WorkDirectory does not exist. Set logPath to "". + logPath = "" + } + criuClient := os.NewFile(uintptr(fds[0]), "criu-transport-client") + criuClientFileCon, err := net.FileConn(criuClient) + criuClient.Close() + if err != nil { + return err + } + + criuClientCon := criuClientFileCon.(*net.UnixConn) + defer criuClientCon.Close() + + criuServer := os.NewFile(uintptr(fds[1]), "criu-transport-server") + defer criuServer.Close() + + args := []string{"swrk", "3"} + if c.criuVersion != 0 { + // If the CRIU Version is still '0' then this is probably + // the initial CRIU run to detect the version. Skip it. + logrus.Debugf("Using CRIU %d at: %s", c.criuVersion, c.criuPath) + } + logrus.Debugf("Using CRIU with following args: %s", args) + cmd := exec.Command(c.criuPath, args...) + if process != nil { + cmd.Stdin = process.Stdin + cmd.Stdout = process.Stdout + cmd.Stderr = process.Stderr + } + cmd.ExtraFiles = append(cmd.ExtraFiles, criuServer) + if extraFiles != nil { + cmd.ExtraFiles = append(cmd.ExtraFiles, extraFiles...) + } + + if err := cmd.Start(); err != nil { + return err + } + criuServer.Close() + + defer func() { + criuClientCon.Close() + _, err := cmd.Process.Wait() + if err != nil { + return + } + }() + + if applyCgroups { + err := c.criuApplyCgroups(cmd.Process.Pid, req) + if err != nil { + return err + } + } + + var extFds []string + if process != nil { + extFds, err = getPipeFds(cmd.Process.Pid) + if err != nil { + return err + } + } + + logrus.Debugf("Using CRIU in %s mode", req.GetType().String()) + // In the case of criurpc.CriuReqType_FEATURE_CHECK req.GetOpts() + // should be empty. For older CRIU versions it still will be + // available but empty. criurpc.CriuReqType_VERSION actually + // has no req.GetOpts(). + if !(req.GetType() == criurpc.CriuReqType_FEATURE_CHECK || + req.GetType() == criurpc.CriuReqType_VERSION) { + + val := reflect.ValueOf(req.GetOpts()) + v := reflect.Indirect(val) + for i := 0; i < v.NumField(); i++ { + st := v.Type() + name := st.Field(i).Name + if strings.HasPrefix(name, "XXX_") { + continue + } + value := val.MethodByName("Get" + name).Call([]reflect.Value{}) + logrus.Debugf("CRIU option %s with value %v", name, value[0]) + } + } + data, err := proto.Marshal(req) + if err != nil { + return err + } + _, err = criuClientCon.Write(data) + if err != nil { + return err + } + + buf := make([]byte, 10*4096) + oob := make([]byte, 4096) + for true { + n, oobn, _, _, err := criuClientCon.ReadMsgUnix(buf, oob) + if err != nil { + return err + } + if n == 0 { + return fmt.Errorf("unexpected EOF") + } + if n == len(buf) { + return fmt.Errorf("buffer is too small") + } + + resp := new(criurpc.CriuResp) + err = proto.Unmarshal(buf[:n], resp) + if err != nil { + return err + } + if !resp.GetSuccess() { + typeString := req.GetType().String() + if typeString == "VERSION" { + // If the VERSION RPC fails this probably means that the CRIU + // version is too old for this RPC. Just return 'nil'. + return nil + } + return fmt.Errorf("criu failed: type %s errno %d\nlog file: %s", typeString, resp.GetCrErrno(), logPath) + } + + t := resp.GetType() + switch { + case t == criurpc.CriuReqType_VERSION: + logrus.Debugf("CRIU version: %s", resp) + criuVersionRPC = resp.GetVersion() + break + case t == criurpc.CriuReqType_FEATURE_CHECK: + logrus.Debugf("Feature check says: %s", resp) + criuFeatures = resp.GetFeatures() + case t == criurpc.CriuReqType_NOTIFY: + if err := c.criuNotifications(resp, process, opts, extFds, oob[:oobn]); err != nil { + return err + } + t = criurpc.CriuReqType_NOTIFY + req = &criurpc.CriuReq{ + Type: &t, + NotifySuccess: proto.Bool(true), + } + data, err = proto.Marshal(req) + if err != nil { + return err + } + _, err = criuClientCon.Write(data) + if err != nil { + return err + } + continue + case t == criurpc.CriuReqType_RESTORE: + case t == criurpc.CriuReqType_DUMP: + case t == criurpc.CriuReqType_PRE_DUMP: + default: + return fmt.Errorf("unable to parse the response %s", resp.String()) + } + + break + } + + criuClientCon.CloseWrite() + // cmd.Wait() waits cmd.goroutines which are used for proxying file descriptors. + // Here we want to wait only the CRIU process. + st, err := cmd.Process.Wait() + if err != nil { + return err + } + + // In pre-dump mode CRIU is in a loop and waits for + // the final DUMP command. + // The current runc pre-dump approach, however, is + // start criu in PRE_DUMP once for a single pre-dump + // and not the whole series of pre-dump, pre-dump, ...m, dump + // If we got the message CriuReqType_PRE_DUMP it means + // CRIU was successful and we need to forcefully stop CRIU + if !st.Success() && *req.Type != criurpc.CriuReqType_PRE_DUMP { + return fmt.Errorf("criu failed: %s\nlog file: %s", st.String(), logPath) + } + return nil +} + +// block any external network activity +func lockNetwork(config *configs.Config) error { + for _, config := range config.Networks { + strategy, err := getStrategy(config.Type) + if err != nil { + return err + } + + if err := strategy.detach(config); err != nil { + return err + } + } + return nil +} + +func unlockNetwork(config *configs.Config) error { + for _, config := range config.Networks { + strategy, err := getStrategy(config.Type) + if err != nil { + return err + } + if err = strategy.attach(config); err != nil { + return err + } + } + return nil +} + +func (c *linuxContainer) criuNotifications(resp *criurpc.CriuResp, process *Process, opts *CriuOpts, fds []string, oob []byte) error { + notify := resp.GetNotify() + if notify == nil { + return fmt.Errorf("invalid response: %s", resp.String()) + } + logrus.Debugf("notify: %s\n", notify.GetScript()) + switch { + case notify.GetScript() == "post-dump": + f, err := os.Create(filepath.Join(c.root, "checkpoint")) + if err != nil { + return err + } + f.Close() + case notify.GetScript() == "network-unlock": + if err := unlockNetwork(c.config); err != nil { + return err + } + case notify.GetScript() == "network-lock": + if err := lockNetwork(c.config); err != nil { + return err + } + case notify.GetScript() == "setup-namespaces": + if c.config.Hooks != nil { + bundle, annotations := utils.Annotations(c.config.Labels) + s := configs.HookState{ + Version: c.config.Version, + ID: c.id, + Pid: int(notify.GetPid()), + Bundle: bundle, + Annotations: annotations, + } + for i, hook := range c.config.Hooks.Prestart { + if err := hook.Run(s); err != nil { + return newSystemErrorWithCausef(err, "running prestart hook %d", i) + } + } + } + case notify.GetScript() == "post-restore": + pid := notify.GetPid() + r, err := newRestoredProcess(int(pid), fds) + if err != nil { + return err + } + process.ops = r + if err := c.state.transition(&restoredState{ + imageDir: opts.ImagesDirectory, + c: c, + }); err != nil { + return err + } + // create a timestamp indicating when the restored checkpoint was started + c.created = time.Now().UTC() + if _, err := c.updateState(r); err != nil { + return err + } + if err := os.Remove(filepath.Join(c.root, "checkpoint")); err != nil { + if !os.IsNotExist(err) { + logrus.Error(err) + } + } + case notify.GetScript() == "orphan-pts-master": + scm, err := unix.ParseSocketControlMessage(oob) + if err != nil { + return err + } + fds, err := unix.ParseUnixRights(&scm[0]) + if err != nil { + return err + } + + master := os.NewFile(uintptr(fds[0]), "orphan-pts-master") + defer master.Close() + + // While we can access console.master, using the API is a good idea. + if err := utils.SendFd(process.ConsoleSocket, master.Name(), master.Fd()); err != nil { + return err + } + } + return nil +} + +func (c *linuxContainer) updateState(process parentProcess) (*State, error) { + if process != nil { + c.initProcess = process + } + state, err := c.currentState() + if err != nil { + return nil, err + } + err = c.saveState(state) + if err != nil { + return nil, err + } + return state, nil +} + +func (c *linuxContainer) saveState(s *State) error { + f, err := os.Create(filepath.Join(c.root, stateFilename)) + if err != nil { + return err + } + defer f.Close() + return utils.WriteJSON(f, s) +} + +func (c *linuxContainer) deleteState() error { + return os.Remove(filepath.Join(c.root, stateFilename)) +} + +func (c *linuxContainer) currentStatus() (Status, error) { + if err := c.refreshState(); err != nil { + return -1, err + } + return c.state.status(), nil +} + +// refreshState needs to be called to verify that the current state on the +// container is what is true. Because consumers of libcontainer can use it +// out of process we need to verify the container's status based on runtime +// information and not rely on our in process info. +func (c *linuxContainer) refreshState() error { + paused, err := c.isPaused() + if err != nil { + return err + } + if paused { + return c.state.transition(&pausedState{c: c}) + } + t, err := c.runType() + if err != nil { + return err + } + switch t { + case Created: + return c.state.transition(&createdState{c: c}) + case Running: + return c.state.transition(&runningState{c: c}) + } + return c.state.transition(&stoppedState{c: c}) +} + +func (c *linuxContainer) runType() (Status, error) { + if c.initProcess == nil { + return Stopped, nil + } + pid := c.initProcess.pid() + stat, err := system.Stat(pid) + if err != nil { + return Stopped, nil + } + if stat.StartTime != c.initProcessStartTime || stat.State == system.Zombie || stat.State == system.Dead { + return Stopped, nil + } + // We'll create exec fifo and blocking on it after container is created, + // and delete it after start container. + if _, err := os.Stat(filepath.Join(c.root, execFifoFilename)); err == nil { + return Created, nil + } + return Running, nil +} + +func (c *linuxContainer) isPaused() (bool, error) { + fcg := c.cgroupManager.GetPaths()["freezer"] + if fcg == "" { + // A container doesn't have a freezer cgroup + return false, nil + } + data, err := ioutil.ReadFile(filepath.Join(fcg, "freezer.state")) + if err != nil { + // If freezer cgroup is not mounted, the container would just be not paused. + if os.IsNotExist(err) { + return false, nil + } + return false, newSystemErrorWithCause(err, "checking if container is paused") + } + return bytes.Equal(bytes.TrimSpace(data), []byte("FROZEN")), nil +} + +func (c *linuxContainer) currentState() (*State, error) { + var ( + startTime uint64 + externalDescriptors []string + pid = -1 + ) + if c.initProcess != nil { + pid = c.initProcess.pid() + startTime, _ = c.initProcess.startTime() + externalDescriptors = c.initProcess.externalDescriptors() + } + intelRdtPath, err := intelrdt.GetIntelRdtPath(c.ID()) + if err != nil { + intelRdtPath = "" + } + state := &State{ + BaseState: BaseState{ + ID: c.ID(), + Config: *c.config, + InitProcessPid: pid, + InitProcessStartTime: startTime, + Created: c.created, + }, + Rootless: c.config.Rootless, + CgroupPaths: c.cgroupManager.GetPaths(), + IntelRdtPath: intelRdtPath, + NamespacePaths: make(map[configs.NamespaceType]string), + ExternalDescriptors: externalDescriptors, + } + if pid > 0 { + for _, ns := range c.config.Namespaces { + state.NamespacePaths[ns.Type] = ns.GetPath(pid) + } + for _, nsType := range configs.NamespaceTypes() { + if !configs.IsNamespaceSupported(nsType) { + continue + } + if _, ok := state.NamespacePaths[nsType]; !ok { + ns := configs.Namespace{Type: nsType} + state.NamespacePaths[ns.Type] = ns.GetPath(pid) + } + } + } + return state, nil +} + +// orderNamespacePaths sorts namespace paths into a list of paths that we +// can setns in order. +func (c *linuxContainer) orderNamespacePaths(namespaces map[configs.NamespaceType]string) ([]string, error) { + paths := []string{} + + for _, ns := range configs.NamespaceTypes() { + + // Remove namespaces that we don't need to join. + if !c.config.Namespaces.Contains(ns) { + continue + } + + if p, ok := namespaces[ns]; ok && p != "" { + // check if the requested namespace is supported + if !configs.IsNamespaceSupported(ns) { + return nil, newSystemError(fmt.Errorf("namespace %s is not supported", ns)) + } + // only set to join this namespace if it exists + if _, err := os.Lstat(p); err != nil { + return nil, newSystemErrorWithCausef(err, "running lstat on namespace path %q", p) + } + // do not allow namespace path with comma as we use it to separate + // the namespace paths + if strings.ContainsRune(p, ',') { + return nil, newSystemError(fmt.Errorf("invalid path %s", p)) + } + paths = append(paths, fmt.Sprintf("%s:%s", configs.NsName(ns), p)) + } + + } + + return paths, nil +} + +func encodeIDMapping(idMap []configs.IDMap) ([]byte, error) { + data := bytes.NewBuffer(nil) + for _, im := range idMap { + line := fmt.Sprintf("%d %d %d\n", im.ContainerID, im.HostID, im.Size) + if _, err := data.WriteString(line); err != nil { + return nil, err + } + } + return data.Bytes(), nil +} + +// bootstrapData encodes the necessary data in netlink binary format +// as a io.Reader. +// Consumer can write the data to a bootstrap program +// such as one that uses nsenter package to bootstrap the container's +// init process correctly, i.e. with correct namespaces, uid/gid +// mapping etc. +func (c *linuxContainer) bootstrapData(cloneFlags uintptr, nsMaps map[configs.NamespaceType]string) (io.Reader, error) { + // create the netlink message + r := nl.NewNetlinkRequest(int(InitMsg), 0) + + // write cloneFlags + r.AddData(&Int32msg{ + Type: CloneFlagsAttr, + Value: uint32(cloneFlags), + }) + + // write custom namespace paths + if len(nsMaps) > 0 { + nsPaths, err := c.orderNamespacePaths(nsMaps) + if err != nil { + return nil, err + } + r.AddData(&Bytemsg{ + Type: NsPathsAttr, + Value: []byte(strings.Join(nsPaths, ",")), + }) + } + + // write namespace paths only when we are not joining an existing user ns + _, joinExistingUser := nsMaps[configs.NEWUSER] + if !joinExistingUser { + // write uid mappings + if len(c.config.UidMappings) > 0 { + if c.config.Rootless && c.newuidmapPath != "" { + r.AddData(&Bytemsg{ + Type: UidmapPathAttr, + Value: []byte(c.newuidmapPath), + }) + } + b, err := encodeIDMapping(c.config.UidMappings) + if err != nil { + return nil, err + } + r.AddData(&Bytemsg{ + Type: UidmapAttr, + Value: b, + }) + } + + // write gid mappings + if len(c.config.GidMappings) > 0 { + b, err := encodeIDMapping(c.config.GidMappings) + if err != nil { + return nil, err + } + r.AddData(&Bytemsg{ + Type: GidmapAttr, + Value: b, + }) + if c.config.Rootless && c.newgidmapPath != "" { + r.AddData(&Bytemsg{ + Type: GidmapPathAttr, + Value: []byte(c.newgidmapPath), + }) + } + if requiresRootOrMappingTool(c.config) { + r.AddData(&Boolmsg{ + Type: SetgroupAttr, + Value: true, + }) + } + } + } + + if c.config.OomScoreAdj != nil { + // write oom_score_adj + r.AddData(&Bytemsg{ + Type: OomScoreAdjAttr, + Value: []byte(fmt.Sprintf("%d", *c.config.OomScoreAdj)), + }) + } + + // write rootless + r.AddData(&Boolmsg{ + Type: RootlessAttr, + Value: c.config.Rootless, + }) + + return bytes.NewReader(r.Serialize()), nil +} + +// ignoreTerminateErrors returns nil if the given err matches an error known +// to indicate that the terminate occurred successfully or err was nil, otherwise +// err is returned unaltered. +func ignoreTerminateErrors(err error) error { + if err == nil { + return nil + } + s := err.Error() + switch { + case strings.Contains(s, "process already finished"), strings.Contains(s, "Wait was already called"): + return nil + } + return err +} + +func requiresRootOrMappingTool(c *configs.Config) bool { + gidMap := []configs.IDMap{ + {ContainerID: 0, HostID: os.Getegid(), Size: 1}, + } + return !reflect.DeepEqual(c.GidMappings, gidMap) +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/criu_opts_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/criu_opts_linux.go new file mode 100644 index 000000000..a2e344fc4 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/criu_opts_linux.go @@ -0,0 +1,40 @@ +package libcontainer + +// cgroup restoring strategy provided by criu +type cgMode uint32 + +const ( + CRIU_CG_MODE_SOFT cgMode = 3 + iota // restore cgroup properties if only dir created by criu + CRIU_CG_MODE_FULL // always restore all cgroups and their properties + CRIU_CG_MODE_STRICT // restore all, requiring them to not present in the system + CRIU_CG_MODE_DEFAULT // the same as CRIU_CG_MODE_SOFT +) + +type CriuPageServerInfo struct { + Address string // IP address of CRIU page server + Port int32 // port number of CRIU page server +} + +type VethPairName struct { + ContainerInterfaceName string + HostInterfaceName string +} + +type CriuOpts struct { + ImagesDirectory string // directory for storing image files + WorkDirectory string // directory to cd and write logs/pidfiles/stats to + ParentImage string // directory for storing parent image files in pre-dump and dump + LeaveRunning bool // leave container in running state after checkpoint + TcpEstablished bool // checkpoint/restore established TCP connections + ExternalUnixConnections bool // allow external unix connections + ShellJob bool // allow to dump and restore shell jobs + FileLocks bool // handle file locks, for safety + PreDump bool // call criu predump to perform iterative checkpoint + PageServer CriuPageServerInfo // allow to dump to criu page server + VethPairs []VethPairName // pass the veth to criu when restore + ManageCgroupsMode cgMode // dump or restore cgroup mode + EmptyNs uint32 // don't c/r properties for namespace from this mask + AutoDedup bool // auto deduplication for incremental dumps + LazyPages bool // restore memory pages lazily using userfaultfd + StatusFd string // fd for feedback when lazy server is ready +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/criurpc/Makefile b/vendor/github.com/opencontainers/runc/libcontainer/criurpc/Makefile new file mode 100644 index 000000000..3e5346a3f --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/criurpc/Makefile @@ -0,0 +1,2 @@ +gen: criurpc.proto + protoc --go_out=. criurpc.proto diff --git a/vendor/github.com/opencontainers/runc/libcontainer/criurpc/criurpc.pb.go b/vendor/github.com/opencontainers/runc/libcontainer/criurpc/criurpc.pb.go new file mode 100644 index 000000000..21af9db97 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/criurpc/criurpc.pb.go @@ -0,0 +1,1178 @@ +// Code generated by protoc-gen-go. +// source: criurpc.proto +// DO NOT EDIT! + +/* +Package criurpc is a generated protocol buffer package. + +It is generated from these files: + criurpc.proto + +It has these top-level messages: + CriuPageServerInfo + CriuVethPair + ExtMountMap + JoinNamespace + InheritFd + CgroupRoot + UnixSk + CriuOpts + CriuDumpResp + CriuRestoreResp + CriuNotify + CriuFeatures + CriuReq + CriuResp + CriuVersion +*/ +package criurpc + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +type CriuCgMode int32 + +const ( + CriuCgMode_IGNORE CriuCgMode = 0 + CriuCgMode_CG_NONE CriuCgMode = 1 + CriuCgMode_PROPS CriuCgMode = 2 + CriuCgMode_SOFT CriuCgMode = 3 + CriuCgMode_FULL CriuCgMode = 4 + CriuCgMode_STRICT CriuCgMode = 5 + CriuCgMode_DEFAULT CriuCgMode = 6 +) + +var CriuCgMode_name = map[int32]string{ + 0: "IGNORE", + 1: "CG_NONE", + 2: "PROPS", + 3: "SOFT", + 4: "FULL", + 5: "STRICT", + 6: "DEFAULT", +} +var CriuCgMode_value = map[string]int32{ + "IGNORE": 0, + "CG_NONE": 1, + "PROPS": 2, + "SOFT": 3, + "FULL": 4, + "STRICT": 5, + "DEFAULT": 6, +} + +func (x CriuCgMode) Enum() *CriuCgMode { + p := new(CriuCgMode) + *p = x + return p +} +func (x CriuCgMode) String() string { + return proto.EnumName(CriuCgMode_name, int32(x)) +} +func (x *CriuCgMode) UnmarshalJSON(data []byte) error { + value, err := proto.UnmarshalJSONEnum(CriuCgMode_value, data, "CriuCgMode") + if err != nil { + return err + } + *x = CriuCgMode(value) + return nil +} +func (CriuCgMode) EnumDescriptor() ([]byte, []int) { return fileDescriptor0, []int{0} } + +type CriuReqType int32 + +const ( + CriuReqType_EMPTY CriuReqType = 0 + CriuReqType_DUMP CriuReqType = 1 + CriuReqType_RESTORE CriuReqType = 2 + CriuReqType_CHECK CriuReqType = 3 + CriuReqType_PRE_DUMP CriuReqType = 4 + CriuReqType_PAGE_SERVER CriuReqType = 5 + CriuReqType_NOTIFY CriuReqType = 6 + CriuReqType_CPUINFO_DUMP CriuReqType = 7 + CriuReqType_CPUINFO_CHECK CriuReqType = 8 + CriuReqType_FEATURE_CHECK CriuReqType = 9 + CriuReqType_VERSION CriuReqType = 10 +) + +var CriuReqType_name = map[int32]string{ + 0: "EMPTY", + 1: "DUMP", + 2: "RESTORE", + 3: "CHECK", + 4: "PRE_DUMP", + 5: "PAGE_SERVER", + 6: "NOTIFY", + 7: "CPUINFO_DUMP", + 8: "CPUINFO_CHECK", + 9: "FEATURE_CHECK", + 10: "VERSION", +} +var CriuReqType_value = map[string]int32{ + "EMPTY": 0, + "DUMP": 1, + "RESTORE": 2, + "CHECK": 3, + "PRE_DUMP": 4, + "PAGE_SERVER": 5, + "NOTIFY": 6, + "CPUINFO_DUMP": 7, + "CPUINFO_CHECK": 8, + "FEATURE_CHECK": 9, + "VERSION": 10, +} + +func (x CriuReqType) Enum() *CriuReqType { + p := new(CriuReqType) + *p = x + return p +} +func (x CriuReqType) String() string { + return proto.EnumName(CriuReqType_name, int32(x)) +} +func (x *CriuReqType) UnmarshalJSON(data []byte) error { + value, err := proto.UnmarshalJSONEnum(CriuReqType_value, data, "CriuReqType") + if err != nil { + return err + } + *x = CriuReqType(value) + return nil +} +func (CriuReqType) EnumDescriptor() ([]byte, []int) { return fileDescriptor0, []int{1} } + +type CriuPageServerInfo struct { + Address *string `protobuf:"bytes,1,opt,name=address" json:"address,omitempty"` + Port *int32 `protobuf:"varint,2,opt,name=port" json:"port,omitempty"` + Pid *int32 `protobuf:"varint,3,opt,name=pid" json:"pid,omitempty"` + Fd *int32 `protobuf:"varint,4,opt,name=fd" json:"fd,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *CriuPageServerInfo) Reset() { *m = CriuPageServerInfo{} } +func (m *CriuPageServerInfo) String() string { return proto.CompactTextString(m) } +func (*CriuPageServerInfo) ProtoMessage() {} +func (*CriuPageServerInfo) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{0} } + +func (m *CriuPageServerInfo) GetAddress() string { + if m != nil && m.Address != nil { + return *m.Address + } + return "" +} + +func (m *CriuPageServerInfo) GetPort() int32 { + if m != nil && m.Port != nil { + return *m.Port + } + return 0 +} + +func (m *CriuPageServerInfo) GetPid() int32 { + if m != nil && m.Pid != nil { + return *m.Pid + } + return 0 +} + +func (m *CriuPageServerInfo) GetFd() int32 { + if m != nil && m.Fd != nil { + return *m.Fd + } + return 0 +} + +type CriuVethPair struct { + IfIn *string `protobuf:"bytes,1,req,name=if_in,json=ifIn" json:"if_in,omitempty"` + IfOut *string `protobuf:"bytes,2,req,name=if_out,json=ifOut" json:"if_out,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *CriuVethPair) Reset() { *m = CriuVethPair{} } +func (m *CriuVethPair) String() string { return proto.CompactTextString(m) } +func (*CriuVethPair) ProtoMessage() {} +func (*CriuVethPair) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{1} } + +func (m *CriuVethPair) GetIfIn() string { + if m != nil && m.IfIn != nil { + return *m.IfIn + } + return "" +} + +func (m *CriuVethPair) GetIfOut() string { + if m != nil && m.IfOut != nil { + return *m.IfOut + } + return "" +} + +type ExtMountMap struct { + Key *string `protobuf:"bytes,1,req,name=key" json:"key,omitempty"` + Val *string `protobuf:"bytes,2,req,name=val" json:"val,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *ExtMountMap) Reset() { *m = ExtMountMap{} } +func (m *ExtMountMap) String() string { return proto.CompactTextString(m) } +func (*ExtMountMap) ProtoMessage() {} +func (*ExtMountMap) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{2} } + +func (m *ExtMountMap) GetKey() string { + if m != nil && m.Key != nil { + return *m.Key + } + return "" +} + +func (m *ExtMountMap) GetVal() string { + if m != nil && m.Val != nil { + return *m.Val + } + return "" +} + +type JoinNamespace struct { + Ns *string `protobuf:"bytes,1,req,name=ns" json:"ns,omitempty"` + NsFile *string `protobuf:"bytes,2,req,name=ns_file,json=nsFile" json:"ns_file,omitempty"` + ExtraOpt *string `protobuf:"bytes,3,opt,name=extra_opt,json=extraOpt" json:"extra_opt,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *JoinNamespace) Reset() { *m = JoinNamespace{} } +func (m *JoinNamespace) String() string { return proto.CompactTextString(m) } +func (*JoinNamespace) ProtoMessage() {} +func (*JoinNamespace) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{3} } + +func (m *JoinNamespace) GetNs() string { + if m != nil && m.Ns != nil { + return *m.Ns + } + return "" +} + +func (m *JoinNamespace) GetNsFile() string { + if m != nil && m.NsFile != nil { + return *m.NsFile + } + return "" +} + +func (m *JoinNamespace) GetExtraOpt() string { + if m != nil && m.ExtraOpt != nil { + return *m.ExtraOpt + } + return "" +} + +type InheritFd struct { + Key *string `protobuf:"bytes,1,req,name=key" json:"key,omitempty"` + Fd *int32 `protobuf:"varint,2,req,name=fd" json:"fd,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *InheritFd) Reset() { *m = InheritFd{} } +func (m *InheritFd) String() string { return proto.CompactTextString(m) } +func (*InheritFd) ProtoMessage() {} +func (*InheritFd) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{4} } + +func (m *InheritFd) GetKey() string { + if m != nil && m.Key != nil { + return *m.Key + } + return "" +} + +func (m *InheritFd) GetFd() int32 { + if m != nil && m.Fd != nil { + return *m.Fd + } + return 0 +} + +type CgroupRoot struct { + Ctrl *string `protobuf:"bytes,1,opt,name=ctrl" json:"ctrl,omitempty"` + Path *string `protobuf:"bytes,2,req,name=path" json:"path,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *CgroupRoot) Reset() { *m = CgroupRoot{} } +func (m *CgroupRoot) String() string { return proto.CompactTextString(m) } +func (*CgroupRoot) ProtoMessage() {} +func (*CgroupRoot) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{5} } + +func (m *CgroupRoot) GetCtrl() string { + if m != nil && m.Ctrl != nil { + return *m.Ctrl + } + return "" +} + +func (m *CgroupRoot) GetPath() string { + if m != nil && m.Path != nil { + return *m.Path + } + return "" +} + +type UnixSk struct { + Inode *uint32 `protobuf:"varint,1,req,name=inode" json:"inode,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *UnixSk) Reset() { *m = UnixSk{} } +func (m *UnixSk) String() string { return proto.CompactTextString(m) } +func (*UnixSk) ProtoMessage() {} +func (*UnixSk) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{6} } + +func (m *UnixSk) GetInode() uint32 { + if m != nil && m.Inode != nil { + return *m.Inode + } + return 0 +} + +type CriuOpts struct { + ImagesDirFd *int32 `protobuf:"varint,1,req,name=images_dir_fd,json=imagesDirFd" json:"images_dir_fd,omitempty"` + Pid *int32 `protobuf:"varint,2,opt,name=pid" json:"pid,omitempty"` + LeaveRunning *bool `protobuf:"varint,3,opt,name=leave_running,json=leaveRunning" json:"leave_running,omitempty"` + ExtUnixSk *bool `protobuf:"varint,4,opt,name=ext_unix_sk,json=extUnixSk" json:"ext_unix_sk,omitempty"` + TcpEstablished *bool `protobuf:"varint,5,opt,name=tcp_established,json=tcpEstablished" json:"tcp_established,omitempty"` + EvasiveDevices *bool `protobuf:"varint,6,opt,name=evasive_devices,json=evasiveDevices" json:"evasive_devices,omitempty"` + ShellJob *bool `protobuf:"varint,7,opt,name=shell_job,json=shellJob" json:"shell_job,omitempty"` + FileLocks *bool `protobuf:"varint,8,opt,name=file_locks,json=fileLocks" json:"file_locks,omitempty"` + LogLevel *int32 `protobuf:"varint,9,opt,name=log_level,json=logLevel,def=2" json:"log_level,omitempty"` + LogFile *string `protobuf:"bytes,10,opt,name=log_file,json=logFile" json:"log_file,omitempty"` + Ps *CriuPageServerInfo `protobuf:"bytes,11,opt,name=ps" json:"ps,omitempty"` + NotifyScripts *bool `protobuf:"varint,12,opt,name=notify_scripts,json=notifyScripts" json:"notify_scripts,omitempty"` + Root *string `protobuf:"bytes,13,opt,name=root" json:"root,omitempty"` + ParentImg *string `protobuf:"bytes,14,opt,name=parent_img,json=parentImg" json:"parent_img,omitempty"` + TrackMem *bool `protobuf:"varint,15,opt,name=track_mem,json=trackMem" json:"track_mem,omitempty"` + AutoDedup *bool `protobuf:"varint,16,opt,name=auto_dedup,json=autoDedup" json:"auto_dedup,omitempty"` + WorkDirFd *int32 `protobuf:"varint,17,opt,name=work_dir_fd,json=workDirFd" json:"work_dir_fd,omitempty"` + LinkRemap *bool `protobuf:"varint,18,opt,name=link_remap,json=linkRemap" json:"link_remap,omitempty"` + Veths []*CriuVethPair `protobuf:"bytes,19,rep,name=veths" json:"veths,omitempty"` + CpuCap *uint32 `protobuf:"varint,20,opt,name=cpu_cap,json=cpuCap,def=4294967295" json:"cpu_cap,omitempty"` + ForceIrmap *bool `protobuf:"varint,21,opt,name=force_irmap,json=forceIrmap" json:"force_irmap,omitempty"` + ExecCmd []string `protobuf:"bytes,22,rep,name=exec_cmd,json=execCmd" json:"exec_cmd,omitempty"` + ExtMnt []*ExtMountMap `protobuf:"bytes,23,rep,name=ext_mnt,json=extMnt" json:"ext_mnt,omitempty"` + ManageCgroups *bool `protobuf:"varint,24,opt,name=manage_cgroups,json=manageCgroups" json:"manage_cgroups,omitempty"` + CgRoot []*CgroupRoot `protobuf:"bytes,25,rep,name=cg_root,json=cgRoot" json:"cg_root,omitempty"` + RstSibling *bool `protobuf:"varint,26,opt,name=rst_sibling,json=rstSibling" json:"rst_sibling,omitempty"` + InheritFd []*InheritFd `protobuf:"bytes,27,rep,name=inherit_fd,json=inheritFd" json:"inherit_fd,omitempty"` + AutoExtMnt *bool `protobuf:"varint,28,opt,name=auto_ext_mnt,json=autoExtMnt" json:"auto_ext_mnt,omitempty"` + ExtSharing *bool `protobuf:"varint,29,opt,name=ext_sharing,json=extSharing" json:"ext_sharing,omitempty"` + ExtMasters *bool `protobuf:"varint,30,opt,name=ext_masters,json=extMasters" json:"ext_masters,omitempty"` + SkipMnt []string `protobuf:"bytes,31,rep,name=skip_mnt,json=skipMnt" json:"skip_mnt,omitempty"` + EnableFs []string `protobuf:"bytes,32,rep,name=enable_fs,json=enableFs" json:"enable_fs,omitempty"` + UnixSkIno []*UnixSk `protobuf:"bytes,33,rep,name=unix_sk_ino,json=unixSkIno" json:"unix_sk_ino,omitempty"` + ManageCgroupsMode *CriuCgMode `protobuf:"varint,34,opt,name=manage_cgroups_mode,json=manageCgroupsMode,enum=CriuCgMode" json:"manage_cgroups_mode,omitempty"` + GhostLimit *uint32 `protobuf:"varint,35,opt,name=ghost_limit,json=ghostLimit,def=1048576" json:"ghost_limit,omitempty"` + IrmapScanPaths []string `protobuf:"bytes,36,rep,name=irmap_scan_paths,json=irmapScanPaths" json:"irmap_scan_paths,omitempty"` + External []string `protobuf:"bytes,37,rep,name=external" json:"external,omitempty"` + EmptyNs *uint32 `protobuf:"varint,38,opt,name=empty_ns,json=emptyNs" json:"empty_ns,omitempty"` + JoinNs []*JoinNamespace `protobuf:"bytes,39,rep,name=join_ns,json=joinNs" json:"join_ns,omitempty"` + CgroupProps *string `protobuf:"bytes,41,opt,name=cgroup_props,json=cgroupProps" json:"cgroup_props,omitempty"` + CgroupPropsFile *string `protobuf:"bytes,42,opt,name=cgroup_props_file,json=cgroupPropsFile" json:"cgroup_props_file,omitempty"` + CgroupDumpController []string `protobuf:"bytes,43,rep,name=cgroup_dump_controller,json=cgroupDumpController" json:"cgroup_dump_controller,omitempty"` + FreezeCgroup *string `protobuf:"bytes,44,opt,name=freeze_cgroup,json=freezeCgroup" json:"freeze_cgroup,omitempty"` + Timeout *uint32 `protobuf:"varint,45,opt,name=timeout" json:"timeout,omitempty"` + TcpSkipInFlight *bool `protobuf:"varint,46,opt,name=tcp_skip_in_flight,json=tcpSkipInFlight" json:"tcp_skip_in_flight,omitempty"` + WeakSysctls *bool `protobuf:"varint,47,opt,name=weak_sysctls,json=weakSysctls" json:"weak_sysctls,omitempty"` + LazyPages *bool `protobuf:"varint,48,opt,name=lazy_pages,json=lazyPages" json:"lazy_pages,omitempty"` + StatusFd *int32 `protobuf:"varint,49,opt,name=status_fd,json=statusFd" json:"status_fd,omitempty"` + OrphanPtsMaster *bool `protobuf:"varint,50,opt,name=orphan_pts_master,json=orphanPtsMaster" json:"orphan_pts_master,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *CriuOpts) Reset() { *m = CriuOpts{} } +func (m *CriuOpts) String() string { return proto.CompactTextString(m) } +func (*CriuOpts) ProtoMessage() {} +func (*CriuOpts) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{7} } + +const Default_CriuOpts_LogLevel int32 = 2 +const Default_CriuOpts_CpuCap uint32 = 4294967295 +const Default_CriuOpts_GhostLimit uint32 = 1048576 + +func (m *CriuOpts) GetImagesDirFd() int32 { + if m != nil && m.ImagesDirFd != nil { + return *m.ImagesDirFd + } + return 0 +} + +func (m *CriuOpts) GetPid() int32 { + if m != nil && m.Pid != nil { + return *m.Pid + } + return 0 +} + +func (m *CriuOpts) GetLeaveRunning() bool { + if m != nil && m.LeaveRunning != nil { + return *m.LeaveRunning + } + return false +} + +func (m *CriuOpts) GetExtUnixSk() bool { + if m != nil && m.ExtUnixSk != nil { + return *m.ExtUnixSk + } + return false +} + +func (m *CriuOpts) GetTcpEstablished() bool { + if m != nil && m.TcpEstablished != nil { + return *m.TcpEstablished + } + return false +} + +func (m *CriuOpts) GetEvasiveDevices() bool { + if m != nil && m.EvasiveDevices != nil { + return *m.EvasiveDevices + } + return false +} + +func (m *CriuOpts) GetShellJob() bool { + if m != nil && m.ShellJob != nil { + return *m.ShellJob + } + return false +} + +func (m *CriuOpts) GetFileLocks() bool { + if m != nil && m.FileLocks != nil { + return *m.FileLocks + } + return false +} + +func (m *CriuOpts) GetLogLevel() int32 { + if m != nil && m.LogLevel != nil { + return *m.LogLevel + } + return Default_CriuOpts_LogLevel +} + +func (m *CriuOpts) GetLogFile() string { + if m != nil && m.LogFile != nil { + return *m.LogFile + } + return "" +} + +func (m *CriuOpts) GetPs() *CriuPageServerInfo { + if m != nil { + return m.Ps + } + return nil +} + +func (m *CriuOpts) GetNotifyScripts() bool { + if m != nil && m.NotifyScripts != nil { + return *m.NotifyScripts + } + return false +} + +func (m *CriuOpts) GetRoot() string { + if m != nil && m.Root != nil { + return *m.Root + } + return "" +} + +func (m *CriuOpts) GetParentImg() string { + if m != nil && m.ParentImg != nil { + return *m.ParentImg + } + return "" +} + +func (m *CriuOpts) GetTrackMem() bool { + if m != nil && m.TrackMem != nil { + return *m.TrackMem + } + return false +} + +func (m *CriuOpts) GetAutoDedup() bool { + if m != nil && m.AutoDedup != nil { + return *m.AutoDedup + } + return false +} + +func (m *CriuOpts) GetWorkDirFd() int32 { + if m != nil && m.WorkDirFd != nil { + return *m.WorkDirFd + } + return 0 +} + +func (m *CriuOpts) GetLinkRemap() bool { + if m != nil && m.LinkRemap != nil { + return *m.LinkRemap + } + return false +} + +func (m *CriuOpts) GetVeths() []*CriuVethPair { + if m != nil { + return m.Veths + } + return nil +} + +func (m *CriuOpts) GetCpuCap() uint32 { + if m != nil && m.CpuCap != nil { + return *m.CpuCap + } + return Default_CriuOpts_CpuCap +} + +func (m *CriuOpts) GetForceIrmap() bool { + if m != nil && m.ForceIrmap != nil { + return *m.ForceIrmap + } + return false +} + +func (m *CriuOpts) GetExecCmd() []string { + if m != nil { + return m.ExecCmd + } + return nil +} + +func (m *CriuOpts) GetExtMnt() []*ExtMountMap { + if m != nil { + return m.ExtMnt + } + return nil +} + +func (m *CriuOpts) GetManageCgroups() bool { + if m != nil && m.ManageCgroups != nil { + return *m.ManageCgroups + } + return false +} + +func (m *CriuOpts) GetCgRoot() []*CgroupRoot { + if m != nil { + return m.CgRoot + } + return nil +} + +func (m *CriuOpts) GetRstSibling() bool { + if m != nil && m.RstSibling != nil { + return *m.RstSibling + } + return false +} + +func (m *CriuOpts) GetInheritFd() []*InheritFd { + if m != nil { + return m.InheritFd + } + return nil +} + +func (m *CriuOpts) GetAutoExtMnt() bool { + if m != nil && m.AutoExtMnt != nil { + return *m.AutoExtMnt + } + return false +} + +func (m *CriuOpts) GetExtSharing() bool { + if m != nil && m.ExtSharing != nil { + return *m.ExtSharing + } + return false +} + +func (m *CriuOpts) GetExtMasters() bool { + if m != nil && m.ExtMasters != nil { + return *m.ExtMasters + } + return false +} + +func (m *CriuOpts) GetSkipMnt() []string { + if m != nil { + return m.SkipMnt + } + return nil +} + +func (m *CriuOpts) GetEnableFs() []string { + if m != nil { + return m.EnableFs + } + return nil +} + +func (m *CriuOpts) GetUnixSkIno() []*UnixSk { + if m != nil { + return m.UnixSkIno + } + return nil +} + +func (m *CriuOpts) GetManageCgroupsMode() CriuCgMode { + if m != nil && m.ManageCgroupsMode != nil { + return *m.ManageCgroupsMode + } + return CriuCgMode_IGNORE +} + +func (m *CriuOpts) GetGhostLimit() uint32 { + if m != nil && m.GhostLimit != nil { + return *m.GhostLimit + } + return Default_CriuOpts_GhostLimit +} + +func (m *CriuOpts) GetIrmapScanPaths() []string { + if m != nil { + return m.IrmapScanPaths + } + return nil +} + +func (m *CriuOpts) GetExternal() []string { + if m != nil { + return m.External + } + return nil +} + +func (m *CriuOpts) GetEmptyNs() uint32 { + if m != nil && m.EmptyNs != nil { + return *m.EmptyNs + } + return 0 +} + +func (m *CriuOpts) GetJoinNs() []*JoinNamespace { + if m != nil { + return m.JoinNs + } + return nil +} + +func (m *CriuOpts) GetCgroupProps() string { + if m != nil && m.CgroupProps != nil { + return *m.CgroupProps + } + return "" +} + +func (m *CriuOpts) GetCgroupPropsFile() string { + if m != nil && m.CgroupPropsFile != nil { + return *m.CgroupPropsFile + } + return "" +} + +func (m *CriuOpts) GetCgroupDumpController() []string { + if m != nil { + return m.CgroupDumpController + } + return nil +} + +func (m *CriuOpts) GetFreezeCgroup() string { + if m != nil && m.FreezeCgroup != nil { + return *m.FreezeCgroup + } + return "" +} + +func (m *CriuOpts) GetTimeout() uint32 { + if m != nil && m.Timeout != nil { + return *m.Timeout + } + return 0 +} + +func (m *CriuOpts) GetTcpSkipInFlight() bool { + if m != nil && m.TcpSkipInFlight != nil { + return *m.TcpSkipInFlight + } + return false +} + +func (m *CriuOpts) GetWeakSysctls() bool { + if m != nil && m.WeakSysctls != nil { + return *m.WeakSysctls + } + return false +} + +func (m *CriuOpts) GetLazyPages() bool { + if m != nil && m.LazyPages != nil { + return *m.LazyPages + } + return false +} + +func (m *CriuOpts) GetStatusFd() int32 { + if m != nil && m.StatusFd != nil { + return *m.StatusFd + } + return 0 +} + +func (m *CriuOpts) GetOrphanPtsMaster() bool { + if m != nil && m.OrphanPtsMaster != nil { + return *m.OrphanPtsMaster + } + return false +} + +type CriuDumpResp struct { + Restored *bool `protobuf:"varint,1,opt,name=restored" json:"restored,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *CriuDumpResp) Reset() { *m = CriuDumpResp{} } +func (m *CriuDumpResp) String() string { return proto.CompactTextString(m) } +func (*CriuDumpResp) ProtoMessage() {} +func (*CriuDumpResp) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{8} } + +func (m *CriuDumpResp) GetRestored() bool { + if m != nil && m.Restored != nil { + return *m.Restored + } + return false +} + +type CriuRestoreResp struct { + Pid *int32 `protobuf:"varint,1,req,name=pid" json:"pid,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *CriuRestoreResp) Reset() { *m = CriuRestoreResp{} } +func (m *CriuRestoreResp) String() string { return proto.CompactTextString(m) } +func (*CriuRestoreResp) ProtoMessage() {} +func (*CriuRestoreResp) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{9} } + +func (m *CriuRestoreResp) GetPid() int32 { + if m != nil && m.Pid != nil { + return *m.Pid + } + return 0 +} + +type CriuNotify struct { + Script *string `protobuf:"bytes,1,opt,name=script" json:"script,omitempty"` + Pid *int32 `protobuf:"varint,2,opt,name=pid" json:"pid,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *CriuNotify) Reset() { *m = CriuNotify{} } +func (m *CriuNotify) String() string { return proto.CompactTextString(m) } +func (*CriuNotify) ProtoMessage() {} +func (*CriuNotify) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{10} } + +func (m *CriuNotify) GetScript() string { + if m != nil && m.Script != nil { + return *m.Script + } + return "" +} + +func (m *CriuNotify) GetPid() int32 { + if m != nil && m.Pid != nil { + return *m.Pid + } + return 0 +} + +// +// List of features which can queried via +// CRIU_REQ_TYPE__FEATURE_CHECK +type CriuFeatures struct { + MemTrack *bool `protobuf:"varint,1,opt,name=mem_track,json=memTrack" json:"mem_track,omitempty"` + LazyPages *bool `protobuf:"varint,2,opt,name=lazy_pages,json=lazyPages" json:"lazy_pages,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *CriuFeatures) Reset() { *m = CriuFeatures{} } +func (m *CriuFeatures) String() string { return proto.CompactTextString(m) } +func (*CriuFeatures) ProtoMessage() {} +func (*CriuFeatures) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{11} } + +func (m *CriuFeatures) GetMemTrack() bool { + if m != nil && m.MemTrack != nil { + return *m.MemTrack + } + return false +} + +func (m *CriuFeatures) GetLazyPages() bool { + if m != nil && m.LazyPages != nil { + return *m.LazyPages + } + return false +} + +type CriuReq struct { + Type *CriuReqType `protobuf:"varint,1,req,name=type,enum=CriuReqType" json:"type,omitempty"` + Opts *CriuOpts `protobuf:"bytes,2,opt,name=opts" json:"opts,omitempty"` + NotifySuccess *bool `protobuf:"varint,3,opt,name=notify_success,json=notifySuccess" json:"notify_success,omitempty"` + // + // When set service won't close the connection but + // will wait for more req-s to appear. Works not + // for all request types. + KeepOpen *bool `protobuf:"varint,4,opt,name=keep_open,json=keepOpen" json:"keep_open,omitempty"` + // + // 'features' can be used to query which features + // are supported by the installed criu/kernel + // via RPC. + Features *CriuFeatures `protobuf:"bytes,5,opt,name=features" json:"features,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *CriuReq) Reset() { *m = CriuReq{} } +func (m *CriuReq) String() string { return proto.CompactTextString(m) } +func (*CriuReq) ProtoMessage() {} +func (*CriuReq) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{12} } + +func (m *CriuReq) GetType() CriuReqType { + if m != nil && m.Type != nil { + return *m.Type + } + return CriuReqType_EMPTY +} + +func (m *CriuReq) GetOpts() *CriuOpts { + if m != nil { + return m.Opts + } + return nil +} + +func (m *CriuReq) GetNotifySuccess() bool { + if m != nil && m.NotifySuccess != nil { + return *m.NotifySuccess + } + return false +} + +func (m *CriuReq) GetKeepOpen() bool { + if m != nil && m.KeepOpen != nil { + return *m.KeepOpen + } + return false +} + +func (m *CriuReq) GetFeatures() *CriuFeatures { + if m != nil { + return m.Features + } + return nil +} + +type CriuResp struct { + Type *CriuReqType `protobuf:"varint,1,req,name=type,enum=CriuReqType" json:"type,omitempty"` + Success *bool `protobuf:"varint,2,req,name=success" json:"success,omitempty"` + Dump *CriuDumpResp `protobuf:"bytes,3,opt,name=dump" json:"dump,omitempty"` + Restore *CriuRestoreResp `protobuf:"bytes,4,opt,name=restore" json:"restore,omitempty"` + Notify *CriuNotify `protobuf:"bytes,5,opt,name=notify" json:"notify,omitempty"` + Ps *CriuPageServerInfo `protobuf:"bytes,6,opt,name=ps" json:"ps,omitempty"` + CrErrno *int32 `protobuf:"varint,7,opt,name=cr_errno,json=crErrno" json:"cr_errno,omitempty"` + Features *CriuFeatures `protobuf:"bytes,8,opt,name=features" json:"features,omitempty"` + CrErrmsg *string `protobuf:"bytes,9,opt,name=cr_errmsg,json=crErrmsg" json:"cr_errmsg,omitempty"` + Version *CriuVersion `protobuf:"bytes,10,opt,name=version" json:"version,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *CriuResp) Reset() { *m = CriuResp{} } +func (m *CriuResp) String() string { return proto.CompactTextString(m) } +func (*CriuResp) ProtoMessage() {} +func (*CriuResp) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{13} } + +func (m *CriuResp) GetType() CriuReqType { + if m != nil && m.Type != nil { + return *m.Type + } + return CriuReqType_EMPTY +} + +func (m *CriuResp) GetSuccess() bool { + if m != nil && m.Success != nil { + return *m.Success + } + return false +} + +func (m *CriuResp) GetDump() *CriuDumpResp { + if m != nil { + return m.Dump + } + return nil +} + +func (m *CriuResp) GetRestore() *CriuRestoreResp { + if m != nil { + return m.Restore + } + return nil +} + +func (m *CriuResp) GetNotify() *CriuNotify { + if m != nil { + return m.Notify + } + return nil +} + +func (m *CriuResp) GetPs() *CriuPageServerInfo { + if m != nil { + return m.Ps + } + return nil +} + +func (m *CriuResp) GetCrErrno() int32 { + if m != nil && m.CrErrno != nil { + return *m.CrErrno + } + return 0 +} + +func (m *CriuResp) GetFeatures() *CriuFeatures { + if m != nil { + return m.Features + } + return nil +} + +func (m *CriuResp) GetCrErrmsg() string { + if m != nil && m.CrErrmsg != nil { + return *m.CrErrmsg + } + return "" +} + +func (m *CriuResp) GetVersion() *CriuVersion { + if m != nil { + return m.Version + } + return nil +} + +// Answer for criu_req_type.VERSION requests +type CriuVersion struct { + Major *int32 `protobuf:"varint,1,req,name=major" json:"major,omitempty"` + Minor *int32 `protobuf:"varint,2,req,name=minor" json:"minor,omitempty"` + Gitid *string `protobuf:"bytes,3,opt,name=gitid" json:"gitid,omitempty"` + Sublevel *int32 `protobuf:"varint,4,opt,name=sublevel" json:"sublevel,omitempty"` + Extra *int32 `protobuf:"varint,5,opt,name=extra" json:"extra,omitempty"` + Name *string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *CriuVersion) Reset() { *m = CriuVersion{} } +func (m *CriuVersion) String() string { return proto.CompactTextString(m) } +func (*CriuVersion) ProtoMessage() {} +func (*CriuVersion) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{14} } + +func (m *CriuVersion) GetMajor() int32 { + if m != nil && m.Major != nil { + return *m.Major + } + return 0 +} + +func (m *CriuVersion) GetMinor() int32 { + if m != nil && m.Minor != nil { + return *m.Minor + } + return 0 +} + +func (m *CriuVersion) GetGitid() string { + if m != nil && m.Gitid != nil { + return *m.Gitid + } + return "" +} + +func (m *CriuVersion) GetSublevel() int32 { + if m != nil && m.Sublevel != nil { + return *m.Sublevel + } + return 0 +} + +func (m *CriuVersion) GetExtra() int32 { + if m != nil && m.Extra != nil { + return *m.Extra + } + return 0 +} + +func (m *CriuVersion) GetName() string { + if m != nil && m.Name != nil { + return *m.Name + } + return "" +} + +func init() { + proto.RegisterType((*CriuPageServerInfo)(nil), "criu_page_server_info") + proto.RegisterType((*CriuVethPair)(nil), "criu_veth_pair") + proto.RegisterType((*ExtMountMap)(nil), "ext_mount_map") + proto.RegisterType((*JoinNamespace)(nil), "join_namespace") + proto.RegisterType((*InheritFd)(nil), "inherit_fd") + proto.RegisterType((*CgroupRoot)(nil), "cgroup_root") + proto.RegisterType((*UnixSk)(nil), "unix_sk") + proto.RegisterType((*CriuOpts)(nil), "criu_opts") + proto.RegisterType((*CriuDumpResp)(nil), "criu_dump_resp") + proto.RegisterType((*CriuRestoreResp)(nil), "criu_restore_resp") + proto.RegisterType((*CriuNotify)(nil), "criu_notify") + proto.RegisterType((*CriuFeatures)(nil), "criu_features") + proto.RegisterType((*CriuReq)(nil), "criu_req") + proto.RegisterType((*CriuResp)(nil), "criu_resp") + proto.RegisterType((*CriuVersion)(nil), "criu_version") + proto.RegisterEnum("CriuCgMode", CriuCgMode_name, CriuCgMode_value) + proto.RegisterEnum("CriuReqType", CriuReqType_name, CriuReqType_value) +} + +func init() { proto.RegisterFile("criurpc.proto", fileDescriptor0) } + +var fileDescriptor0 = []byte{ + // 1781 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0x56, 0xdd, 0x72, 0x5b, 0xb7, + 0x11, 0x0e, 0x29, 0xfe, 0x1c, 0x82, 0x3f, 0xa6, 0x10, 0xdb, 0x81, 0x93, 0xda, 0x62, 0xe8, 0x28, + 0x51, 0x15, 0x97, 0x4d, 0x58, 0x3b, 0xae, 0x33, 0xed, 0x85, 0x47, 0x22, 0x5d, 0x36, 0x92, 0xc8, + 0x01, 0x25, 0xcf, 0xe4, 0x0a, 0x73, 0x74, 0x0e, 0x48, 0xc1, 0x3c, 0x7f, 0x05, 0x40, 0x45, 0xf2, + 0x83, 0xf4, 0x29, 0xfa, 0x0c, 0x7d, 0x84, 0xbe, 0x4e, 0x6f, 0x3b, 0xbb, 0x00, 0x65, 0x29, 0xc9, + 0xb4, 0xbd, 0xc3, 0x7e, 0x58, 0x00, 0xbb, 0xfb, 0xed, 0x0f, 0x48, 0x3b, 0xd2, 0x6a, 0xad, 0x8b, + 0x68, 0x50, 0xe8, 0xdc, 0xe6, 0xfd, 0x25, 0x79, 0x00, 0x80, 0x28, 0xc2, 0xa5, 0x14, 0x46, 0xea, + 0x4b, 0xa9, 0x85, 0xca, 0x16, 0x39, 0x65, 0xa4, 0x1e, 0xc6, 0xb1, 0x96, 0xc6, 0xb0, 0x52, 0xaf, + 0xb4, 0xd7, 0xe0, 0x1b, 0x91, 0x52, 0x52, 0x29, 0x72, 0x6d, 0x59, 0xb9, 0x57, 0xda, 0xab, 0x72, + 0x5c, 0xd3, 0x2e, 0xd9, 0x2a, 0x54, 0xcc, 0xb6, 0x10, 0x82, 0x25, 0xed, 0x90, 0xf2, 0x22, 0x66, + 0x15, 0x04, 0xca, 0x8b, 0xb8, 0xff, 0x27, 0xd2, 0xc1, 0x87, 0x2e, 0xa5, 0xbd, 0x10, 0x45, 0xa8, + 0x34, 0xfd, 0x98, 0x54, 0xd5, 0x42, 0xa8, 0x8c, 0x95, 0x7a, 0xe5, 0xbd, 0x06, 0xaf, 0xa8, 0xc5, + 0x24, 0xa3, 0x0f, 0x48, 0x4d, 0x2d, 0x44, 0xbe, 0x86, 0xeb, 0x01, 0xad, 0xaa, 0xc5, 0x74, 0x6d, + 0xfb, 0x7f, 0x20, 0x6d, 0x79, 0x65, 0x45, 0x9a, 0xaf, 0x33, 0x2b, 0xd2, 0xb0, 0x80, 0x07, 0x57, + 0xf2, 0xda, 0x1f, 0x85, 0x25, 0x20, 0x97, 0x61, 0xe2, 0x8f, 0xc1, 0xb2, 0xff, 0x96, 0x74, 0xde, + 0xe5, 0x2a, 0x13, 0x59, 0x98, 0x4a, 0x53, 0x84, 0x91, 0x04, 0xa3, 0x32, 0xe3, 0x0f, 0x95, 0x33, + 0x43, 0x3f, 0x21, 0xf5, 0xcc, 0x88, 0x85, 0x4a, 0xa4, 0x3f, 0x57, 0xcb, 0xcc, 0x58, 0x25, 0x92, + 0x7e, 0x46, 0x1a, 0xf2, 0xca, 0xea, 0x50, 0xe4, 0x85, 0x45, 0xaf, 0x1a, 0x3c, 0x40, 0x60, 0x5a, + 0xd8, 0xfe, 0x80, 0x10, 0x95, 0x5d, 0x48, 0xad, 0xac, 0x58, 0xc4, 0xbf, 0x62, 0x89, 0x73, 0x1d, + 0x2e, 0x74, 0xae, 0xbf, 0x20, 0xcd, 0x68, 0xa9, 0xf3, 0x75, 0x21, 0x74, 0x9e, 0x5b, 0x88, 0x5f, + 0x64, 0x75, 0xe2, 0xc3, 0x8a, 0x6b, 0x8c, 0x69, 0x68, 0x2f, 0xbc, 0x15, 0xb8, 0xee, 0xef, 0x90, + 0xfa, 0x3a, 0x53, 0x57, 0xc2, 0xac, 0xe8, 0x7d, 0x52, 0x55, 0x59, 0x1e, 0x4b, 0x7c, 0xa5, 0xcd, + 0x9d, 0xd0, 0xff, 0x57, 0x9b, 0x34, 0x30, 0xa6, 0x79, 0x61, 0x0d, 0xed, 0x93, 0xb6, 0x4a, 0xc3, + 0xa5, 0x34, 0x22, 0x56, 0x5a, 0x2c, 0x62, 0xd4, 0xad, 0xf2, 0xa6, 0x03, 0x0f, 0x95, 0x1e, 0xc7, + 0x1b, 0x9a, 0xca, 0x1f, 0x68, 0x7a, 0x4a, 0xda, 0x89, 0x0c, 0x2f, 0xa5, 0xd0, 0xeb, 0x2c, 0x53, + 0xd9, 0x12, 0x9d, 0x0d, 0x78, 0x0b, 0x41, 0xee, 0x30, 0xfa, 0x84, 0x34, 0x21, 0xfa, 0xde, 0x1a, + 0x24, 0x35, 0xe0, 0x10, 0xa0, 0xb3, 0x4c, 0x5d, 0xcd, 0x57, 0xf4, 0x2b, 0x72, 0xcf, 0x46, 0x85, + 0x90, 0xc6, 0x86, 0xe7, 0x89, 0x32, 0x17, 0x32, 0x66, 0x55, 0xd4, 0xe9, 0xd8, 0xa8, 0x18, 0x7d, + 0x40, 0x41, 0x51, 0x5e, 0x86, 0x46, 0x5d, 0x4a, 0x11, 0xcb, 0x4b, 0x15, 0x49, 0xc3, 0x6a, 0x4e, + 0xd1, 0xc3, 0x87, 0x0e, 0x85, 0xf8, 0x9b, 0x0b, 0x99, 0x24, 0xe2, 0x5d, 0x7e, 0xce, 0xea, 0xa8, + 0x12, 0x20, 0xf0, 0xd7, 0xfc, 0x9c, 0x3e, 0x26, 0x04, 0x28, 0x13, 0x49, 0x1e, 0xad, 0x0c, 0x0b, + 0x9c, 0x35, 0x80, 0x1c, 0x01, 0x40, 0x9f, 0x90, 0x46, 0x92, 0x2f, 0x45, 0x22, 0x2f, 0x65, 0xc2, + 0x1a, 0xe0, 0xea, 0xf7, 0xa5, 0x21, 0x0f, 0x92, 0x7c, 0x79, 0x04, 0x10, 0x7d, 0x44, 0x60, 0xed, + 0x58, 0x27, 0x2e, 0xb5, 0x93, 0x7c, 0x89, 0xb4, 0x7f, 0x49, 0xca, 0x85, 0x61, 0xcd, 0x5e, 0x69, + 0xaf, 0x39, 0x7c, 0x38, 0xf8, 0xd5, 0xc2, 0xe0, 0xe5, 0xc2, 0xd0, 0x5d, 0xd2, 0xc9, 0x72, 0xab, + 0x16, 0xd7, 0xc2, 0x44, 0x5a, 0x15, 0xd6, 0xb0, 0x16, 0x5a, 0xd1, 0x76, 0xe8, 0xdc, 0x81, 0xc0, + 0x2a, 0x30, 0xce, 0xda, 0x8e, 0x69, 0x64, 0xff, 0x31, 0x21, 0x45, 0xa8, 0x65, 0x66, 0x85, 0x4a, + 0x97, 0xac, 0x83, 0x3b, 0x0d, 0x87, 0x4c, 0xd2, 0x25, 0x38, 0x6e, 0x75, 0x18, 0xad, 0x44, 0x2a, + 0x53, 0x76, 0xcf, 0x39, 0x8e, 0xc0, 0xb1, 0x4c, 0xe1, 0x6c, 0xb8, 0xb6, 0xb9, 0x88, 0x65, 0xbc, + 0x2e, 0x58, 0xd7, 0x39, 0x0e, 0xc8, 0x21, 0x00, 0x40, 0xd3, 0x4f, 0xb9, 0x5e, 0x6d, 0xf8, 0xdf, + 0x46, 0x96, 0x1b, 0x00, 0x39, 0xf6, 0x1f, 0x13, 0x92, 0xa8, 0x6c, 0x25, 0xb4, 0x4c, 0xc3, 0x82, + 0x51, 0x77, 0x1c, 0x10, 0x0e, 0x00, 0xdd, 0x25, 0x55, 0x28, 0x4e, 0xc3, 0x3e, 0xee, 0x6d, 0xed, + 0x35, 0x87, 0xf7, 0x06, 0x77, 0xeb, 0x95, 0xbb, 0x5d, 0xfa, 0x94, 0xd4, 0xa3, 0x62, 0x2d, 0xa2, + 0xb0, 0x60, 0xf7, 0x7b, 0xa5, 0xbd, 0xf6, 0xf7, 0xe4, 0xf9, 0xf0, 0xd5, 0xf3, 0x57, 0xdf, 0xbd, + 0x1c, 0xbe, 0x7a, 0xc1, 0x6b, 0x51, 0xb1, 0x3e, 0x08, 0x0b, 0xba, 0x43, 0x9a, 0x8b, 0x5c, 0x47, + 0x52, 0x28, 0x0d, 0x6f, 0x3d, 0xc0, 0xb7, 0x08, 0x42, 0x13, 0x40, 0x80, 0x04, 0x79, 0x25, 0x23, + 0x11, 0xa5, 0x31, 0x7b, 0xd8, 0xdb, 0x02, 0x12, 0x40, 0x3e, 0x48, 0x21, 0x49, 0xea, 0x58, 0xeb, + 0x99, 0x65, 0x9f, 0xa0, 0x25, 0x9d, 0xc1, 0x9d, 0xda, 0xe7, 0x35, 0x79, 0x65, 0x8f, 0x33, 0x0b, + 0x2c, 0xa4, 0x61, 0x06, 0xfc, 0xb8, 0xf2, 0x32, 0x8c, 0x39, 0x16, 0x1c, 0x7a, 0xe0, 0x40, 0xba, + 0x4b, 0xea, 0xd1, 0x12, 0x4b, 0x8f, 0x3d, 0xc2, 0xfb, 0x5a, 0x83, 0x5b, 0xe5, 0xc8, 0x6b, 0xd1, + 0x92, 0x03, 0x31, 0x3b, 0xa4, 0xa9, 0x8d, 0x15, 0x46, 0x9d, 0x27, 0x50, 0x07, 0x9f, 0x3a, 0x93, + 0xb5, 0xb1, 0x73, 0x87, 0xd0, 0xfd, 0xdb, 0x65, 0xcf, 0x3e, 0xc3, 0xab, 0x9a, 0x83, 0x0f, 0x10, + 0x6f, 0xf8, 0xf5, 0x38, 0xa6, 0x3d, 0xd2, 0x42, 0xa6, 0x36, 0x8e, 0xfc, 0xc6, 0xdd, 0x06, 0xd8, + 0xc8, 0x19, 0xbf, 0xe3, 0x6a, 0xca, 0x5c, 0x84, 0x1a, 0x9e, 0x7b, 0xec, 0x14, 0xe4, 0x95, 0x9d, + 0x3b, 0x64, 0xa3, 0x90, 0x86, 0xc6, 0x4a, 0x6d, 0xd8, 0x93, 0x1b, 0x85, 0x63, 0x87, 0x40, 0x08, + 0xcd, 0x4a, 0x15, 0x78, 0xff, 0x8e, 0x0b, 0x21, 0xc8, 0x70, 0x39, 0xb4, 0xaf, 0x2c, 0x3c, 0x4f, + 0xa4, 0x58, 0x18, 0xd6, 0xc3, 0xbd, 0xc0, 0x01, 0x63, 0x43, 0xf7, 0x48, 0xd3, 0x57, 0xb2, 0x50, + 0x59, 0xce, 0x3e, 0x47, 0x47, 0x82, 0x81, 0xc7, 0x78, 0x63, 0x8d, 0x45, 0x3d, 0xc9, 0x72, 0xfa, + 0x67, 0xf2, 0xf1, 0xdd, 0x00, 0x8b, 0x14, 0x9a, 0x50, 0xbf, 0x57, 0xda, 0xeb, 0x0c, 0xdb, 0x2e, + 0x3f, 0xa2, 0x25, 0x82, 0x7c, 0xfb, 0x4e, 0xd0, 0x8f, 0xf3, 0x58, 0xc2, 0x43, 0xcb, 0x8b, 0xdc, + 0x58, 0x91, 0xa8, 0x54, 0x59, 0xf6, 0x14, 0xb3, 0xa5, 0xfe, 0xed, 0x37, 0xcf, 0xff, 0xf8, 0xe2, + 0xe5, 0x77, 0x9c, 0xe0, 0xde, 0x11, 0x6c, 0xd1, 0x3d, 0xd2, 0xc5, 0x44, 0x11, 0x26, 0x0a, 0x33, + 0x01, 0xdd, 0xcf, 0xb0, 0x2f, 0xd0, 0xec, 0x0e, 0xe2, 0xf3, 0x28, 0xcc, 0x66, 0x80, 0xd2, 0x4f, + 0x21, 0x6f, 0xac, 0xd4, 0x59, 0x98, 0xb0, 0x5d, 0xef, 0x98, 0x97, 0x31, 0xa7, 0xd2, 0xc2, 0x5e, + 0x8b, 0xcc, 0xb0, 0x2f, 0xe1, 0x31, 0x5e, 0x47, 0xf9, 0x04, 0x7c, 0xae, 0xbb, 0x51, 0x60, 0xd8, + 0x57, 0x3e, 0xbb, 0xef, 0x8e, 0x06, 0x5e, 0x03, 0xf9, 0xc4, 0xd0, 0xcf, 0x49, 0xcb, 0x67, 0x47, + 0xa1, 0xf3, 0xc2, 0xb0, 0xdf, 0x62, 0x85, 0xfa, 0x06, 0x3e, 0x03, 0x88, 0xee, 0x93, 0xed, 0xdb, + 0x2a, 0xae, 0x93, 0xec, 0xa3, 0xde, 0xbd, 0x5b, 0x7a, 0xd8, 0x51, 0x9e, 0x93, 0x87, 0x5e, 0x37, + 0x5e, 0xa7, 0x85, 0x88, 0xf2, 0xcc, 0xea, 0x3c, 0x49, 0xa4, 0x66, 0x5f, 0xa3, 0xf5, 0xf7, 0xdd, + 0xee, 0xe1, 0x3a, 0x2d, 0x0e, 0x6e, 0xf6, 0xa0, 0x2b, 0x2f, 0xb4, 0x94, 0xef, 0x37, 0x81, 0x67, + 0xcf, 0xf0, 0xf6, 0x96, 0x03, 0x5d, 0x8c, 0x61, 0x42, 0x5b, 0x95, 0x4a, 0x98, 0x95, 0xbf, 0x73, + 0xde, 0x7a, 0x91, 0x7e, 0x4d, 0x28, 0xf4, 0x63, 0xcc, 0x0e, 0x95, 0x89, 0x45, 0xa2, 0x96, 0x17, + 0x96, 0x0d, 0x30, 0x83, 0xa0, 0x53, 0xcf, 0x57, 0xaa, 0x98, 0x64, 0x63, 0x84, 0xc1, 0xe1, 0x9f, + 0x64, 0xb8, 0x12, 0xe6, 0xda, 0x44, 0x36, 0x31, 0xec, 0xf7, 0xa8, 0xd6, 0x04, 0x6c, 0xee, 0x20, + 0x6c, 0x1c, 0xe1, 0xfb, 0x6b, 0xec, 0x85, 0x86, 0x7d, 0xe3, 0x1b, 0x47, 0xf8, 0xfe, 0x7a, 0x06, + 0x00, 0x36, 0x6b, 0x1b, 0xda, 0xb5, 0x81, 0xba, 0xf8, 0x16, 0xbb, 0x4e, 0xe0, 0x80, 0x71, 0x0c, + 0xc1, 0xca, 0x75, 0x71, 0x01, 0xb4, 0x5a, 0xe3, 0xb3, 0x99, 0x0d, 0x9d, 0x29, 0x6e, 0x63, 0x66, + 0x8d, 0x4b, 0xe9, 0xfe, 0x33, 0xff, 0x47, 0xc0, 0x50, 0x69, 0x69, 0x0a, 0xa0, 0x5b, 0x4b, 0x63, + 0x73, 0x2d, 0x63, 0x9c, 0x97, 0x01, 0xbf, 0x91, 0xfb, 0xbb, 0x64, 0x1b, 0xb5, 0x3d, 0xe0, 0x0e, + 0xf8, 0x09, 0xe7, 0x66, 0x1f, 0x2c, 0xfb, 0x2f, 0x49, 0x13, 0xd5, 0x5c, 0x6b, 0xa6, 0x0f, 0x49, + 0xcd, 0xf5, 0x6c, 0x3f, 0x7f, 0xbd, 0xf4, 0xcb, 0xd1, 0xd8, 0xff, 0xc1, 0xfd, 0x95, 0xc4, 0x42, + 0x86, 0x76, 0xad, 0x9d, 0x9f, 0xa9, 0x4c, 0x05, 0xb6, 0xe3, 0x8d, 0x35, 0xa9, 0x4c, 0x4f, 0x41, + 0xfe, 0x59, 0x8c, 0xca, 0x3f, 0x8b, 0x51, 0xff, 0x9f, 0x25, 0x12, 0x78, 0x6b, 0xff, 0x46, 0xfb, + 0xa4, 0x62, 0xaf, 0x0b, 0x37, 0xcd, 0x3b, 0xc3, 0xce, 0x60, 0xb3, 0x21, 0x00, 0xe5, 0xb8, 0x47, + 0x9f, 0x90, 0x0a, 0x8c, 0x75, 0xbc, 0xa9, 0x39, 0x24, 0x83, 0x9b, 0x41, 0xcf, 0x11, 0xbf, 0x3d, + 0x82, 0xd6, 0x51, 0x04, 0xdf, 0xb4, 0xad, 0x3b, 0x23, 0xc8, 0x81, 0x60, 0xf3, 0x4a, 0xca, 0x42, + 0xe4, 0x85, 0xcc, 0xfc, 0xe0, 0x0e, 0x00, 0x98, 0x16, 0x32, 0xa3, 0xfb, 0x24, 0xd8, 0x38, 0x87, + 0x03, 0xbb, 0xb9, 0xb1, 0x65, 0x83, 0xf2, 0x9b, 0xfd, 0xfe, 0xbf, 0xcb, 0xfe, 0xb3, 0x81, 0x61, + 0xfe, 0x7f, 0x3c, 0x60, 0xa4, 0xbe, 0x31, 0x0d, 0xbe, 0x35, 0x01, 0xdf, 0x88, 0xf4, 0x29, 0xa9, + 0x00, 0xc5, 0x68, 0xf1, 0xcd, 0xa0, 0xb9, 0x21, 0x9d, 0xe3, 0x26, 0x7d, 0x46, 0xea, 0x9e, 0x59, + 0xb4, 0xbb, 0x39, 0xa4, 0x83, 0x5f, 0xd0, 0xcd, 0x37, 0x2a, 0xf4, 0x0b, 0x52, 0x73, 0x8e, 0x7b, + 0x47, 0x5a, 0x83, 0x5b, 0xa4, 0x73, 0xbf, 0xe7, 0xe7, 0x7b, 0xed, 0x7f, 0xce, 0xf7, 0x47, 0x40, + 0x96, 0x90, 0x5a, 0x67, 0x39, 0xfe, 0x3e, 0xaa, 0xbc, 0x1e, 0xe9, 0x11, 0x88, 0x77, 0x62, 0x16, + 0xfc, 0xf7, 0x98, 0x41, 0xf0, 0xdd, 0x35, 0xa9, 0x59, 0xe2, 0x4f, 0xa4, 0xc1, 0x03, 0xbc, 0x27, + 0x35, 0x4b, 0x18, 0x73, 0x97, 0x52, 0x1b, 0x95, 0x67, 0xf8, 0x0b, 0x69, 0x6e, 0x1a, 0xaa, 0x07, + 0xf9, 0x66, 0xb7, 0xff, 0xf7, 0x12, 0x69, 0xdd, 0xde, 0x81, 0xdf, 0x60, 0x1a, 0xbe, 0xcb, 0xb5, + 0xcf, 0x72, 0x27, 0x20, 0xaa, 0xb2, 0x5c, 0xfb, 0x8f, 0xa7, 0x13, 0x00, 0x5d, 0x2a, 0xeb, 0xbf, + 0xe6, 0x0d, 0xee, 0x04, 0x28, 0x2b, 0xb3, 0x3e, 0x77, 0x3f, 0xa4, 0x8a, 0x2f, 0x58, 0x2f, 0xc3, + 0x09, 0xfc, 0xe9, 0x62, 0x20, 0xab, 0xdc, 0x09, 0xf0, 0x95, 0x81, 0x5e, 0x89, 0xb1, 0x6b, 0x70, + 0x5c, 0xef, 0x0b, 0x6f, 0x97, 0x1f, 0x01, 0x94, 0x90, 0xda, 0xe4, 0xcd, 0xc9, 0x94, 0x8f, 0xba, + 0x1f, 0xd1, 0x26, 0xa9, 0x1f, 0xbc, 0x11, 0x27, 0xd3, 0x93, 0x51, 0xb7, 0x44, 0x1b, 0xa4, 0x3a, + 0xe3, 0xd3, 0xd9, 0xbc, 0x5b, 0xa6, 0x01, 0xa9, 0xcc, 0xa7, 0xe3, 0xd3, 0xee, 0x16, 0xac, 0xc6, + 0x67, 0x47, 0x47, 0xdd, 0x0a, 0x9c, 0x9b, 0x9f, 0xf2, 0xc9, 0xc1, 0x69, 0xb7, 0x0a, 0xe7, 0x0e, + 0x47, 0xe3, 0xd7, 0x67, 0x47, 0xa7, 0xdd, 0xda, 0xfe, 0x3f, 0x4a, 0xbe, 0x04, 0x37, 0x99, 0x05, + 0x37, 0x8d, 0x8e, 0x67, 0xa7, 0x3f, 0x76, 0x3f, 0x82, 0xf3, 0x87, 0x67, 0xc7, 0xb3, 0x6e, 0x09, + 0xce, 0xf0, 0xd1, 0xfc, 0x14, 0x1e, 0x2e, 0x83, 0xc6, 0xc1, 0x5f, 0x46, 0x07, 0x3f, 0x74, 0xb7, + 0x68, 0x8b, 0x04, 0x33, 0x3e, 0x12, 0xa8, 0x55, 0xa1, 0xf7, 0x48, 0x73, 0xf6, 0xfa, 0xcd, 0x48, + 0xcc, 0x47, 0xfc, 0xed, 0x88, 0x77, 0xab, 0xf0, 0xec, 0xc9, 0xf4, 0x74, 0x32, 0xfe, 0xb1, 0x5b, + 0xa3, 0x5d, 0xd2, 0x3a, 0x98, 0x9d, 0x4d, 0x4e, 0xc6, 0x53, 0xa7, 0x5e, 0xa7, 0xdb, 0xa4, 0xbd, + 0x41, 0xdc, 0x7d, 0x01, 0x40, 0xe3, 0xd1, 0xeb, 0xd3, 0x33, 0x3e, 0xf2, 0x50, 0x03, 0x9e, 0x7e, + 0x3b, 0xe2, 0xf3, 0xc9, 0xf4, 0xa4, 0x4b, 0xfe, 0x13, 0x00, 0x00, 0xff, 0xff, 0x5f, 0x2a, 0xaf, + 0x49, 0x5b, 0x0d, 0x00, 0x00, +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/criurpc/criurpc.proto b/vendor/github.com/opencontainers/runc/libcontainer/criurpc/criurpc.proto new file mode 100644 index 000000000..48e42e26e --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/criurpc/criurpc.proto @@ -0,0 +1,209 @@ +syntax = "proto2"; + +message criu_page_server_info { + optional string address = 1; + optional int32 port = 2; + optional int32 pid = 3; + optional int32 fd = 4; +} + +message criu_veth_pair { + required string if_in = 1; + required string if_out = 2; +}; + +message ext_mount_map { + required string key = 1; + required string val = 2; +}; + +message join_namespace { + required string ns = 1; + required string ns_file = 2; + optional string extra_opt = 3; +} + +message inherit_fd { + required string key = 1; + required int32 fd = 2; +}; + +message cgroup_root { + optional string ctrl = 1; + required string path = 2; +}; + +message unix_sk { + required uint32 inode = 1; +}; + +enum criu_cg_mode { + IGNORE = 0; + CG_NONE = 1; + PROPS = 2; + SOFT = 3; + FULL = 4; + STRICT = 5; + DEFAULT = 6; +}; + +message criu_opts { + required int32 images_dir_fd = 1; + optional int32 pid = 2; /* if not set on dump, will dump requesting process */ + + optional bool leave_running = 3; + optional bool ext_unix_sk = 4; + optional bool tcp_established = 5; + optional bool evasive_devices = 6; + optional bool shell_job = 7; + optional bool file_locks = 8; + optional int32 log_level = 9 [default = 2]; + optional string log_file = 10; /* No subdirs are allowed. Consider using work-dir */ + + optional criu_page_server_info ps = 11; + + optional bool notify_scripts = 12; + + optional string root = 13; + optional string parent_img = 14; + optional bool track_mem = 15; + optional bool auto_dedup = 16; + + optional int32 work_dir_fd = 17; + optional bool link_remap = 18; + repeated criu_veth_pair veths = 19; /* DEPRECATED, use external instead */ + + optional uint32 cpu_cap = 20 [default = 0xffffffff]; + optional bool force_irmap = 21; + repeated string exec_cmd = 22; + + repeated ext_mount_map ext_mnt = 23; /* DEPRECATED, use external instead */ + optional bool manage_cgroups = 24; /* backward compatibility */ + repeated cgroup_root cg_root = 25; + + optional bool rst_sibling = 26; /* swrk only */ + repeated inherit_fd inherit_fd = 27; /* swrk only */ + + optional bool auto_ext_mnt = 28; + optional bool ext_sharing = 29; + optional bool ext_masters = 30; + + repeated string skip_mnt = 31; + repeated string enable_fs = 32; + + repeated unix_sk unix_sk_ino = 33; /* DEPRECATED, use external instead */ + + optional criu_cg_mode manage_cgroups_mode = 34; + optional uint32 ghost_limit = 35 [default = 0x100000]; + repeated string irmap_scan_paths = 36; + repeated string external = 37; + optional uint32 empty_ns = 38; + repeated join_namespace join_ns = 39; + + optional string cgroup_props = 41; + optional string cgroup_props_file = 42; + repeated string cgroup_dump_controller = 43; + + optional string freeze_cgroup = 44; + optional uint32 timeout = 45; + optional bool tcp_skip_in_flight = 46; + optional bool weak_sysctls = 47; + optional bool lazy_pages = 48; + optional int32 status_fd = 49; + optional bool orphan_pts_master = 50; +} + +message criu_dump_resp { + optional bool restored = 1; +} + +message criu_restore_resp { + required int32 pid = 1; +} + +message criu_notify { + optional string script = 1; + optional int32 pid = 2; +} + +enum criu_req_type { + EMPTY = 0; + DUMP = 1; + RESTORE = 2; + CHECK = 3; + PRE_DUMP = 4; + PAGE_SERVER = 5; + + NOTIFY = 6; + + CPUINFO_DUMP = 7; + CPUINFO_CHECK = 8; + + FEATURE_CHECK = 9; + + VERSION = 10; +} + +/* + * List of features which can queried via + * CRIU_REQ_TYPE__FEATURE_CHECK + */ +message criu_features { + optional bool mem_track = 1; + optional bool lazy_pages = 2; +} + +/* + * Request -- each type corresponds to must-be-there + * request arguments of respective type + */ + +message criu_req { + required criu_req_type type = 1; + + optional criu_opts opts = 2; + optional bool notify_success = 3; + + /* + * When set service won't close the connection but + * will wait for more req-s to appear. Works not + * for all request types. + */ + optional bool keep_open = 4; + /* + * 'features' can be used to query which features + * are supported by the installed criu/kernel + * via RPC. + */ + optional criu_features features = 5; +} + +/* + * Response -- it states whether the request was served + * and additional request-specific information + */ + +message criu_resp { + required criu_req_type type = 1; + required bool success = 2; + + optional criu_dump_resp dump = 3; + optional criu_restore_resp restore = 4; + optional criu_notify notify = 5; + optional criu_page_server_info ps = 6; + + optional int32 cr_errno = 7; + optional criu_features features = 8; + optional string cr_errmsg = 9; + optional criu_version version = 10; +} + +/* Answer for criu_req_type.VERSION requests */ +message criu_version { + required int32 major = 1; + required int32 minor = 2; + optional string gitid = 3; + optional int32 sublevel = 4; + optional int32 extra = 5; + optional string name = 6; +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/error.go b/vendor/github.com/opencontainers/runc/libcontainer/error.go new file mode 100644 index 000000000..21a3789ba --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/error.go @@ -0,0 +1,70 @@ +package libcontainer + +import "io" + +// ErrorCode is the API error code type. +type ErrorCode int + +// API error codes. +const ( + // Factory errors + IdInUse ErrorCode = iota + InvalidIdFormat + + // Container errors + ContainerNotExists + ContainerPaused + ContainerNotStopped + ContainerNotRunning + ContainerNotPaused + + // Process errors + NoProcessOps + + // Common errors + ConfigInvalid + ConsoleExists + SystemError +) + +func (c ErrorCode) String() string { + switch c { + case IdInUse: + return "Id already in use" + case InvalidIdFormat: + return "Invalid format" + case ContainerPaused: + return "Container paused" + case ConfigInvalid: + return "Invalid configuration" + case SystemError: + return "System error" + case ContainerNotExists: + return "Container does not exist" + case ContainerNotStopped: + return "Container is not stopped" + case ContainerNotRunning: + return "Container is not running" + case ConsoleExists: + return "Console exists for process" + case ContainerNotPaused: + return "Container is not paused" + case NoProcessOps: + return "No process operations" + default: + return "Unknown error" + } +} + +// Error is the API error type. +type Error interface { + error + + // Returns an error if it failed to write the detail of the Error to w. + // The detail of the Error may include the error message and a + // representation of the stack trace. + Detail(w io.Writer) error + + // Returns the error code for this error. + Code() ErrorCode +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/factory.go b/vendor/github.com/opencontainers/runc/libcontainer/factory.go new file mode 100644 index 000000000..0986cd77e --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/factory.go @@ -0,0 +1,44 @@ +package libcontainer + +import ( + "github.com/opencontainers/runc/libcontainer/configs" +) + +type Factory interface { + // Creates a new container with the given id and starts the initial process inside it. + // id must be a string containing only letters, digits and underscores and must contain + // between 1 and 1024 characters, inclusive. + // + // The id must not already be in use by an existing container. Containers created using + // a factory with the same path (and filesystem) must have distinct ids. + // + // Returns the new container with a running process. + // + // errors: + // IdInUse - id is already in use by a container + // InvalidIdFormat - id has incorrect format + // ConfigInvalid - config is invalid + // Systemerror - System error + // + // On error, any partially created container parts are cleaned up (the operation is atomic). + Create(id string, config *configs.Config) (Container, error) + + // Load takes an ID for an existing container and returns the container information + // from the state. This presents a read only view of the container. + // + // errors: + // Path does not exist + // System error + Load(id string) (Container, error) + + // StartInitialization is an internal API to libcontainer used during the reexec of the + // container. + // + // Errors: + // Pipe connection error + // System error + StartInitialization() error + + // Type returns info string about factory type (e.g. lxc, libcontainer...) + Type() string +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/factory_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/factory_linux.go new file mode 100644 index 000000000..4cbd70cd7 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/factory_linux.go @@ -0,0 +1,381 @@ +// +build linux + +package libcontainer + +import ( + "encoding/json" + "fmt" + "os" + "path/filepath" + "regexp" + "runtime/debug" + "strconv" + + "github.com/opencontainers/runc/libcontainer/cgroups" + "github.com/opencontainers/runc/libcontainer/cgroups/fs" + "github.com/opencontainers/runc/libcontainer/cgroups/systemd" + "github.com/opencontainers/runc/libcontainer/configs" + "github.com/opencontainers/runc/libcontainer/configs/validate" + "github.com/opencontainers/runc/libcontainer/intelrdt" + "github.com/opencontainers/runc/libcontainer/mount" + "github.com/opencontainers/runc/libcontainer/utils" + + "golang.org/x/sys/unix" +) + +const ( + stateFilename = "state.json" + execFifoFilename = "exec.fifo" +) + +var idRegex = regexp.MustCompile(`^[\w+-\.]+$`) + +// InitArgs returns an options func to configure a LinuxFactory with the +// provided init binary path and arguments. +func InitArgs(args ...string) func(*LinuxFactory) error { + return func(l *LinuxFactory) (err error) { + if len(args) > 0 { + // Resolve relative paths to ensure that its available + // after directory changes. + if args[0], err = filepath.Abs(args[0]); err != nil { + return newGenericError(err, ConfigInvalid) + } + } + + l.InitArgs = args + return nil + } +} + +// SystemdCgroups is an options func to configure a LinuxFactory to return +// containers that use systemd to create and manage cgroups. +func SystemdCgroups(l *LinuxFactory) error { + l.NewCgroupsManager = func(config *configs.Cgroup, paths map[string]string) cgroups.Manager { + return &systemd.Manager{ + Cgroups: config, + Paths: paths, + } + } + return nil +} + +// Cgroupfs is an options func to configure a LinuxFactory to return containers +// that use the native cgroups filesystem implementation to create and manage +// cgroups. +func Cgroupfs(l *LinuxFactory) error { + l.NewCgroupsManager = func(config *configs.Cgroup, paths map[string]string) cgroups.Manager { + return &fs.Manager{ + Cgroups: config, + Paths: paths, + } + } + return nil +} + +// RootlessCgroupfs is an options func to configure a LinuxFactory to return +// containers that use the native cgroups filesystem implementation to create +// and manage cgroups. The difference between RootlessCgroupfs and Cgroupfs is +// that RootlessCgroupfs can transparently handle permission errors that occur +// during rootless container setup (while still allowing cgroup usage if +// they've been set up properly). +func RootlessCgroupfs(l *LinuxFactory) error { + l.NewCgroupsManager = func(config *configs.Cgroup, paths map[string]string) cgroups.Manager { + return &fs.Manager{ + Cgroups: config, + Rootless: true, + Paths: paths, + } + } + return nil +} + +// IntelRdtfs is an options func to configure a LinuxFactory to return +// containers that use the Intel RDT "resource control" filesystem to +// create and manage Intel Xeon platform shared resources (e.g., L3 cache). +func IntelRdtFs(l *LinuxFactory) error { + l.NewIntelRdtManager = func(config *configs.Config, id string, path string) intelrdt.Manager { + return &intelrdt.IntelRdtManager{ + Config: config, + Id: id, + Path: path, + } + } + return nil +} + +// TmpfsRoot is an option func to mount LinuxFactory.Root to tmpfs. +func TmpfsRoot(l *LinuxFactory) error { + mounted, err := mount.Mounted(l.Root) + if err != nil { + return err + } + if !mounted { + if err := unix.Mount("tmpfs", l.Root, "tmpfs", 0, ""); err != nil { + return err + } + } + return nil +} + +// CriuPath returns an option func to configure a LinuxFactory with the +// provided criupath +func CriuPath(criupath string) func(*LinuxFactory) error { + return func(l *LinuxFactory) error { + l.CriuPath = criupath + return nil + } +} + +// New returns a linux based container factory based in the root directory and +// configures the factory with the provided option funcs. +func New(root string, options ...func(*LinuxFactory) error) (Factory, error) { + if root != "" { + if err := os.MkdirAll(root, 0700); err != nil { + return nil, newGenericError(err, SystemError) + } + } + l := &LinuxFactory{ + Root: root, + InitPath: "/proc/self/exe", + InitArgs: []string{os.Args[0], "init"}, + Validator: validate.New(), + CriuPath: "criu", + } + Cgroupfs(l) + for _, opt := range options { + if opt == nil { + continue + } + if err := opt(l); err != nil { + return nil, err + } + } + return l, nil +} + +// LinuxFactory implements the default factory interface for linux based systems. +type LinuxFactory struct { + // Root directory for the factory to store state. + Root string + + // InitPath is the path for calling the init responsibilities for spawning + // a container. + InitPath string + + // InitArgs are arguments for calling the init responsibilities for spawning + // a container. + InitArgs []string + + // CriuPath is the path to the criu binary used for checkpoint and restore of + // containers. + CriuPath string + + // New{u,g}uidmapPath is the path to the binaries used for mapping with + // rootless containers. + NewuidmapPath string + NewgidmapPath string + + // Validator provides validation to container configurations. + Validator validate.Validator + + // NewCgroupsManager returns an initialized cgroups manager for a single container. + NewCgroupsManager func(config *configs.Cgroup, paths map[string]string) cgroups.Manager + + // NewIntelRdtManager returns an initialized Intel RDT manager for a single container. + NewIntelRdtManager func(config *configs.Config, id string, path string) intelrdt.Manager +} + +func (l *LinuxFactory) Create(id string, config *configs.Config) (Container, error) { + if l.Root == "" { + return nil, newGenericError(fmt.Errorf("invalid root"), ConfigInvalid) + } + if err := l.validateID(id); err != nil { + return nil, err + } + if err := l.Validator.Validate(config); err != nil { + return nil, newGenericError(err, ConfigInvalid) + } + containerRoot := filepath.Join(l.Root, id) + if _, err := os.Stat(containerRoot); err == nil { + return nil, newGenericError(fmt.Errorf("container with id exists: %v", id), IdInUse) + } else if !os.IsNotExist(err) { + return nil, newGenericError(err, SystemError) + } + if err := os.MkdirAll(containerRoot, 0711); err != nil { + return nil, newGenericError(err, SystemError) + } + if err := os.Chown(containerRoot, unix.Geteuid(), unix.Getegid()); err != nil { + return nil, newGenericError(err, SystemError) + } + c := &linuxContainer{ + id: id, + root: containerRoot, + config: config, + initPath: l.InitPath, + initArgs: l.InitArgs, + criuPath: l.CriuPath, + newuidmapPath: l.NewuidmapPath, + newgidmapPath: l.NewgidmapPath, + cgroupManager: l.NewCgroupsManager(config.Cgroups, nil), + } + if intelrdt.IsEnabled() { + c.intelRdtManager = l.NewIntelRdtManager(config, id, "") + } + c.state = &stoppedState{c: c} + return c, nil +} + +func (l *LinuxFactory) Load(id string) (Container, error) { + if l.Root == "" { + return nil, newGenericError(fmt.Errorf("invalid root"), ConfigInvalid) + } + containerRoot := filepath.Join(l.Root, id) + state, err := l.loadState(containerRoot, id) + if err != nil { + return nil, err + } + r := &nonChildProcess{ + processPid: state.InitProcessPid, + processStartTime: state.InitProcessStartTime, + fds: state.ExternalDescriptors, + } + c := &linuxContainer{ + initProcess: r, + initProcessStartTime: state.InitProcessStartTime, + id: id, + config: &state.Config, + initPath: l.InitPath, + initArgs: l.InitArgs, + criuPath: l.CriuPath, + newuidmapPath: l.NewuidmapPath, + newgidmapPath: l.NewgidmapPath, + cgroupManager: l.NewCgroupsManager(state.Config.Cgroups, state.CgroupPaths), + root: containerRoot, + created: state.Created, + } + c.state = &loadedState{c: c} + if err := c.refreshState(); err != nil { + return nil, err + } + if intelrdt.IsEnabled() { + c.intelRdtManager = l.NewIntelRdtManager(&state.Config, id, state.IntelRdtPath) + } + return c, nil +} + +func (l *LinuxFactory) Type() string { + return "libcontainer" +} + +// StartInitialization loads a container by opening the pipe fd from the parent to read the configuration and state +// This is a low level implementation detail of the reexec and should not be consumed externally +func (l *LinuxFactory) StartInitialization() (err error) { + var ( + pipefd, fifofd int + consoleSocket *os.File + envInitPipe = os.Getenv("_LIBCONTAINER_INITPIPE") + envFifoFd = os.Getenv("_LIBCONTAINER_FIFOFD") + envConsole = os.Getenv("_LIBCONTAINER_CONSOLE") + ) + + // Get the INITPIPE. + pipefd, err = strconv.Atoi(envInitPipe) + if err != nil { + return fmt.Errorf("unable to convert _LIBCONTAINER_INITPIPE=%s to int: %s", envInitPipe, err) + } + + var ( + pipe = os.NewFile(uintptr(pipefd), "pipe") + it = initType(os.Getenv("_LIBCONTAINER_INITTYPE")) + ) + defer pipe.Close() + + // Only init processes have FIFOFD. + fifofd = -1 + if it == initStandard { + if fifofd, err = strconv.Atoi(envFifoFd); err != nil { + return fmt.Errorf("unable to convert _LIBCONTAINER_FIFOFD=%s to int: %s", envFifoFd, err) + } + } + + if envConsole != "" { + console, err := strconv.Atoi(envConsole) + if err != nil { + return fmt.Errorf("unable to convert _LIBCONTAINER_CONSOLE=%s to int: %s", envConsole, err) + } + consoleSocket = os.NewFile(uintptr(console), "console-socket") + defer consoleSocket.Close() + } + + // clear the current process's environment to clean any libcontainer + // specific env vars. + os.Clearenv() + + defer func() { + // We have an error during the initialization of the container's init, + // send it back to the parent process in the form of an initError. + if werr := utils.WriteJSON(pipe, syncT{procError}); werr != nil { + fmt.Fprintln(os.Stderr, err) + return + } + if werr := utils.WriteJSON(pipe, newSystemError(err)); werr != nil { + fmt.Fprintln(os.Stderr, err) + return + } + }() + defer func() { + if e := recover(); e != nil { + err = fmt.Errorf("panic from initialization: %v, %v", e, string(debug.Stack())) + } + }() + + i, err := newContainerInit(it, pipe, consoleSocket, fifofd) + if err != nil { + return err + } + + // If Init succeeds, syscall.Exec will not return, hence none of the defers will be called. + return i.Init() +} + +func (l *LinuxFactory) loadState(root, id string) (*State, error) { + f, err := os.Open(filepath.Join(root, stateFilename)) + if err != nil { + if os.IsNotExist(err) { + return nil, newGenericError(fmt.Errorf("container %q does not exist", id), ContainerNotExists) + } + return nil, newGenericError(err, SystemError) + } + defer f.Close() + var state *State + if err := json.NewDecoder(f).Decode(&state); err != nil { + return nil, newGenericError(err, SystemError) + } + return state, nil +} + +func (l *LinuxFactory) validateID(id string) error { + if !idRegex.MatchString(id) { + return newGenericError(fmt.Errorf("invalid id format: %v", id), InvalidIdFormat) + } + + return nil +} + +// NewuidmapPath returns an option func to configure a LinuxFactory with the +// provided .. +func NewuidmapPath(newuidmapPath string) func(*LinuxFactory) error { + return func(l *LinuxFactory) error { + l.NewuidmapPath = newuidmapPath + return nil + } +} + +// NewgidmapPath returns an option func to configure a LinuxFactory with the +// provided .. +func NewgidmapPath(newgidmapPath string) func(*LinuxFactory) error { + return func(l *LinuxFactory) error { + l.NewgidmapPath = newgidmapPath + return nil + } +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/generic_error.go b/vendor/github.com/opencontainers/runc/libcontainer/generic_error.go new file mode 100644 index 000000000..6e7de2fe7 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/generic_error.go @@ -0,0 +1,92 @@ +package libcontainer + +import ( + "fmt" + "io" + "text/template" + "time" + + "github.com/opencontainers/runc/libcontainer/stacktrace" +) + +var errorTemplate = template.Must(template.New("error").Parse(`Timestamp: {{.Timestamp}} +Code: {{.ECode}} +{{if .Message }} +Message: {{.Message}} +{{end}} +Frames:{{range $i, $frame := .Stack.Frames}} +--- +{{$i}}: {{$frame.Function}} +Package: {{$frame.Package}} +File: {{$frame.File}}@{{$frame.Line}}{{end}} +`)) + +func newGenericError(err error, c ErrorCode) Error { + if le, ok := err.(Error); ok { + return le + } + gerr := &genericError{ + Timestamp: time.Now(), + Err: err, + ECode: c, + Stack: stacktrace.Capture(1), + } + if err != nil { + gerr.Message = err.Error() + } + return gerr +} + +func newSystemError(err error) Error { + return createSystemError(err, "") +} + +func newSystemErrorWithCausef(err error, cause string, v ...interface{}) Error { + return createSystemError(err, fmt.Sprintf(cause, v...)) +} + +func newSystemErrorWithCause(err error, cause string) Error { + return createSystemError(err, cause) +} + +// createSystemError creates the specified error with the correct number of +// stack frames skipped. This is only to be called by the other functions for +// formatting the error. +func createSystemError(err error, cause string) Error { + gerr := &genericError{ + Timestamp: time.Now(), + Err: err, + ECode: SystemError, + Cause: cause, + Stack: stacktrace.Capture(2), + } + if err != nil { + gerr.Message = err.Error() + } + return gerr +} + +type genericError struct { + Timestamp time.Time + ECode ErrorCode + Err error `json:"-"` + Cause string + Message string + Stack stacktrace.Stacktrace +} + +func (e *genericError) Error() string { + if e.Cause == "" { + return e.Message + } + frame := e.Stack.Frames[0] + return fmt.Sprintf("%s:%d: %s caused %q", frame.File, frame.Line, e.Cause, e.Message) +} + +func (e *genericError) Code() ErrorCode { + return e.ECode +} + +func (e *genericError) Detail(w io.Writer) error { + return errorTemplate.Execute(w, e) +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/init_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/init_linux.go new file mode 100644 index 000000000..b19f00ed9 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/init_linux.go @@ -0,0 +1,535 @@ +// +build linux + +package libcontainer + +import ( + "encoding/json" + "fmt" + "io" + "net" + "os" + "strings" + "syscall" // only for Errno + "unsafe" + + "golang.org/x/sys/unix" + + "github.com/containerd/console" + "github.com/opencontainers/runc/libcontainer/cgroups" + "github.com/opencontainers/runc/libcontainer/configs" + "github.com/opencontainers/runc/libcontainer/system" + "github.com/opencontainers/runc/libcontainer/user" + "github.com/opencontainers/runc/libcontainer/utils" + "github.com/pkg/errors" + "github.com/sirupsen/logrus" + "github.com/vishvananda/netlink" +) + +type initType string + +const ( + initSetns initType = "setns" + initStandard initType = "standard" +) + +type pid struct { + Pid int `json:"pid"` + PidFirstChild int `json:"pid_first"` +} + +// network is an internal struct used to setup container networks. +type network struct { + configs.Network + + // TempVethPeerName is a unique temporary veth peer name that was placed into + // the container's namespace. + TempVethPeerName string `json:"temp_veth_peer_name"` +} + +// initConfig is used for transferring parameters from Exec() to Init() +type initConfig struct { + Args []string `json:"args"` + Env []string `json:"env"` + Cwd string `json:"cwd"` + Capabilities *configs.Capabilities `json:"capabilities"` + ProcessLabel string `json:"process_label"` + AppArmorProfile string `json:"apparmor_profile"` + NoNewPrivileges bool `json:"no_new_privileges"` + User string `json:"user"` + AdditionalGroups []string `json:"additional_groups"` + Config *configs.Config `json:"config"` + Networks []*network `json:"network"` + PassedFilesCount int `json:"passed_files_count"` + ContainerId string `json:"containerid"` + Rlimits []configs.Rlimit `json:"rlimits"` + CreateConsole bool `json:"create_console"` + ConsoleWidth uint16 `json:"console_width"` + ConsoleHeight uint16 `json:"console_height"` + Rootless bool `json:"rootless"` +} + +type initer interface { + Init() error +} + +func newContainerInit(t initType, pipe *os.File, consoleSocket *os.File, fifoFd int) (initer, error) { + var config *initConfig + if err := json.NewDecoder(pipe).Decode(&config); err != nil { + return nil, err + } + if err := populateProcessEnvironment(config.Env); err != nil { + return nil, err + } + switch t { + case initSetns: + return &linuxSetnsInit{ + pipe: pipe, + consoleSocket: consoleSocket, + config: config, + }, nil + case initStandard: + return &linuxStandardInit{ + pipe: pipe, + consoleSocket: consoleSocket, + parentPid: unix.Getppid(), + config: config, + fifoFd: fifoFd, + }, nil + } + return nil, fmt.Errorf("unknown init type %q", t) +} + +// populateProcessEnvironment loads the provided environment variables into the +// current processes's environment. +func populateProcessEnvironment(env []string) error { + for _, pair := range env { + p := strings.SplitN(pair, "=", 2) + if len(p) < 2 { + return fmt.Errorf("invalid environment '%v'", pair) + } + if err := os.Setenv(p[0], p[1]); err != nil { + return err + } + } + return nil +} + +// finalizeNamespace drops the caps, sets the correct user +// and working dir, and closes any leaked file descriptors +// before executing the command inside the namespace +func finalizeNamespace(config *initConfig) error { + // Ensure that all unwanted fds we may have accidentally + // inherited are marked close-on-exec so they stay out of the + // container + if err := utils.CloseExecFrom(config.PassedFilesCount + 3); err != nil { + return errors.Wrap(err, "close exec fds") + } + + capabilities := &configs.Capabilities{} + if config.Capabilities != nil { + capabilities = config.Capabilities + } else if config.Config.Capabilities != nil { + capabilities = config.Config.Capabilities + } + w, err := newContainerCapList(capabilities) + if err != nil { + return err + } + // drop capabilities in bounding set before changing user + if err := w.ApplyBoundingSet(); err != nil { + return errors.Wrap(err, "apply bounding set") + } + // preserve existing capabilities while we change users + if err := system.SetKeepCaps(); err != nil { + return errors.Wrap(err, "set keep caps") + } + if err := setupUser(config); err != nil { + return errors.Wrap(err, "setup user") + } + if err := system.ClearKeepCaps(); err != nil { + return errors.Wrap(err, "clear keep caps") + } + if err := w.ApplyCaps(); err != nil { + return errors.Wrap(err, "apply caps") + } + if config.Cwd != "" { + if err := unix.Chdir(config.Cwd); err != nil { + return fmt.Errorf("chdir to cwd (%q) set in config.json failed: %v", config.Cwd, err) + } + } + return nil +} + +// setupConsole sets up the console from inside the container, and sends the +// master pty fd to the config.Pipe (using cmsg). This is done to ensure that +// consoles are scoped to a container properly (see runc#814 and the many +// issues related to that). This has to be run *after* we've pivoted to the new +// rootfs (and the users' configuration is entirely set up). +func setupConsole(socket *os.File, config *initConfig, mount bool) error { + defer socket.Close() + // At this point, /dev/ptmx points to something that we would expect. We + // used to change the owner of the slave path, but since the /dev/pts mount + // can have gid=X set (at the users' option). So touching the owner of the + // slave PTY is not necessary, as the kernel will handle that for us. Note + // however, that setupUser (specifically fixStdioPermissions) *will* change + // the UID owner of the console to be the user the process will run as (so + // they can actually control their console). + + pty, slavePath, err := console.NewPty() + if err != nil { + return err + } + + if config.ConsoleHeight != 0 && config.ConsoleWidth != 0 { + err = pty.Resize(console.WinSize{ + Height: config.ConsoleHeight, + Width: config.ConsoleWidth, + }) + + if err != nil { + return err + } + } + + // After we return from here, we don't need the console anymore. + defer pty.Close() + + // Mount the console inside our rootfs. + if mount { + if err := mountConsole(slavePath); err != nil { + return err + } + } + // While we can access console.master, using the API is a good idea. + if err := utils.SendFd(socket, pty.Name(), pty.Fd()); err != nil { + return err + } + // Now, dup over all the things. + return dupStdio(slavePath) +} + +// syncParentReady sends to the given pipe a JSON payload which indicates that +// the init is ready to Exec the child process. It then waits for the parent to +// indicate that it is cleared to Exec. +func syncParentReady(pipe io.ReadWriter) error { + // Tell parent. + if err := writeSync(pipe, procReady); err != nil { + return err + } + + // Wait for parent to give the all-clear. + if err := readSync(pipe, procRun); err != nil { + return err + } + + return nil +} + +// syncParentHooks sends to the given pipe a JSON payload which indicates that +// the parent should execute pre-start hooks. It then waits for the parent to +// indicate that it is cleared to resume. +func syncParentHooks(pipe io.ReadWriter) error { + // Tell parent. + if err := writeSync(pipe, procHooks); err != nil { + return err + } + + // Wait for parent to give the all-clear. + if err := readSync(pipe, procResume); err != nil { + return err + } + + return nil +} + +// setupUser changes the groups, gid, and uid for the user inside the container +func setupUser(config *initConfig) error { + // Set up defaults. + defaultExecUser := user.ExecUser{ + Uid: 0, + Gid: 0, + Home: "/", + } + + passwdPath, err := user.GetPasswdPath() + if err != nil { + return err + } + + groupPath, err := user.GetGroupPath() + if err != nil { + return err + } + + execUser, err := user.GetExecUserPath(config.User, &defaultExecUser, passwdPath, groupPath) + if err != nil { + return err + } + + var addGroups []int + if len(config.AdditionalGroups) > 0 { + addGroups, err = user.GetAdditionalGroupsPath(config.AdditionalGroups, groupPath) + if err != nil { + return err + } + } + + // Rather than just erroring out later in setuid(2) and setgid(2), check + // that the user is mapped here. + if _, err := config.Config.HostUID(execUser.Uid); err != nil { + return fmt.Errorf("cannot set uid to unmapped user in user namespace") + } + if _, err := config.Config.HostGID(execUser.Gid); err != nil { + return fmt.Errorf("cannot set gid to unmapped user in user namespace") + } + + if config.Rootless { + // We cannot set any additional groups in a rootless container and thus + // we bail if the user asked us to do so. TODO: We currently can't do + // this check earlier, but if libcontainer.Process.User was typesafe + // this might work. + if len(addGroups) > 0 { + return fmt.Errorf("cannot set any additional groups in a rootless container") + } + } + + // Before we change to the container's user make sure that the processes + // STDIO is correctly owned by the user that we are switching to. + if err := fixStdioPermissions(config, execUser); err != nil { + return err + } + + // This isn't allowed in an unprivileged user namespace since Linux 3.19. + // There's nothing we can do about /etc/group entries, so we silently + // ignore setting groups here (since the user didn't explicitly ask us to + // set the group). + if !config.Rootless { + suppGroups := append(execUser.Sgids, addGroups...) + if err := unix.Setgroups(suppGroups); err != nil { + return err + } + } + + if err := system.Setgid(execUser.Gid); err != nil { + return err + } + if err := system.Setuid(execUser.Uid); err != nil { + return err + } + + // if we didn't get HOME already, set it based on the user's HOME + if envHome := os.Getenv("HOME"); envHome == "" { + if err := os.Setenv("HOME", execUser.Home); err != nil { + return err + } + } + return nil +} + +// fixStdioPermissions fixes the permissions of PID 1's STDIO within the container to the specified user. +// The ownership needs to match because it is created outside of the container and needs to be +// localized. +func fixStdioPermissions(config *initConfig, u *user.ExecUser) error { + var null unix.Stat_t + if err := unix.Stat("/dev/null", &null); err != nil { + return err + } + for _, fd := range []uintptr{ + os.Stdin.Fd(), + os.Stderr.Fd(), + os.Stdout.Fd(), + } { + var s unix.Stat_t + if err := unix.Fstat(int(fd), &s); err != nil { + return err + } + + // Skip chown of /dev/null if it was used as one of the STDIO fds. + if s.Rdev == null.Rdev { + continue + } + + // We only change the uid owner (as it is possible for the mount to + // prefer a different gid, and there's no reason for us to change it). + // The reason why we don't just leave the default uid=X mount setup is + // that users expect to be able to actually use their console. Without + // this code, you couldn't effectively run as a non-root user inside a + // container and also have a console set up. + if err := unix.Fchown(int(fd), u.Uid, int(s.Gid)); err != nil { + // If we've hit an EINVAL then s.Gid isn't mapped in the user + // namespace. If we've hit an EPERM then the inode's current owner + // is not mapped in our user namespace (in particular, + // privileged_wrt_inode_uidgid() has failed). In either case, we + // are in a configuration where it's better for us to just not + // touch the stdio rather than bail at this point. + if err == unix.EINVAL || err == unix.EPERM { + continue + } + return err + } + } + return nil +} + +// setupNetwork sets up and initializes any network interface inside the container. +func setupNetwork(config *initConfig) error { + for _, config := range config.Networks { + strategy, err := getStrategy(config.Type) + if err != nil { + return err + } + if err := strategy.initialize(config); err != nil { + return err + } + } + return nil +} + +func setupRoute(config *configs.Config) error { + for _, config := range config.Routes { + _, dst, err := net.ParseCIDR(config.Destination) + if err != nil { + return err + } + src := net.ParseIP(config.Source) + if src == nil { + return fmt.Errorf("Invalid source for route: %s", config.Source) + } + gw := net.ParseIP(config.Gateway) + if gw == nil { + return fmt.Errorf("Invalid gateway for route: %s", config.Gateway) + } + l, err := netlink.LinkByName(config.InterfaceName) + if err != nil { + return err + } + route := &netlink.Route{ + Scope: netlink.SCOPE_UNIVERSE, + Dst: dst, + Src: src, + Gw: gw, + LinkIndex: l.Attrs().Index, + } + if err := netlink.RouteAdd(route); err != nil { + return err + } + } + return nil +} + +func setupRlimits(limits []configs.Rlimit, pid int) error { + for _, rlimit := range limits { + if err := system.Prlimit(pid, rlimit.Type, unix.Rlimit{Max: rlimit.Hard, Cur: rlimit.Soft}); err != nil { + return fmt.Errorf("error setting rlimit type %v: %v", rlimit.Type, err) + } + } + return nil +} + +const _P_PID = 1 + +type siginfo struct { + si_signo int32 + si_errno int32 + si_code int32 + // below here is a union; si_pid is the only field we use + si_pid int32 + // Pad to 128 bytes as detailed in blockUntilWaitable + pad [96]byte +} + +// isWaitable returns true if the process has exited false otherwise. +// Its based off blockUntilWaitable in src/os/wait_waitid.go +func isWaitable(pid int) (bool, error) { + si := &siginfo{} + _, _, e := unix.Syscall6(unix.SYS_WAITID, _P_PID, uintptr(pid), uintptr(unsafe.Pointer(si)), unix.WEXITED|unix.WNOWAIT|unix.WNOHANG, 0, 0) + if e != 0 { + return false, os.NewSyscallError("waitid", e) + } + + return si.si_pid != 0, nil +} + +// isNoChildren returns true if err represents a unix.ECHILD (formerly syscall.ECHILD) false otherwise +func isNoChildren(err error) bool { + switch err := err.(type) { + case syscall.Errno: + if err == unix.ECHILD { + return true + } + case *os.SyscallError: + if err.Err == unix.ECHILD { + return true + } + } + return false +} + +// signalAllProcesses freezes then iterates over all the processes inside the +// manager's cgroups sending the signal s to them. +// If s is SIGKILL then it will wait for each process to exit. +// For all other signals it will check if the process is ready to report its +// exit status and only if it is will a wait be performed. +func signalAllProcesses(m cgroups.Manager, s os.Signal) error { + var procs []*os.Process + if err := m.Freeze(configs.Frozen); err != nil { + logrus.Warn(err) + } + pids, err := m.GetAllPids() + if err != nil { + m.Freeze(configs.Thawed) + return err + } + for _, pid := range pids { + p, err := os.FindProcess(pid) + if err != nil { + logrus.Warn(err) + continue + } + procs = append(procs, p) + if err := p.Signal(s); err != nil { + logrus.Warn(err) + } + } + if err := m.Freeze(configs.Thawed); err != nil { + logrus.Warn(err) + } + + subreaper, err := system.GetSubreaper() + if err != nil { + // The error here means that PR_GET_CHILD_SUBREAPER is not + // supported because this code might run on a kernel older + // than 3.4. We don't want to throw an error in that case, + // and we simplify things, considering there is no subreaper + // set. + subreaper = 0 + } + + for _, p := range procs { + if s != unix.SIGKILL { + if ok, err := isWaitable(p.Pid); err != nil { + if !isNoChildren(err) { + logrus.Warn("signalAllProcesses: ", p.Pid, err) + } + continue + } else if !ok { + // Not ready to report so don't wait + continue + } + } + + // In case a subreaper has been setup, this code must not + // wait for the process. Otherwise, we cannot be sure the + // current process will be reaped by the subreaper, while + // the subreaper might be waiting for this process in order + // to retrieve its exit code. + if subreaper == 0 { + if _, err := p.Wait(); err != nil { + if !isNoChildren(err) { + logrus.Warn("wait: ", err) + } + } + } + } + return nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/intelrdt/intelrdt.go b/vendor/github.com/opencontainers/runc/libcontainer/intelrdt/intelrdt.go new file mode 100644 index 000000000..487c630af --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/intelrdt/intelrdt.go @@ -0,0 +1,553 @@ +// +build linux + +package intelrdt + +import ( + "bufio" + "fmt" + "io/ioutil" + "os" + "path/filepath" + "strconv" + "strings" + "sync" + + "github.com/opencontainers/runc/libcontainer/configs" +) + +/* + * About Intel RDT/CAT feature: + * Intel platforms with new Xeon CPU support Resource Director Technology (RDT). + * Intel Cache Allocation Technology (CAT) is a sub-feature of RDT. Currently L3 + * Cache is the only resource that is supported in RDT. + * + * This feature provides a way for the software to restrict cache allocation to a + * defined 'subset' of L3 cache which may be overlapping with other 'subsets'. + * The different subsets are identified by class of service (CLOS) and each CLOS + * has a capacity bitmask (CBM). + * + * For more information about Intel RDT/CAT can be found in the section 17.17 + * of Intel Software Developer Manual. + * + * About Intel RDT/CAT kernel interface: + * In Linux 4.10 kernel or newer, the interface is defined and exposed via + * "resource control" filesystem, which is a "cgroup-like" interface. + * + * Comparing with cgroups, it has similar process management lifecycle and + * interfaces in a container. But unlike cgroups' hierarchy, it has single level + * filesystem layout. + * + * Intel RDT "resource control" filesystem hierarchy: + * mount -t resctrl resctrl /sys/fs/resctrl + * tree /sys/fs/resctrl + * /sys/fs/resctrl/ + * |-- info + * | |-- L3 + * | |-- cbm_mask + * | |-- min_cbm_bits + * | |-- num_closids + * |-- cpus + * |-- schemata + * |-- tasks + * |-- + * |-- cpus + * |-- schemata + * |-- tasks + * + * For runc, we can make use of `tasks` and `schemata` configuration for L3 cache + * resource constraints. + * + * The file `tasks` has a list of tasks that belongs to this group (e.g., + * " group). Tasks can be added to a group by writing the task ID + * to the "tasks" file (which will automatically remove them from the previous + * group to which they belonged). New tasks created by fork(2) and clone(2) are + * added to the same group as their parent. If a pid is not in any sub group, it is + * in root group. + * + * The file `schemata` has allocation bitmasks/values for L3 cache on each socket, + * which contains L3 cache id and capacity bitmask (CBM). + * Format: "L3:=;=;..." + * For example, on a two-socket machine, L3's schema line could be `L3:0=ff;1=c0` + * which means L3 cache id 0's CBM is 0xff, and L3 cache id 1's CBM is 0xc0. + * + * The valid L3 cache CBM is a *contiguous bits set* and number of bits that can + * be set is less than the max bit. The max bits in the CBM is varied among + * supported Intel Xeon platforms. In Intel RDT "resource control" filesystem + * layout, the CBM in a group should be a subset of the CBM in root. Kernel will + * check if it is valid when writing. e.g., 0xfffff in root indicates the max bits + * of CBM is 20 bits, which mapping to entire L3 cache capacity. Some valid CBM + * values to set in a group: 0xf, 0xf0, 0x3ff, 0x1f00 and etc. + * + * For more information about Intel RDT/CAT kernel interface: + * https://www.kernel.org/doc/Documentation/x86/intel_rdt_ui.txt + * + * An example for runc: + * Consider a two-socket machine with two L3 caches where the default CBM is + * 0xfffff and the max CBM length is 20 bits. With this configuration, tasks + * inside the container only have access to the "upper" 80% of L3 cache id 0 and + * the "lower" 50% L3 cache id 1: + * + * "linux": { + * "intelRdt": { + * "l3CacheSchema": "L3:0=ffff0;1=3ff" + * } + * } + */ + +type Manager interface { + // Applies Intel RDT configuration to the process with the specified pid + Apply(pid int) error + + // Returns statistics for Intel RDT + GetStats() (*Stats, error) + + // Destroys the Intel RDT 'container_id' group + Destroy() error + + // Returns Intel RDT path to save in a state file and to be able to + // restore the object later + GetPath() string + + // Set Intel RDT "resource control" filesystem as configured. + Set(container *configs.Config) error +} + +// This implements interface Manager +type IntelRdtManager struct { + mu sync.Mutex + Config *configs.Config + Id string + Path string +} + +const ( + IntelRdtTasks = "tasks" +) + +var ( + // The absolute root path of the Intel RDT "resource control" filesystem + intelRdtRoot string + intelRdtRootLock sync.Mutex + + // The flag to indicate if Intel RDT is supported + isEnabled bool +) + +type intelRdtData struct { + root string + config *configs.Config + pid int +} + +// Check if Intel RDT is enabled in init() +func init() { + // 1. Check if hardware and kernel support Intel RDT/CAT feature + // "cat_l3" flag is set if supported + isFlagSet, err := parseCpuInfoFile("/proc/cpuinfo") + if !isFlagSet || err != nil { + isEnabled = false + return + } + + // 2. Check if Intel RDT "resource control" filesystem is mounted + // The user guarantees to mount the filesystem + isEnabled = isIntelRdtMounted() +} + +// Return the mount point path of Intel RDT "resource control" filesysem +func findIntelRdtMountpointDir() (string, error) { + f, err := os.Open("/proc/self/mountinfo") + if err != nil { + return "", err + } + defer f.Close() + + s := bufio.NewScanner(f) + for s.Scan() { + text := s.Text() + fields := strings.Split(text, " ") + // Safe as mountinfo encodes mountpoints with spaces as \040. + index := strings.Index(text, " - ") + postSeparatorFields := strings.Fields(text[index+3:]) + numPostFields := len(postSeparatorFields) + + // This is an error as we can't detect if the mount is for "Intel RDT" + if numPostFields == 0 { + return "", fmt.Errorf("Found no fields post '-' in %q", text) + } + + if postSeparatorFields[0] == "resctrl" { + // Check that the mount is properly formated. + if numPostFields < 3 { + return "", fmt.Errorf("Error found less than 3 fields post '-' in %q", text) + } + + return fields[4], nil + } + } + if err := s.Err(); err != nil { + return "", err + } + + return "", NewNotFoundError("Intel RDT") +} + +// Gets the root path of Intel RDT "resource control" filesystem +func getIntelRdtRoot() (string, error) { + intelRdtRootLock.Lock() + defer intelRdtRootLock.Unlock() + + if intelRdtRoot != "" { + return intelRdtRoot, nil + } + + root, err := findIntelRdtMountpointDir() + if err != nil { + return "", err + } + + if _, err := os.Stat(root); err != nil { + return "", err + } + + intelRdtRoot = root + return intelRdtRoot, nil +} + +func isIntelRdtMounted() bool { + _, err := getIntelRdtRoot() + if err != nil { + return false + } + + return true +} + +func parseCpuInfoFile(path string) (bool, error) { + f, err := os.Open(path) + if err != nil { + return false, err + } + defer f.Close() + + s := bufio.NewScanner(f) + for s.Scan() { + if err := s.Err(); err != nil { + return false, err + } + + text := s.Text() + flags := strings.Split(text, " ") + + // "cat_l3" flag is set if Intel RDT/CAT is supported + for _, flag := range flags { + if flag == "cat_l3" { + return true, nil + } + } + } + return false, nil +} + +func parseUint(s string, base, bitSize int) (uint64, error) { + value, err := strconv.ParseUint(s, base, bitSize) + if err != nil { + intValue, intErr := strconv.ParseInt(s, base, bitSize) + // 1. Handle negative values greater than MinInt64 (and) + // 2. Handle negative values lesser than MinInt64 + if intErr == nil && intValue < 0 { + return 0, nil + } else if intErr != nil && intErr.(*strconv.NumError).Err == strconv.ErrRange && intValue < 0 { + return 0, nil + } + + return value, err + } + + return value, nil +} + +// Gets a single uint64 value from the specified file. +func getIntelRdtParamUint(path, file string) (uint64, error) { + fileName := filepath.Join(path, file) + contents, err := ioutil.ReadFile(fileName) + if err != nil { + return 0, err + } + + res, err := parseUint(strings.TrimSpace(string(contents)), 10, 64) + if err != nil { + return res, fmt.Errorf("unable to parse %q as a uint from file %q", string(contents), fileName) + } + return res, nil +} + +// Gets a string value from the specified file +func getIntelRdtParamString(path, file string) (string, error) { + contents, err := ioutil.ReadFile(filepath.Join(path, file)) + if err != nil { + return "", err + } + + return strings.TrimSpace(string(contents)), nil +} + +func readTasksFile(dir string) ([]int, error) { + f, err := os.Open(filepath.Join(dir, IntelRdtTasks)) + if err != nil { + return nil, err + } + defer f.Close() + + var ( + s = bufio.NewScanner(f) + out = []int{} + ) + + for s.Scan() { + if t := s.Text(); t != "" { + pid, err := strconv.Atoi(t) + if err != nil { + return nil, err + } + out = append(out, pid) + } + } + return out, nil +} + +func writeFile(dir, file, data string) error { + if dir == "" { + return fmt.Errorf("no such directory for %s", file) + } + if err := ioutil.WriteFile(filepath.Join(dir, file), []byte(data+"\n"), 0700); err != nil { + return fmt.Errorf("failed to write %v to %v: %v", data, file, err) + } + return nil +} + +func getIntelRdtData(c *configs.Config, pid int) (*intelRdtData, error) { + rootPath, err := getIntelRdtRoot() + if err != nil { + return nil, err + } + return &intelRdtData{ + root: rootPath, + config: c, + pid: pid, + }, nil +} + +// Get the read-only L3 cache information +func getL3CacheInfo() (*L3CacheInfo, error) { + l3CacheInfo := &L3CacheInfo{} + + rootPath, err := getIntelRdtRoot() + if err != nil { + return l3CacheInfo, err + } + + path := filepath.Join(rootPath, "info", "L3") + cbmMask, err := getIntelRdtParamString(path, "cbm_mask") + if err != nil { + return l3CacheInfo, err + } + minCbmBits, err := getIntelRdtParamUint(path, "min_cbm_bits") + if err != nil { + return l3CacheInfo, err + } + numClosids, err := getIntelRdtParamUint(path, "num_closids") + if err != nil { + return l3CacheInfo, err + } + + l3CacheInfo.CbmMask = cbmMask + l3CacheInfo.MinCbmBits = minCbmBits + l3CacheInfo.NumClosids = numClosids + + return l3CacheInfo, nil +} + +// WriteIntelRdtTasks writes the specified pid into the "tasks" file +func WriteIntelRdtTasks(dir string, pid int) error { + if dir == "" { + return fmt.Errorf("no such directory for %s", IntelRdtTasks) + } + + // Dont attach any pid if -1 is specified as a pid + if pid != -1 { + if err := ioutil.WriteFile(filepath.Join(dir, IntelRdtTasks), []byte(strconv.Itoa(pid)), 0700); err != nil { + return fmt.Errorf("failed to write %v to %v: %v", pid, IntelRdtTasks, err) + } + } + return nil +} + +// Check if Intel RDT is enabled +func IsEnabled() bool { + return isEnabled +} + +// Get the 'container_id' path in Intel RDT "resource control" filesystem +func GetIntelRdtPath(id string) (string, error) { + rootPath, err := getIntelRdtRoot() + if err != nil { + return "", err + } + + path := filepath.Join(rootPath, id) + return path, nil +} + +// Applies Intel RDT configuration to the process with the specified pid +func (m *IntelRdtManager) Apply(pid int) (err error) { + // If intelRdt is not specified in config, we do nothing + if m.Config.IntelRdt == nil { + return nil + } + d, err := getIntelRdtData(m.Config, pid) + if err != nil && !IsNotFound(err) { + return err + } + + m.mu.Lock() + defer m.mu.Unlock() + path, err := d.join(m.Id) + if err != nil { + return err + } + + m.Path = path + return nil +} + +// Destroys the Intel RDT 'container_id' group +func (m *IntelRdtManager) Destroy() error { + m.mu.Lock() + defer m.mu.Unlock() + if err := os.RemoveAll(m.Path); err != nil { + return err + } + m.Path = "" + return nil +} + +// Returns Intel RDT path to save in a state file and to be able to +// restore the object later +func (m *IntelRdtManager) GetPath() string { + if m.Path == "" { + m.Path, _ = GetIntelRdtPath(m.Id) + } + return m.Path +} + +// Returns statistics for Intel RDT +func (m *IntelRdtManager) GetStats() (*Stats, error) { + // If intelRdt is not specified in config + if m.Config.IntelRdt == nil { + return nil, nil + } + + m.mu.Lock() + defer m.mu.Unlock() + stats := NewStats() + + // The read-only L3 cache information + l3CacheInfo, err := getL3CacheInfo() + if err != nil { + return nil, err + } + stats.L3CacheInfo = l3CacheInfo + + // The read-only L3 cache schema in root + rootPath, err := getIntelRdtRoot() + if err != nil { + return nil, err + } + tmpRootStrings, err := getIntelRdtParamString(rootPath, "schemata") + if err != nil { + return nil, err + } + // L3 cache schema is in the first line + schemaRootStrings := strings.Split(tmpRootStrings, "\n") + stats.L3CacheSchemaRoot = schemaRootStrings[0] + + // The L3 cache schema in 'container_id' group + tmpStrings, err := getIntelRdtParamString(m.GetPath(), "schemata") + if err != nil { + return nil, err + } + // L3 cache schema is in the first line + schemaStrings := strings.Split(tmpStrings, "\n") + stats.L3CacheSchema = schemaStrings[0] + + return stats, nil +} + +// Set Intel RDT "resource control" filesystem as configured. +func (m *IntelRdtManager) Set(container *configs.Config) error { + path := m.GetPath() + + // About L3 cache schema file: + // The schema has allocation masks/values for L3 cache on each socket, + // which contains L3 cache id and capacity bitmask (CBM). + // Format: "L3:=;=;..." + // For example, on a two-socket machine, L3's schema line could be: + // L3:0=ff;1=c0 + // Which means L3 cache id 0's CBM is 0xff, and L3 cache id 1's CBM is 0xc0. + // + // About L3 cache CBM validity: + // The valid L3 cache CBM is a *contiguous bits set* and number of + // bits that can be set is less than the max bit. The max bits in the + // CBM is varied among supported Intel Xeon platforms. In Intel RDT + // "resource control" filesystem layout, the CBM in a group should + // be a subset of the CBM in root. Kernel will check if it is valid + // when writing. + // e.g., 0xfffff in root indicates the max bits of CBM is 20 bits, + // which mapping to entire L3 cache capacity. Some valid CBM values + // to set in a group: 0xf, 0xf0, 0x3ff, 0x1f00 and etc. + if container.IntelRdt != nil { + l3CacheSchema := container.IntelRdt.L3CacheSchema + if l3CacheSchema != "" { + if err := writeFile(path, "schemata", l3CacheSchema); err != nil { + return err + } + } + } + + return nil +} + +func (raw *intelRdtData) join(id string) (string, error) { + path := filepath.Join(raw.root, id) + if err := os.MkdirAll(path, 0755); err != nil { + return "", err + } + + if err := WriteIntelRdtTasks(path, raw.pid); err != nil { + return "", err + } + return path, nil +} + +type NotFoundError struct { + ResourceControl string +} + +func (e *NotFoundError) Error() string { + return fmt.Sprintf("mountpoint for %s not found", e.ResourceControl) +} + +func NewNotFoundError(res string) error { + return &NotFoundError{ + ResourceControl: res, + } +} + +func IsNotFound(err error) bool { + if err == nil { + return false + } + _, ok := err.(*NotFoundError) + return ok +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/intelrdt/stats.go b/vendor/github.com/opencontainers/runc/libcontainer/intelrdt/stats.go new file mode 100644 index 000000000..095c0a380 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/intelrdt/stats.go @@ -0,0 +1,24 @@ +// +build linux + +package intelrdt + +type L3CacheInfo struct { + CbmMask string `json:"cbm_mask,omitempty"` + MinCbmBits uint64 `json:"min_cbm_bits,omitempty"` + NumClosids uint64 `json:"num_closids,omitempty"` +} + +type Stats struct { + // The read-only L3 cache information + L3CacheInfo *L3CacheInfo `json:"l3_cache_info,omitempty"` + + // The read-only L3 cache schema in root + L3CacheSchemaRoot string `json:"l3_cache_schema_root,omitempty"` + + // The L3 cache schema in 'container_id' group + L3CacheSchema string `json:"l3_cache_schema,omitempty"` +} + +func NewStats() *Stats { + return &Stats{} +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/keys/keyctl.go b/vendor/github.com/opencontainers/runc/libcontainer/keys/keyctl.go new file mode 100644 index 000000000..ce8b4e6b0 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/keys/keyctl.go @@ -0,0 +1,50 @@ +// +build linux + +package keys + +import ( + "fmt" + "strconv" + "strings" + + "golang.org/x/sys/unix" +) + +type KeySerial uint32 + +func JoinSessionKeyring(name string) (KeySerial, error) { + sessKeyId, err := unix.KeyctlJoinSessionKeyring(name) + if err != nil { + return 0, fmt.Errorf("could not create session key: %v", err) + } + return KeySerial(sessKeyId), nil +} + +// ModKeyringPerm modifies permissions on a keyring by reading the current permissions, +// anding the bits with the given mask (clearing permissions) and setting +// additional permission bits +func ModKeyringPerm(ringId KeySerial, mask, setbits uint32) error { + dest, err := unix.KeyctlString(unix.KEYCTL_DESCRIBE, int(ringId)) + if err != nil { + return err + } + + res := strings.Split(dest, ";") + if len(res) < 5 { + return fmt.Errorf("Destination buffer for key description is too small") + } + + // parse permissions + perm64, err := strconv.ParseUint(res[3], 16, 32) + if err != nil { + return err + } + + perm := (uint32(perm64) & mask) | setbits + + if err := unix.KeyctlSetperm(int(ringId), perm); err != nil { + return err + } + + return nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/message_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/message_linux.go new file mode 100644 index 000000000..ed7f986df --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/message_linux.go @@ -0,0 +1,89 @@ +// +build linux + +package libcontainer + +import ( + "github.com/vishvananda/netlink/nl" + "golang.org/x/sys/unix" +) + +// list of known message types we want to send to bootstrap program +// The number is randomly chosen to not conflict with known netlink types +const ( + InitMsg uint16 = 62000 + CloneFlagsAttr uint16 = 27281 + NsPathsAttr uint16 = 27282 + UidmapAttr uint16 = 27283 + GidmapAttr uint16 = 27284 + SetgroupAttr uint16 = 27285 + OomScoreAdjAttr uint16 = 27286 + RootlessAttr uint16 = 27287 + UidmapPathAttr uint16 = 27288 + GidmapPathAttr uint16 = 27289 +) + +type Int32msg struct { + Type uint16 + Value uint32 +} + +// Serialize serializes the message. +// Int32msg has the following representation +// | nlattr len | nlattr type | +// | uint32 value | +func (msg *Int32msg) Serialize() []byte { + buf := make([]byte, msg.Len()) + native := nl.NativeEndian() + native.PutUint16(buf[0:2], uint16(msg.Len())) + native.PutUint16(buf[2:4], msg.Type) + native.PutUint32(buf[4:8], msg.Value) + return buf +} + +func (msg *Int32msg) Len() int { + return unix.NLA_HDRLEN + 4 +} + +// Bytemsg has the following representation +// | nlattr len | nlattr type | +// | value | pad | +type Bytemsg struct { + Type uint16 + Value []byte +} + +func (msg *Bytemsg) Serialize() []byte { + l := msg.Len() + buf := make([]byte, (l+unix.NLA_ALIGNTO-1) & ^(unix.NLA_ALIGNTO-1)) + native := nl.NativeEndian() + native.PutUint16(buf[0:2], uint16(l)) + native.PutUint16(buf[2:4], msg.Type) + copy(buf[4:], msg.Value) + return buf +} + +func (msg *Bytemsg) Len() int { + return unix.NLA_HDRLEN + len(msg.Value) + 1 // null-terminated +} + +type Boolmsg struct { + Type uint16 + Value bool +} + +func (msg *Boolmsg) Serialize() []byte { + buf := make([]byte, msg.Len()) + native := nl.NativeEndian() + native.PutUint16(buf[0:2], uint16(msg.Len())) + native.PutUint16(buf[2:4], msg.Type) + if msg.Value { + native.PutUint32(buf[4:8], uint32(1)) + } else { + native.PutUint32(buf[4:8], uint32(0)) + } + return buf +} + +func (msg *Boolmsg) Len() int { + return unix.NLA_HDRLEN + 4 // alignment +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/mount/mount.go b/vendor/github.com/opencontainers/runc/libcontainer/mount/mount.go new file mode 100644 index 000000000..e8965e081 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/mount/mount.go @@ -0,0 +1,23 @@ +package mount + +// GetMounts retrieves a list of mounts for the current running process. +func GetMounts() ([]*Info, error) { + return parseMountTable() +} + +// Mounted looks at /proc/self/mountinfo to determine of the specified +// mountpoint has been mounted +func Mounted(mountpoint string) (bool, error) { + entries, err := parseMountTable() + if err != nil { + return false, err + } + + // Search the table for the mountpoint + for _, e := range entries { + if e.Mountpoint == mountpoint { + return true, nil + } + } + return false, nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/mount/mount_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/mount/mount_linux.go new file mode 100644 index 000000000..1e5191928 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/mount/mount_linux.go @@ -0,0 +1,82 @@ +// +build linux + +package mount + +import ( + "bufio" + "fmt" + "io" + "os" + "strings" +) + +const ( + /* 36 35 98:0 /mnt1 /mnt2 rw,noatime master:1 - ext3 /dev/root rw,errors=continue + (1)(2)(3) (4) (5) (6) (7) (8) (9) (10) (11) + + (1) mount ID: unique identifier of the mount (may be reused after umount) + (2) parent ID: ID of parent (or of self for the top of the mount tree) + (3) major:minor: value of st_dev for files on filesystem + (4) root: root of the mount within the filesystem + (5) mount point: mount point relative to the process's root + (6) mount options: per mount options + (7) optional fields: zero or more fields of the form "tag[:value]" + (8) separator: marks the end of the optional fields + (9) filesystem type: name of filesystem of the form "type[.subtype]" + (10) mount source: filesystem specific information or "none" + (11) super options: per super block options*/ + mountinfoFormat = "%d %d %d:%d %s %s %s %s" +) + +// Parse /proc/self/mountinfo because comparing Dev and ino does not work from +// bind mounts +func parseMountTable() ([]*Info, error) { + f, err := os.Open("/proc/self/mountinfo") + if err != nil { + return nil, err + } + defer f.Close() + + return parseInfoFile(f) +} + +func parseInfoFile(r io.Reader) ([]*Info, error) { + var ( + s = bufio.NewScanner(r) + out = []*Info{} + ) + + for s.Scan() { + if err := s.Err(); err != nil { + return nil, err + } + + var ( + p = &Info{} + text = s.Text() + optionalFields string + ) + + if _, err := fmt.Sscanf(text, mountinfoFormat, + &p.ID, &p.Parent, &p.Major, &p.Minor, + &p.Root, &p.Mountpoint, &p.Opts, &optionalFields); err != nil { + return nil, fmt.Errorf("Scanning '%s' failed: %s", text, err) + } + // Safe as mountinfo encodes mountpoints with spaces as \040. + index := strings.Index(text, " - ") + postSeparatorFields := strings.Fields(text[index+3:]) + if len(postSeparatorFields) < 3 { + return nil, fmt.Errorf("Error found less than 3 fields post '-' in %q", text) + } + + if optionalFields != "-" { + p.Optional = optionalFields + } + + p.Fstype = postSeparatorFields[0] + p.Source = postSeparatorFields[1] + p.VfsOpts = strings.Join(postSeparatorFields[2:], " ") + out = append(out, p) + } + return out, nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/mount/mountinfo.go b/vendor/github.com/opencontainers/runc/libcontainer/mount/mountinfo.go new file mode 100644 index 000000000..e3fc3535e --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/mount/mountinfo.go @@ -0,0 +1,40 @@ +package mount + +// Info reveals information about a particular mounted filesystem. This +// struct is populated from the content in the /proc//mountinfo file. +type Info struct { + // ID is a unique identifier of the mount (may be reused after umount). + ID int + + // Parent indicates the ID of the mount parent (or of self for the top of the + // mount tree). + Parent int + + // Major indicates one half of the device ID which identifies the device class. + Major int + + // Minor indicates one half of the device ID which identifies a specific + // instance of device. + Minor int + + // Root of the mount within the filesystem. + Root string + + // Mountpoint indicates the mount point relative to the process's root. + Mountpoint string + + // Opts represents mount-specific options. + Opts string + + // Optional represents optional fields. + Optional string + + // Fstype indicates the type of filesystem, such as EXT3. + Fstype string + + // Source indicates filesystem specific information or "none". + Source string + + // VfsOpts represents per super block options. + VfsOpts string +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/network_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/network_linux.go new file mode 100644 index 000000000..5075bee4d --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/network_linux.go @@ -0,0 +1,259 @@ +// +build linux + +package libcontainer + +import ( + "fmt" + "io/ioutil" + "net" + "path/filepath" + "strconv" + "strings" + + "github.com/opencontainers/runc/libcontainer/configs" + "github.com/opencontainers/runc/libcontainer/utils" + "github.com/vishvananda/netlink" +) + +var strategies = map[string]networkStrategy{ + "veth": &veth{}, + "loopback": &loopback{}, +} + +// networkStrategy represents a specific network configuration for +// a container's networking stack +type networkStrategy interface { + create(*network, int) error + initialize(*network) error + detach(*configs.Network) error + attach(*configs.Network) error +} + +// getStrategy returns the specific network strategy for the +// provided type. +func getStrategy(tpe string) (networkStrategy, error) { + s, exists := strategies[tpe] + if !exists { + return nil, fmt.Errorf("unknown strategy type %q", tpe) + } + return s, nil +} + +// Returns the network statistics for the network interfaces represented by the NetworkRuntimeInfo. +func getNetworkInterfaceStats(interfaceName string) (*NetworkInterface, error) { + out := &NetworkInterface{Name: interfaceName} + // This can happen if the network runtime information is missing - possible if the + // container was created by an old version of libcontainer. + if interfaceName == "" { + return out, nil + } + type netStatsPair struct { + // Where to write the output. + Out *uint64 + // The network stats file to read. + File string + } + // Ingress for host veth is from the container. Hence tx_bytes stat on the host veth is actually number of bytes received by the container. + netStats := []netStatsPair{ + {Out: &out.RxBytes, File: "tx_bytes"}, + {Out: &out.RxPackets, File: "tx_packets"}, + {Out: &out.RxErrors, File: "tx_errors"}, + {Out: &out.RxDropped, File: "tx_dropped"}, + + {Out: &out.TxBytes, File: "rx_bytes"}, + {Out: &out.TxPackets, File: "rx_packets"}, + {Out: &out.TxErrors, File: "rx_errors"}, + {Out: &out.TxDropped, File: "rx_dropped"}, + } + for _, netStat := range netStats { + data, err := readSysfsNetworkStats(interfaceName, netStat.File) + if err != nil { + return nil, err + } + *(netStat.Out) = data + } + return out, nil +} + +// Reads the specified statistics available under /sys/class/net//statistics +func readSysfsNetworkStats(ethInterface, statsFile string) (uint64, error) { + data, err := ioutil.ReadFile(filepath.Join("/sys/class/net", ethInterface, "statistics", statsFile)) + if err != nil { + return 0, err + } + return strconv.ParseUint(strings.TrimSpace(string(data)), 10, 64) +} + +// loopback is a network strategy that provides a basic loopback device +type loopback struct { +} + +func (l *loopback) create(n *network, nspid int) error { + return nil +} + +func (l *loopback) initialize(config *network) error { + return netlink.LinkSetUp(&netlink.Device{LinkAttrs: netlink.LinkAttrs{Name: "lo"}}) +} + +func (l *loopback) attach(n *configs.Network) (err error) { + return nil +} + +func (l *loopback) detach(n *configs.Network) (err error) { + return nil +} + +// veth is a network strategy that uses a bridge and creates +// a veth pair, one that is attached to the bridge on the host and the other +// is placed inside the container's namespace +type veth struct { +} + +func (v *veth) detach(n *configs.Network) (err error) { + return netlink.LinkSetMaster(&netlink.Device{LinkAttrs: netlink.LinkAttrs{Name: n.HostInterfaceName}}, nil) +} + +// attach a container network interface to an external network +func (v *veth) attach(n *configs.Network) (err error) { + brl, err := netlink.LinkByName(n.Bridge) + if err != nil { + return err + } + br, ok := brl.(*netlink.Bridge) + if !ok { + return fmt.Errorf("Wrong device type %T", brl) + } + host, err := netlink.LinkByName(n.HostInterfaceName) + if err != nil { + return err + } + + if err := netlink.LinkSetMaster(host, br); err != nil { + return err + } + if err := netlink.LinkSetMTU(host, n.Mtu); err != nil { + return err + } + if n.HairpinMode { + if err := netlink.LinkSetHairpin(host, true); err != nil { + return err + } + } + if err := netlink.LinkSetUp(host); err != nil { + return err + } + + return nil +} + +func (v *veth) create(n *network, nspid int) (err error) { + tmpName, err := v.generateTempPeerName() + if err != nil { + return err + } + n.TempVethPeerName = tmpName + if n.Bridge == "" { + return fmt.Errorf("bridge is not specified") + } + veth := &netlink.Veth{ + LinkAttrs: netlink.LinkAttrs{ + Name: n.HostInterfaceName, + TxQLen: n.TxQueueLen, + }, + PeerName: n.TempVethPeerName, + } + if err := netlink.LinkAdd(veth); err != nil { + return err + } + defer func() { + if err != nil { + netlink.LinkDel(veth) + } + }() + if err := v.attach(&n.Network); err != nil { + return err + } + child, err := netlink.LinkByName(n.TempVethPeerName) + if err != nil { + return err + } + return netlink.LinkSetNsPid(child, nspid) +} + +func (v *veth) generateTempPeerName() (string, error) { + return utils.GenerateRandomName("veth", 7) +} + +func (v *veth) initialize(config *network) error { + peer := config.TempVethPeerName + if peer == "" { + return fmt.Errorf("peer is not specified") + } + child, err := netlink.LinkByName(peer) + if err != nil { + return err + } + if err := netlink.LinkSetDown(child); err != nil { + return err + } + if err := netlink.LinkSetName(child, config.Name); err != nil { + return err + } + // get the interface again after we changed the name as the index also changes. + if child, err = netlink.LinkByName(config.Name); err != nil { + return err + } + if config.MacAddress != "" { + mac, err := net.ParseMAC(config.MacAddress) + if err != nil { + return err + } + if err := netlink.LinkSetHardwareAddr(child, mac); err != nil { + return err + } + } + ip, err := netlink.ParseAddr(config.Address) + if err != nil { + return err + } + if err := netlink.AddrAdd(child, ip); err != nil { + return err + } + if config.IPv6Address != "" { + ip6, err := netlink.ParseAddr(config.IPv6Address) + if err != nil { + return err + } + if err := netlink.AddrAdd(child, ip6); err != nil { + return err + } + } + if err := netlink.LinkSetMTU(child, config.Mtu); err != nil { + return err + } + if err := netlink.LinkSetUp(child); err != nil { + return err + } + if config.Gateway != "" { + gw := net.ParseIP(config.Gateway) + if err := netlink.RouteAdd(&netlink.Route{ + Scope: netlink.SCOPE_UNIVERSE, + LinkIndex: child.Attrs().Index, + Gw: gw, + }); err != nil { + return err + } + } + if config.IPv6Gateway != "" { + gw := net.ParseIP(config.IPv6Gateway) + if err := netlink.RouteAdd(&netlink.Route{ + Scope: netlink.SCOPE_UNIVERSE, + LinkIndex: child.Attrs().Index, + Gw: gw, + }); err != nil { + return err + } + } + return nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/notify_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/notify_linux.go new file mode 100644 index 000000000..47a06783d --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/notify_linux.go @@ -0,0 +1,90 @@ +// +build linux + +package libcontainer + +import ( + "fmt" + "io/ioutil" + "os" + "path/filepath" + + "golang.org/x/sys/unix" +) + +const oomCgroupName = "memory" + +type PressureLevel uint + +const ( + LowPressure PressureLevel = iota + MediumPressure + CriticalPressure +) + +func registerMemoryEvent(cgDir string, evName string, arg string) (<-chan struct{}, error) { + evFile, err := os.Open(filepath.Join(cgDir, evName)) + if err != nil { + return nil, err + } + fd, err := unix.Eventfd(0, unix.EFD_CLOEXEC) + if err != nil { + evFile.Close() + return nil, err + } + + eventfd := os.NewFile(uintptr(fd), "eventfd") + + eventControlPath := filepath.Join(cgDir, "cgroup.event_control") + data := fmt.Sprintf("%d %d %s", eventfd.Fd(), evFile.Fd(), arg) + if err := ioutil.WriteFile(eventControlPath, []byte(data), 0700); err != nil { + eventfd.Close() + evFile.Close() + return nil, err + } + ch := make(chan struct{}) + go func() { + defer func() { + eventfd.Close() + evFile.Close() + close(ch) + }() + buf := make([]byte, 8) + for { + if _, err := eventfd.Read(buf); err != nil { + return + } + // When a cgroup is destroyed, an event is sent to eventfd. + // So if the control path is gone, return instead of notifying. + if _, err := os.Lstat(eventControlPath); os.IsNotExist(err) { + return + } + ch <- struct{}{} + } + }() + return ch, nil +} + +// notifyOnOOM returns channel on which you can expect event about OOM, +// if process died without OOM this channel will be closed. +func notifyOnOOM(paths map[string]string) (<-chan struct{}, error) { + dir := paths[oomCgroupName] + if dir == "" { + return nil, fmt.Errorf("path %q missing", oomCgroupName) + } + + return registerMemoryEvent(dir, "memory.oom_control", "") +} + +func notifyMemoryPressure(paths map[string]string, level PressureLevel) (<-chan struct{}, error) { + dir := paths[oomCgroupName] + if dir == "" { + return nil, fmt.Errorf("path %q missing", oomCgroupName) + } + + if level > CriticalPressure { + return nil, fmt.Errorf("invalid pressure level %d", level) + } + + levelStr := []string{"low", "medium", "critical"}[level] + return registerMemoryEvent(dir, "memory.pressure_level", levelStr) +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/process.go b/vendor/github.com/opencontainers/runc/libcontainer/process.go new file mode 100644 index 000000000..9a7c60141 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/process.go @@ -0,0 +1,113 @@ +package libcontainer + +import ( + "fmt" + "io" + "math" + "os" + + "github.com/opencontainers/runc/libcontainer/configs" +) + +type processOperations interface { + wait() (*os.ProcessState, error) + signal(sig os.Signal) error + pid() int +} + +// Process specifies the configuration and IO for a process inside +// a container. +type Process struct { + // The command to be run followed by any arguments. + Args []string + + // Env specifies the environment variables for the process. + Env []string + + // User will set the uid and gid of the executing process running inside the container + // local to the container's user and group configuration. + User string + + // AdditionalGroups specifies the gids that should be added to supplementary groups + // in addition to those that the user belongs to. + AdditionalGroups []string + + // Cwd will change the processes current working directory inside the container's rootfs. + Cwd string + + // Stdin is a pointer to a reader which provides the standard input stream. + Stdin io.Reader + + // Stdout is a pointer to a writer which receives the standard output stream. + Stdout io.Writer + + // Stderr is a pointer to a writer which receives the standard error stream. + Stderr io.Writer + + // ExtraFiles specifies additional open files to be inherited by the container + ExtraFiles []*os.File + + // Initial sizings for the console + ConsoleWidth uint16 + ConsoleHeight uint16 + + // Capabilities specify the capabilities to keep when executing the process inside the container + // All capabilities not specified will be dropped from the processes capability mask + Capabilities *configs.Capabilities + + // AppArmorProfile specifies the profile to apply to the process and is + // changed at the time the process is execed + AppArmorProfile string + + // Label specifies the label to apply to the process. It is commonly used by selinux + Label string + + // NoNewPrivileges controls whether processes can gain additional privileges. + NoNewPrivileges *bool + + // Rlimits specifies the resource limits, such as max open files, to set in the container + // If Rlimits are not set, the container will inherit rlimits from the parent process + Rlimits []configs.Rlimit + + // ConsoleSocket provides the masterfd console. + ConsoleSocket *os.File + + // Init specifies whether the process is the first process in the container. + Init bool + + ops processOperations +} + +// Wait waits for the process to exit. +// Wait releases any resources associated with the Process +func (p Process) Wait() (*os.ProcessState, error) { + if p.ops == nil { + return nil, newGenericError(fmt.Errorf("invalid process"), NoProcessOps) + } + return p.ops.wait() +} + +// Pid returns the process ID +func (p Process) Pid() (int, error) { + // math.MinInt32 is returned here, because it's invalid value + // for the kill() system call. + if p.ops == nil { + return math.MinInt32, newGenericError(fmt.Errorf("invalid process"), NoProcessOps) + } + return p.ops.pid(), nil +} + +// Signal sends a signal to the Process. +func (p Process) Signal(sig os.Signal) error { + if p.ops == nil { + return newGenericError(fmt.Errorf("invalid process"), NoProcessOps) + } + return p.ops.signal(sig) +} + +// IO holds the process's STDIO +type IO struct { + Stdin io.WriteCloser + Stdout io.ReadCloser + Stderr io.ReadCloser +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/process_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/process_linux.go new file mode 100644 index 000000000..58980b059 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/process_linux.go @@ -0,0 +1,547 @@ +// +build linux + +package libcontainer + +import ( + "encoding/json" + "errors" + "fmt" + "io" + "os" + "os/exec" + "path/filepath" + "strconv" + "syscall" // only for Signal + + "github.com/opencontainers/runc/libcontainer/cgroups" + "github.com/opencontainers/runc/libcontainer/configs" + "github.com/opencontainers/runc/libcontainer/intelrdt" + "github.com/opencontainers/runc/libcontainer/system" + "github.com/opencontainers/runc/libcontainer/utils" + + "golang.org/x/sys/unix" +) + +type parentProcess interface { + // pid returns the pid for the running process. + pid() int + + // start starts the process execution. + start() error + + // send a SIGKILL to the process and wait for the exit. + terminate() error + + // wait waits on the process returning the process state. + wait() (*os.ProcessState, error) + + // startTime returns the process start time. + startTime() (uint64, error) + + signal(os.Signal) error + + externalDescriptors() []string + + setExternalDescriptors(fds []string) +} + +type setnsProcess struct { + cmd *exec.Cmd + parentPipe *os.File + childPipe *os.File + cgroupPaths map[string]string + intelRdtPath string + config *initConfig + fds []string + process *Process + bootstrapData io.Reader +} + +func (p *setnsProcess) startTime() (uint64, error) { + stat, err := system.Stat(p.pid()) + return stat.StartTime, err +} + +func (p *setnsProcess) signal(sig os.Signal) error { + s, ok := sig.(syscall.Signal) + if !ok { + return errors.New("os: unsupported signal type") + } + return unix.Kill(p.pid(), s) +} + +func (p *setnsProcess) start() (err error) { + defer p.parentPipe.Close() + err = p.cmd.Start() + p.childPipe.Close() + if err != nil { + return newSystemErrorWithCause(err, "starting setns process") + } + if p.bootstrapData != nil { + if _, err := io.Copy(p.parentPipe, p.bootstrapData); err != nil { + return newSystemErrorWithCause(err, "copying bootstrap data to pipe") + } + } + if err = p.execSetns(); err != nil { + return newSystemErrorWithCause(err, "executing setns process") + } + if len(p.cgroupPaths) > 0 { + if err := cgroups.EnterPid(p.cgroupPaths, p.pid()); err != nil { + return newSystemErrorWithCausef(err, "adding pid %d to cgroups", p.pid()) + } + } + if p.intelRdtPath != "" { + // if Intel RDT "resource control" filesystem path exists + _, err := os.Stat(p.intelRdtPath) + if err == nil { + if err := intelrdt.WriteIntelRdtTasks(p.intelRdtPath, p.pid()); err != nil { + return newSystemErrorWithCausef(err, "adding pid %d to Intel RDT resource control filesystem", p.pid()) + } + } + } + // set rlimits, this has to be done here because we lose permissions + // to raise the limits once we enter a user-namespace + if err := setupRlimits(p.config.Rlimits, p.pid()); err != nil { + return newSystemErrorWithCause(err, "setting rlimits for process") + } + if err := utils.WriteJSON(p.parentPipe, p.config); err != nil { + return newSystemErrorWithCause(err, "writing config to pipe") + } + + ierr := parseSync(p.parentPipe, func(sync *syncT) error { + switch sync.Type { + case procReady: + // This shouldn't happen. + panic("unexpected procReady in setns") + case procHooks: + // This shouldn't happen. + panic("unexpected procHooks in setns") + default: + return newSystemError(fmt.Errorf("invalid JSON payload from child")) + } + }) + + if err := unix.Shutdown(int(p.parentPipe.Fd()), unix.SHUT_WR); err != nil { + return newSystemErrorWithCause(err, "calling shutdown on init pipe") + } + // Must be done after Shutdown so the child will exit and we can wait for it. + if ierr != nil { + p.wait() + return ierr + } + return nil +} + +// execSetns runs the process that executes C code to perform the setns calls +// because setns support requires the C process to fork off a child and perform the setns +// before the go runtime boots, we wait on the process to die and receive the child's pid +// over the provided pipe. +func (p *setnsProcess) execSetns() error { + status, err := p.cmd.Process.Wait() + if err != nil { + p.cmd.Wait() + return newSystemErrorWithCause(err, "waiting on setns process to finish") + } + if !status.Success() { + p.cmd.Wait() + return newSystemError(&exec.ExitError{ProcessState: status}) + } + var pid *pid + if err := json.NewDecoder(p.parentPipe).Decode(&pid); err != nil { + p.cmd.Wait() + return newSystemErrorWithCause(err, "reading pid from init pipe") + } + + // Clean up the zombie parent process + firstChildProcess, err := os.FindProcess(pid.PidFirstChild) + if err != nil { + return err + } + + // Ignore the error in case the child has already been reaped for any reason + _, _ = firstChildProcess.Wait() + + process, err := os.FindProcess(pid.Pid) + if err != nil { + return err + } + p.cmd.Process = process + p.process.ops = p + return nil +} + +// terminate sends a SIGKILL to the forked process for the setns routine then waits to +// avoid the process becoming a zombie. +func (p *setnsProcess) terminate() error { + if p.cmd.Process == nil { + return nil + } + err := p.cmd.Process.Kill() + if _, werr := p.wait(); err == nil { + err = werr + } + return err +} + +func (p *setnsProcess) wait() (*os.ProcessState, error) { + err := p.cmd.Wait() + + // Return actual ProcessState even on Wait error + return p.cmd.ProcessState, err +} + +func (p *setnsProcess) pid() int { + return p.cmd.Process.Pid +} + +func (p *setnsProcess) externalDescriptors() []string { + return p.fds +} + +func (p *setnsProcess) setExternalDescriptors(newFds []string) { + p.fds = newFds +} + +type initProcess struct { + cmd *exec.Cmd + parentPipe *os.File + childPipe *os.File + config *initConfig + manager cgroups.Manager + intelRdtManager intelrdt.Manager + container *linuxContainer + fds []string + process *Process + bootstrapData io.Reader + sharePidns bool +} + +func (p *initProcess) pid() int { + return p.cmd.Process.Pid +} + +func (p *initProcess) externalDescriptors() []string { + return p.fds +} + +// execSetns runs the process that executes C code to perform the setns calls +// because setns support requires the C process to fork off a child and perform the setns +// before the go runtime boots, we wait on the process to die and receive the child's pid +// over the provided pipe. +// This is called by initProcess.start function +func (p *initProcess) execSetns() error { + status, err := p.cmd.Process.Wait() + if err != nil { + p.cmd.Wait() + return err + } + if !status.Success() { + p.cmd.Wait() + return &exec.ExitError{ProcessState: status} + } + var pid *pid + if err := json.NewDecoder(p.parentPipe).Decode(&pid); err != nil { + p.cmd.Wait() + return err + } + + // Clean up the zombie parent process + firstChildProcess, err := os.FindProcess(pid.PidFirstChild) + if err != nil { + return err + } + + // Ignore the error in case the child has already been reaped for any reason + _, _ = firstChildProcess.Wait() + + process, err := os.FindProcess(pid.Pid) + if err != nil { + return err + } + p.cmd.Process = process + p.process.ops = p + return nil +} + +func (p *initProcess) start() error { + defer p.parentPipe.Close() + err := p.cmd.Start() + p.process.ops = p + p.childPipe.Close() + if err != nil { + p.process.ops = nil + return newSystemErrorWithCause(err, "starting init process command") + } + // Do this before syncing with child so that no children can escape the + // cgroup. We don't need to worry about not doing this and not being root + // because we'd be using the rootless cgroup manager in that case. + if err := p.manager.Apply(p.pid()); err != nil { + return newSystemErrorWithCause(err, "applying cgroup configuration for process") + } + if p.intelRdtManager != nil { + if err := p.intelRdtManager.Apply(p.pid()); err != nil { + return newSystemErrorWithCause(err, "applying Intel RDT configuration for process") + } + } + defer func() { + if err != nil { + // TODO: should not be the responsibility to call here + p.manager.Destroy() + if p.intelRdtManager != nil { + p.intelRdtManager.Destroy() + } + } + }() + + if _, err := io.Copy(p.parentPipe, p.bootstrapData); err != nil { + return newSystemErrorWithCause(err, "copying bootstrap data to pipe") + } + + if err := p.execSetns(); err != nil { + return newSystemErrorWithCause(err, "running exec setns process for init") + } + + // Save the standard descriptor names before the container process + // can potentially move them (e.g., via dup2()). If we don't do this now, + // we won't know at checkpoint time which file descriptor to look up. + fds, err := getPipeFds(p.pid()) + if err != nil { + return newSystemErrorWithCausef(err, "getting pipe fds for pid %d", p.pid()) + } + p.setExternalDescriptors(fds) + if err := p.createNetworkInterfaces(); err != nil { + return newSystemErrorWithCause(err, "creating network interfaces") + } + if err := p.sendConfig(); err != nil { + return newSystemErrorWithCause(err, "sending config to init process") + } + var ( + sentRun bool + sentResume bool + ) + + ierr := parseSync(p.parentPipe, func(sync *syncT) error { + switch sync.Type { + case procReady: + // set rlimits, this has to be done here because we lose permissions + // to raise the limits once we enter a user-namespace + if err := setupRlimits(p.config.Rlimits, p.pid()); err != nil { + return newSystemErrorWithCause(err, "setting rlimits for ready process") + } + // call prestart hooks + if !p.config.Config.Namespaces.Contains(configs.NEWNS) { + // Setup cgroup before prestart hook, so that the prestart hook could apply cgroup permissions. + if err := p.manager.Set(p.config.Config); err != nil { + return newSystemErrorWithCause(err, "setting cgroup config for ready process") + } + if p.intelRdtManager != nil { + if err := p.intelRdtManager.Set(p.config.Config); err != nil { + return newSystemErrorWithCause(err, "setting Intel RDT config for ready process") + } + } + + if p.config.Config.Hooks != nil { + bundle, annotations := utils.Annotations(p.container.config.Labels) + s := configs.HookState{ + Version: p.container.config.Version, + ID: p.container.id, + Pid: p.pid(), + Bundle: bundle, + Annotations: annotations, + } + for i, hook := range p.config.Config.Hooks.Prestart { + if err := hook.Run(s); err != nil { + return newSystemErrorWithCausef(err, "running prestart hook %d", i) + } + } + } + } + // Sync with child. + if err := writeSync(p.parentPipe, procRun); err != nil { + return newSystemErrorWithCause(err, "writing syncT 'run'") + } + sentRun = true + case procHooks: + // Setup cgroup before prestart hook, so that the prestart hook could apply cgroup permissions. + if err := p.manager.Set(p.config.Config); err != nil { + return newSystemErrorWithCause(err, "setting cgroup config for procHooks process") + } + if p.intelRdtManager != nil { + if err := p.intelRdtManager.Set(p.config.Config); err != nil { + return newSystemErrorWithCause(err, "setting Intel RDT config for procHooks process") + } + } + if p.config.Config.Hooks != nil { + bundle, annotations := utils.Annotations(p.container.config.Labels) + s := configs.HookState{ + Version: p.container.config.Version, + ID: p.container.id, + Pid: p.pid(), + Bundle: bundle, + Annotations: annotations, + } + for i, hook := range p.config.Config.Hooks.Prestart { + if err := hook.Run(s); err != nil { + return newSystemErrorWithCausef(err, "running prestart hook %d", i) + } + } + } + // Sync with child. + if err := writeSync(p.parentPipe, procResume); err != nil { + return newSystemErrorWithCause(err, "writing syncT 'resume'") + } + sentResume = true + default: + return newSystemError(fmt.Errorf("invalid JSON payload from child")) + } + + return nil + }) + + if !sentRun { + return newSystemErrorWithCause(ierr, "container init") + } + if p.config.Config.Namespaces.Contains(configs.NEWNS) && !sentResume { + return newSystemError(fmt.Errorf("could not synchronise after executing prestart hooks with container process")) + } + if err := unix.Shutdown(int(p.parentPipe.Fd()), unix.SHUT_WR); err != nil { + return newSystemErrorWithCause(err, "shutting down init pipe") + } + + // Must be done after Shutdown so the child will exit and we can wait for it. + if ierr != nil { + p.wait() + return ierr + } + return nil +} + +func (p *initProcess) wait() (*os.ProcessState, error) { + err := p.cmd.Wait() + if err != nil { + return p.cmd.ProcessState, err + } + // we should kill all processes in cgroup when init is died if we use host PID namespace + if p.sharePidns { + signalAllProcesses(p.manager, unix.SIGKILL) + } + return p.cmd.ProcessState, nil +} + +func (p *initProcess) terminate() error { + if p.cmd.Process == nil { + return nil + } + err := p.cmd.Process.Kill() + if _, werr := p.wait(); err == nil { + err = werr + } + return err +} + +func (p *initProcess) startTime() (uint64, error) { + stat, err := system.Stat(p.pid()) + return stat.StartTime, err +} + +func (p *initProcess) sendConfig() error { + // send the config to the container's init process, we don't use JSON Encode + // here because there might be a problem in JSON decoder in some cases, see: + // https://github.com/docker/docker/issues/14203#issuecomment-174177790 + return utils.WriteJSON(p.parentPipe, p.config) +} + +func (p *initProcess) createNetworkInterfaces() error { + for _, config := range p.config.Config.Networks { + strategy, err := getStrategy(config.Type) + if err != nil { + return err + } + n := &network{ + Network: *config, + } + if err := strategy.create(n, p.pid()); err != nil { + return err + } + p.config.Networks = append(p.config.Networks, n) + } + return nil +} + +func (p *initProcess) signal(sig os.Signal) error { + s, ok := sig.(syscall.Signal) + if !ok { + return errors.New("os: unsupported signal type") + } + return unix.Kill(p.pid(), s) +} + +func (p *initProcess) setExternalDescriptors(newFds []string) { + p.fds = newFds +} + +func getPipeFds(pid int) ([]string, error) { + fds := make([]string, 3) + + dirPath := filepath.Join("/proc", strconv.Itoa(pid), "/fd") + for i := 0; i < 3; i++ { + // XXX: This breaks if the path is not a valid symlink (which can + // happen in certain particularly unlucky mount namespace setups). + f := filepath.Join(dirPath, strconv.Itoa(i)) + target, err := os.Readlink(f) + if err != nil { + // Ignore permission errors, for rootless containers and other + // non-dumpable processes. if we can't get the fd for a particular + // file, there's not much we can do. + if os.IsPermission(err) { + continue + } + return fds, err + } + fds[i] = target + } + return fds, nil +} + +// InitializeIO creates pipes for use with the process's stdio and returns the +// opposite side for each. Do not use this if you want to have a pseudoterminal +// set up for you by libcontainer (TODO: fix that too). +// TODO: This is mostly unnecessary, and should be handled by clients. +func (p *Process) InitializeIO(rootuid, rootgid int) (i *IO, err error) { + var fds []uintptr + i = &IO{} + // cleanup in case of an error + defer func() { + if err != nil { + for _, fd := range fds { + unix.Close(int(fd)) + } + } + }() + // STDIN + r, w, err := os.Pipe() + if err != nil { + return nil, err + } + fds = append(fds, r.Fd(), w.Fd()) + p.Stdin, i.Stdin = r, w + // STDOUT + if r, w, err = os.Pipe(); err != nil { + return nil, err + } + fds = append(fds, r.Fd(), w.Fd()) + p.Stdout, i.Stdout = w, r + // STDERR + if r, w, err = os.Pipe(); err != nil { + return nil, err + } + fds = append(fds, r.Fd(), w.Fd()) + p.Stderr, i.Stderr = w, r + // change ownership of the pipes incase we are in a user namespace + for _, fd := range fds { + if err := unix.Fchown(int(fd), rootuid, rootgid); err != nil { + return nil, err + } + } + return i, nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/restored_process.go b/vendor/github.com/opencontainers/runc/libcontainer/restored_process.go new file mode 100644 index 000000000..408916ad9 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/restored_process.go @@ -0,0 +1,122 @@ +// +build linux + +package libcontainer + +import ( + "fmt" + "os" + + "github.com/opencontainers/runc/libcontainer/system" +) + +func newRestoredProcess(pid int, fds []string) (*restoredProcess, error) { + var ( + err error + ) + proc, err := os.FindProcess(pid) + if err != nil { + return nil, err + } + stat, err := system.Stat(pid) + if err != nil { + return nil, err + } + return &restoredProcess{ + proc: proc, + processStartTime: stat.StartTime, + fds: fds, + }, nil +} + +type restoredProcess struct { + proc *os.Process + processStartTime uint64 + fds []string +} + +func (p *restoredProcess) start() error { + return newGenericError(fmt.Errorf("restored process cannot be started"), SystemError) +} + +func (p *restoredProcess) pid() int { + return p.proc.Pid +} + +func (p *restoredProcess) terminate() error { + err := p.proc.Kill() + if _, werr := p.wait(); err == nil { + err = werr + } + return err +} + +func (p *restoredProcess) wait() (*os.ProcessState, error) { + // TODO: how do we wait on the actual process? + // maybe use --exec-cmd in criu + st, err := p.proc.Wait() + if err != nil { + return nil, err + } + return st, nil +} + +func (p *restoredProcess) startTime() (uint64, error) { + return p.processStartTime, nil +} + +func (p *restoredProcess) signal(s os.Signal) error { + return p.proc.Signal(s) +} + +func (p *restoredProcess) externalDescriptors() []string { + return p.fds +} + +func (p *restoredProcess) setExternalDescriptors(newFds []string) { + p.fds = newFds +} + +// nonChildProcess represents a process where the calling process is not +// the parent process. This process is created when a factory loads a container from +// a persisted state. +type nonChildProcess struct { + processPid int + processStartTime uint64 + fds []string +} + +func (p *nonChildProcess) start() error { + return newGenericError(fmt.Errorf("restored process cannot be started"), SystemError) +} + +func (p *nonChildProcess) pid() int { + return p.processPid +} + +func (p *nonChildProcess) terminate() error { + return newGenericError(fmt.Errorf("restored process cannot be terminated"), SystemError) +} + +func (p *nonChildProcess) wait() (*os.ProcessState, error) { + return nil, newGenericError(fmt.Errorf("restored process cannot be waited on"), SystemError) +} + +func (p *nonChildProcess) startTime() (uint64, error) { + return p.processStartTime, nil +} + +func (p *nonChildProcess) signal(s os.Signal) error { + proc, err := os.FindProcess(p.processPid) + if err != nil { + return err + } + return proc.Signal(s) +} + +func (p *nonChildProcess) externalDescriptors() []string { + return p.fds +} + +func (p *nonChildProcess) setExternalDescriptors(newFds []string) { + p.fds = newFds +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/rootfs_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/rootfs_linux.go new file mode 100644 index 000000000..cf715d664 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/rootfs_linux.go @@ -0,0 +1,838 @@ +// +build linux + +package libcontainer + +import ( + "fmt" + "io" + "io/ioutil" + "os" + "os/exec" + "path" + "path/filepath" + "strings" + "time" + + "github.com/cyphar/filepath-securejoin" + "github.com/mrunalp/fileutils" + "github.com/opencontainers/runc/libcontainer/cgroups" + "github.com/opencontainers/runc/libcontainer/configs" + "github.com/opencontainers/runc/libcontainer/mount" + "github.com/opencontainers/runc/libcontainer/system" + libcontainerUtils "github.com/opencontainers/runc/libcontainer/utils" + "github.com/opencontainers/selinux/go-selinux/label" + + "golang.org/x/sys/unix" +) + +const defaultMountFlags = unix.MS_NOEXEC | unix.MS_NOSUID | unix.MS_NODEV + +// needsSetupDev returns true if /dev needs to be set up. +func needsSetupDev(config *configs.Config) bool { + for _, m := range config.Mounts { + if m.Device == "bind" && libcontainerUtils.CleanPath(m.Destination) == "/dev" { + return false + } + } + return true +} + +// prepareRootfs sets up the devices, mount points, and filesystems for use +// inside a new mount namespace. It doesn't set anything as ro. You must call +// finalizeRootfs after this function to finish setting up the rootfs. +func prepareRootfs(pipe io.ReadWriter, iConfig *initConfig) (err error) { + config := iConfig.Config + if err := prepareRoot(config); err != nil { + return newSystemErrorWithCause(err, "preparing rootfs") + } + + setupDev := needsSetupDev(config) + for _, m := range config.Mounts { + for _, precmd := range m.PremountCmds { + if err := mountCmd(precmd); err != nil { + return newSystemErrorWithCause(err, "running premount command") + } + } + + if err := mountToRootfs(m, config.Rootfs, config.MountLabel); err != nil { + return newSystemErrorWithCausef(err, "mounting %q to rootfs %q at %q", m.Source, config.Rootfs, m.Destination) + } + + for _, postcmd := range m.PostmountCmds { + if err := mountCmd(postcmd); err != nil { + return newSystemErrorWithCause(err, "running postmount command") + } + } + } + + if setupDev { + if err := createDevices(config); err != nil { + return newSystemErrorWithCause(err, "creating device nodes") + } + if err := setupPtmx(config); err != nil { + return newSystemErrorWithCause(err, "setting up ptmx") + } + if err := setupDevSymlinks(config.Rootfs); err != nil { + return newSystemErrorWithCause(err, "setting up /dev symlinks") + } + } + + // Signal the parent to run the pre-start hooks. + // The hooks are run after the mounts are setup, but before we switch to the new + // root, so that the old root is still available in the hooks for any mount + // manipulations. + // Note that iConfig.Cwd is not guaranteed to exist here. + if err := syncParentHooks(pipe); err != nil { + return err + } + + // The reason these operations are done here rather than in finalizeRootfs + // is because the console-handling code gets quite sticky if we have to set + // up the console before doing the pivot_root(2). This is because the + // Console API has to also work with the ExecIn case, which means that the + // API must be able to deal with being inside as well as outside the + // container. It's just cleaner to do this here (at the expense of the + // operation not being perfectly split). + + if err := unix.Chdir(config.Rootfs); err != nil { + return newSystemErrorWithCausef(err, "changing dir to %q", config.Rootfs) + } + + if config.NoPivotRoot { + err = msMoveRoot(config.Rootfs) + } else if config.Namespaces.Contains(configs.NEWNS) { + err = pivotRoot(config.Rootfs) + } else { + err = chroot(config.Rootfs) + } + if err != nil { + return newSystemErrorWithCause(err, "jailing process inside rootfs") + } + + if setupDev { + if err := reOpenDevNull(); err != nil { + return newSystemErrorWithCause(err, "reopening /dev/null inside container") + } + } + + if cwd := iConfig.Cwd; cwd != "" { + // Note that spec.Process.Cwd can contain unclean value like "../../../../foo/bar...". + // However, we are safe to call MkDirAll directly because we are in the jail here. + if err := os.MkdirAll(cwd, 0755); err != nil { + return err + } + } + + return nil +} + +// finalizeRootfs sets anything to ro if necessary. You must call +// prepareRootfs first. +func finalizeRootfs(config *configs.Config) (err error) { + // remount dev as ro if specified + for _, m := range config.Mounts { + if libcontainerUtils.CleanPath(m.Destination) == "/dev" { + if m.Flags&unix.MS_RDONLY == unix.MS_RDONLY { + if err := remountReadonly(m); err != nil { + return newSystemErrorWithCausef(err, "remounting %q as readonly", m.Destination) + } + } + break + } + } + + // set rootfs ( / ) as readonly + if config.Readonlyfs { + if err := setReadonly(); err != nil { + return newSystemErrorWithCause(err, "setting rootfs as readonly") + } + } + + unix.Umask(0022) + return nil +} + +func mountCmd(cmd configs.Command) error { + command := exec.Command(cmd.Path, cmd.Args[:]...) + command.Env = cmd.Env + command.Dir = cmd.Dir + if out, err := command.CombinedOutput(); err != nil { + return fmt.Errorf("%#v failed: %s: %v", cmd, string(out), err) + } + return nil +} + +func mountToRootfs(m *configs.Mount, rootfs, mountLabel string) error { + var ( + dest = m.Destination + ) + if !strings.HasPrefix(dest, rootfs) { + dest = filepath.Join(rootfs, dest) + } + + switch m.Device { + case "proc", "sysfs": + if err := os.MkdirAll(dest, 0755); err != nil { + return err + } + // Selinux kernels do not support labeling of /proc or /sys + return mountPropagate(m, rootfs, "") + case "mqueue": + if err := os.MkdirAll(dest, 0755); err != nil { + return err + } + if err := mountPropagate(m, rootfs, mountLabel); err != nil { + // older kernels do not support labeling of /dev/mqueue + if err := mountPropagate(m, rootfs, ""); err != nil { + return err + } + return label.SetFileLabel(dest, mountLabel) + } + return nil + case "tmpfs": + copyUp := m.Extensions&configs.EXT_COPYUP == configs.EXT_COPYUP + tmpDir := "" + stat, err := os.Stat(dest) + if err != nil { + if err := os.MkdirAll(dest, 0755); err != nil { + return err + } + } + if copyUp { + tmpDir, err = ioutil.TempDir("/tmp", "runctmpdir") + if err != nil { + return newSystemErrorWithCause(err, "tmpcopyup: failed to create tmpdir") + } + defer os.RemoveAll(tmpDir) + m.Destination = tmpDir + } + if err := mountPropagate(m, rootfs, mountLabel); err != nil { + return err + } + if copyUp { + if err := fileutils.CopyDirectory(dest, tmpDir); err != nil { + errMsg := fmt.Errorf("tmpcopyup: failed to copy %s to %s: %v", dest, tmpDir, err) + if err1 := unix.Unmount(tmpDir, unix.MNT_DETACH); err1 != nil { + return newSystemErrorWithCausef(err1, "tmpcopyup: %v: failed to unmount", errMsg) + } + return errMsg + } + if err := unix.Mount(tmpDir, dest, "", unix.MS_MOVE, ""); err != nil { + errMsg := fmt.Errorf("tmpcopyup: failed to move mount %s to %s: %v", tmpDir, dest, err) + if err1 := unix.Unmount(tmpDir, unix.MNT_DETACH); err1 != nil { + return newSystemErrorWithCausef(err1, "tmpcopyup: %v: failed to unmount", errMsg) + } + return errMsg + } + } + if stat != nil { + if err = os.Chmod(dest, stat.Mode()); err != nil { + return err + } + } + return nil + case "bind": + stat, err := os.Stat(m.Source) + if err != nil { + // error out if the source of a bind mount does not exist as we will be + // unable to bind anything to it. + return err + } + // ensure that the destination of the bind mount is resolved of symlinks at mount time because + // any previous mounts can invalidate the next mount's destination. + // this can happen when a user specifies mounts within other mounts to cause breakouts or other + // evil stuff to try to escape the container's rootfs. + if dest, err = securejoin.SecureJoin(rootfs, m.Destination); err != nil { + return err + } + if err := checkMountDestination(rootfs, dest); err != nil { + return err + } + // update the mount with the correct dest after symlinks are resolved. + m.Destination = dest + if err := createIfNotExists(dest, stat.IsDir()); err != nil { + return err + } + if err := mountPropagate(m, rootfs, mountLabel); err != nil { + return err + } + // bind mount won't change mount options, we need remount to make mount options effective. + // first check that we have non-default options required before attempting a remount + if m.Flags&^(unix.MS_REC|unix.MS_REMOUNT|unix.MS_BIND) != 0 { + // only remount if unique mount options are set + if err := remount(m, rootfs); err != nil { + return err + } + } + + if m.Relabel != "" { + if err := label.Validate(m.Relabel); err != nil { + return err + } + shared := label.IsShared(m.Relabel) + if err := label.Relabel(m.Source, mountLabel, shared); err != nil { + return err + } + } + case "cgroup": + binds, err := getCgroupMounts(m) + if err != nil { + return err + } + var merged []string + for _, b := range binds { + ss := filepath.Base(b.Destination) + if strings.Contains(ss, ",") { + merged = append(merged, ss) + } + } + tmpfs := &configs.Mount{ + Source: "tmpfs", + Device: "tmpfs", + Destination: m.Destination, + Flags: defaultMountFlags, + Data: "mode=755", + PropagationFlags: m.PropagationFlags, + } + if err := mountToRootfs(tmpfs, rootfs, mountLabel); err != nil { + return err + } + for _, b := range binds { + if err := mountToRootfs(b, rootfs, mountLabel); err != nil { + return err + } + } + for _, mc := range merged { + for _, ss := range strings.Split(mc, ",") { + // symlink(2) is very dumb, it will just shove the path into + // the link and doesn't do any checks or relative path + // conversion. Also, don't error out if the cgroup already exists. + if err := os.Symlink(mc, filepath.Join(rootfs, m.Destination, ss)); err != nil && !os.IsExist(err) { + return err + } + } + } + if m.Flags&unix.MS_RDONLY != 0 { + // remount cgroup root as readonly + mcgrouproot := &configs.Mount{ + Source: m.Destination, + Device: "bind", + Destination: m.Destination, + Flags: defaultMountFlags | unix.MS_RDONLY | unix.MS_BIND, + } + if err := remount(mcgrouproot, rootfs); err != nil { + return err + } + } + default: + // ensure that the destination of the mount is resolved of symlinks at mount time because + // any previous mounts can invalidate the next mount's destination. + // this can happen when a user specifies mounts within other mounts to cause breakouts or other + // evil stuff to try to escape the container's rootfs. + var err error + if dest, err = securejoin.SecureJoin(rootfs, m.Destination); err != nil { + return err + } + if err := checkMountDestination(rootfs, dest); err != nil { + return err + } + // update the mount with the correct dest after symlinks are resolved. + m.Destination = dest + if err := os.MkdirAll(dest, 0755); err != nil { + return err + } + return mountPropagate(m, rootfs, mountLabel) + } + return nil +} + +func getCgroupMounts(m *configs.Mount) ([]*configs.Mount, error) { + mounts, err := cgroups.GetCgroupMounts(false) + if err != nil { + return nil, err + } + + cgroupPaths, err := cgroups.ParseCgroupFile("/proc/self/cgroup") + if err != nil { + return nil, err + } + + var binds []*configs.Mount + + for _, mm := range mounts { + dir, err := mm.GetOwnCgroup(cgroupPaths) + if err != nil { + return nil, err + } + relDir, err := filepath.Rel(mm.Root, dir) + if err != nil { + return nil, err + } + binds = append(binds, &configs.Mount{ + Device: "bind", + Source: filepath.Join(mm.Mountpoint, relDir), + Destination: filepath.Join(m.Destination, filepath.Base(mm.Mountpoint)), + Flags: unix.MS_BIND | unix.MS_REC | m.Flags, + PropagationFlags: m.PropagationFlags, + }) + } + + return binds, nil +} + +// checkMountDestination checks to ensure that the mount destination is not over the top of /proc. +// dest is required to be an abs path and have any symlinks resolved before calling this function. +func checkMountDestination(rootfs, dest string) error { + invalidDestinations := []string{ + "/proc", + } + // White list, it should be sub directories of invalid destinations + validDestinations := []string{ + // These entries can be bind mounted by files emulated by fuse, + // so commands like top, free displays stats in container. + "/proc/cpuinfo", + "/proc/diskstats", + "/proc/meminfo", + "/proc/stat", + "/proc/swaps", + "/proc/uptime", + "/proc/net/dev", + } + for _, valid := range validDestinations { + path, err := filepath.Rel(filepath.Join(rootfs, valid), dest) + if err != nil { + return err + } + if path == "." { + return nil + } + } + for _, invalid := range invalidDestinations { + path, err := filepath.Rel(filepath.Join(rootfs, invalid), dest) + if err != nil { + return err + } + if path == "." || !strings.HasPrefix(path, "..") { + return fmt.Errorf("%q cannot be mounted because it is located inside %q", dest, invalid) + } + } + return nil +} + +func setupDevSymlinks(rootfs string) error { + var links = [][2]string{ + {"/proc/self/fd", "/dev/fd"}, + {"/proc/self/fd/0", "/dev/stdin"}, + {"/proc/self/fd/1", "/dev/stdout"}, + {"/proc/self/fd/2", "/dev/stderr"}, + } + // kcore support can be toggled with CONFIG_PROC_KCORE; only create a symlink + // in /dev if it exists in /proc. + if _, err := os.Stat("/proc/kcore"); err == nil { + links = append(links, [2]string{"/proc/kcore", "/dev/core"}) + } + for _, link := range links { + var ( + src = link[0] + dst = filepath.Join(rootfs, link[1]) + ) + if err := os.Symlink(src, dst); err != nil && !os.IsExist(err) { + return fmt.Errorf("symlink %s %s %s", src, dst, err) + } + } + return nil +} + +// If stdin, stdout, and/or stderr are pointing to `/dev/null` in the parent's rootfs +// this method will make them point to `/dev/null` in this container's rootfs. This +// needs to be called after we chroot/pivot into the container's rootfs so that any +// symlinks are resolved locally. +func reOpenDevNull() error { + var stat, devNullStat unix.Stat_t + file, err := os.OpenFile("/dev/null", os.O_RDWR, 0) + if err != nil { + return fmt.Errorf("Failed to open /dev/null - %s", err) + } + defer file.Close() + if err := unix.Fstat(int(file.Fd()), &devNullStat); err != nil { + return err + } + for fd := 0; fd < 3; fd++ { + if err := unix.Fstat(fd, &stat); err != nil { + return err + } + if stat.Rdev == devNullStat.Rdev { + // Close and re-open the fd. + if err := unix.Dup3(int(file.Fd()), fd, 0); err != nil { + return err + } + } + } + return nil +} + +// Create the device nodes in the container. +func createDevices(config *configs.Config) error { + useBindMount := system.RunningInUserNS() || config.Namespaces.Contains(configs.NEWUSER) + oldMask := unix.Umask(0000) + for _, node := range config.Devices { + // containers running in a user namespace are not allowed to mknod + // devices so we can just bind mount it from the host. + if err := createDeviceNode(config.Rootfs, node, useBindMount); err != nil { + unix.Umask(oldMask) + return err + } + } + unix.Umask(oldMask) + return nil +} + +func bindMountDeviceNode(dest string, node *configs.Device) error { + f, err := os.Create(dest) + if err != nil && !os.IsExist(err) { + return err + } + if f != nil { + f.Close() + } + return unix.Mount(node.Path, dest, "bind", unix.MS_BIND, "") +} + +// Creates the device node in the rootfs of the container. +func createDeviceNode(rootfs string, node *configs.Device, bind bool) error { + dest := filepath.Join(rootfs, node.Path) + if err := os.MkdirAll(filepath.Dir(dest), 0755); err != nil { + return err + } + + if bind { + return bindMountDeviceNode(dest, node) + } + if err := mknodDevice(dest, node); err != nil { + if os.IsExist(err) { + return nil + } else if os.IsPermission(err) { + return bindMountDeviceNode(dest, node) + } + return err + } + return nil +} + +func mknodDevice(dest string, node *configs.Device) error { + fileMode := node.FileMode + switch node.Type { + case 'c', 'u': + fileMode |= unix.S_IFCHR + case 'b': + fileMode |= unix.S_IFBLK + case 'p': + fileMode |= unix.S_IFIFO + default: + return fmt.Errorf("%c is not a valid device type for device %s", node.Type, node.Path) + } + if err := unix.Mknod(dest, uint32(fileMode), node.Mkdev()); err != nil { + return err + } + return unix.Chown(dest, int(node.Uid), int(node.Gid)) +} + +func getMountInfo(mountinfo []*mount.Info, dir string) *mount.Info { + for _, m := range mountinfo { + if m.Mountpoint == dir { + return m + } + } + return nil +} + +// Get the parent mount point of directory passed in as argument. Also return +// optional fields. +func getParentMount(rootfs string) (string, string, error) { + var path string + + mountinfos, err := mount.GetMounts() + if err != nil { + return "", "", err + } + + mountinfo := getMountInfo(mountinfos, rootfs) + if mountinfo != nil { + return rootfs, mountinfo.Optional, nil + } + + path = rootfs + for { + path = filepath.Dir(path) + + mountinfo = getMountInfo(mountinfos, path) + if mountinfo != nil { + return path, mountinfo.Optional, nil + } + + if path == "/" { + break + } + } + + // If we are here, we did not find parent mount. Something is wrong. + return "", "", fmt.Errorf("Could not find parent mount of %s", rootfs) +} + +// Make parent mount private if it was shared +func rootfsParentMountPrivate(rootfs string) error { + sharedMount := false + + parentMount, optionalOpts, err := getParentMount(rootfs) + if err != nil { + return err + } + + optsSplit := strings.Split(optionalOpts, " ") + for _, opt := range optsSplit { + if strings.HasPrefix(opt, "shared:") { + sharedMount = true + break + } + } + + // Make parent mount PRIVATE if it was shared. It is needed for two + // reasons. First of all pivot_root() will fail if parent mount is + // shared. Secondly when we bind mount rootfs it will propagate to + // parent namespace and we don't want that to happen. + if sharedMount { + return unix.Mount("", parentMount, "", unix.MS_PRIVATE, "") + } + + return nil +} + +func prepareRoot(config *configs.Config) error { + flag := unix.MS_SLAVE | unix.MS_REC + if config.RootPropagation != 0 { + flag = config.RootPropagation + } + if err := unix.Mount("", "/", "", uintptr(flag), ""); err != nil { + return err + } + + // Make parent mount private to make sure following bind mount does + // not propagate in other namespaces. Also it will help with kernel + // check pass in pivot_root. (IS_SHARED(new_mnt->mnt_parent)) + if err := rootfsParentMountPrivate(config.Rootfs); err != nil { + return err + } + + return unix.Mount(config.Rootfs, config.Rootfs, "bind", unix.MS_BIND|unix.MS_REC, "") +} + +func setReadonly() error { + return unix.Mount("/", "/", "bind", unix.MS_BIND|unix.MS_REMOUNT|unix.MS_RDONLY|unix.MS_REC, "") +} + +func setupPtmx(config *configs.Config) error { + ptmx := filepath.Join(config.Rootfs, "dev/ptmx") + if err := os.Remove(ptmx); err != nil && !os.IsNotExist(err) { + return err + } + if err := os.Symlink("pts/ptmx", ptmx); err != nil { + return fmt.Errorf("symlink dev ptmx %s", err) + } + return nil +} + +// pivotRoot will call pivot_root such that rootfs becomes the new root +// filesystem, and everything else is cleaned up. +func pivotRoot(rootfs string) error { + // While the documentation may claim otherwise, pivot_root(".", ".") is + // actually valid. What this results in is / being the new root but + // /proc/self/cwd being the old root. Since we can play around with the cwd + // with pivot_root this allows us to pivot without creating directories in + // the rootfs. Shout-outs to the LXC developers for giving us this idea. + + oldroot, err := unix.Open("/", unix.O_DIRECTORY|unix.O_RDONLY, 0) + if err != nil { + return err + } + defer unix.Close(oldroot) + + newroot, err := unix.Open(rootfs, unix.O_DIRECTORY|unix.O_RDONLY, 0) + if err != nil { + return err + } + defer unix.Close(newroot) + + // Change to the new root so that the pivot_root actually acts on it. + if err := unix.Fchdir(newroot); err != nil { + return err + } + + if err := unix.PivotRoot(".", "."); err != nil { + return fmt.Errorf("pivot_root %s", err) + } + + // Currently our "." is oldroot (according to the current kernel code). + // However, purely for safety, we will fchdir(oldroot) since there isn't + // really any guarantee from the kernel what /proc/self/cwd will be after a + // pivot_root(2). + + if err := unix.Fchdir(oldroot); err != nil { + return err + } + + // Make oldroot rslave to make sure our unmounts don't propagate to the + // host (and thus bork the machine). We don't use rprivate because this is + // known to cause issues due to races where we still have a reference to a + // mount while a process in the host namespace are trying to operate on + // something they think has no mounts (devicemapper in particular). + if err := unix.Mount("", ".", "", unix.MS_SLAVE|unix.MS_REC, ""); err != nil { + return err + } + // Preform the unmount. MNT_DETACH allows us to unmount /proc/self/cwd. + if err := unix.Unmount(".", unix.MNT_DETACH); err != nil { + return err + } + + // Switch back to our shiny new root. + if err := unix.Chdir("/"); err != nil { + return fmt.Errorf("chdir / %s", err) + } + return nil +} + +func msMoveRoot(rootfs string) error { + if err := unix.Mount(rootfs, "/", "", unix.MS_MOVE, ""); err != nil { + return err + } + return chroot(rootfs) +} + +func chroot(rootfs string) error { + if err := unix.Chroot("."); err != nil { + return err + } + return unix.Chdir("/") +} + +// createIfNotExists creates a file or a directory only if it does not already exist. +func createIfNotExists(path string, isDir bool) error { + if _, err := os.Stat(path); err != nil { + if os.IsNotExist(err) { + if isDir { + return os.MkdirAll(path, 0755) + } + if err := os.MkdirAll(filepath.Dir(path), 0755); err != nil { + return err + } + f, err := os.OpenFile(path, os.O_CREATE, 0755) + if err != nil { + return err + } + f.Close() + } + } + return nil +} + +// readonlyPath will make a path read only. +func readonlyPath(path string) error { + if err := unix.Mount(path, path, "", unix.MS_BIND|unix.MS_REC, ""); err != nil { + if os.IsNotExist(err) { + return nil + } + return err + } + return unix.Mount(path, path, "", unix.MS_BIND|unix.MS_REMOUNT|unix.MS_RDONLY|unix.MS_REC, "") +} + +// remountReadonly will remount an existing mount point and ensure that it is read-only. +func remountReadonly(m *configs.Mount) error { + var ( + dest = m.Destination + flags = m.Flags + ) + for i := 0; i < 5; i++ { + // There is a special case in the kernel for + // MS_REMOUNT | MS_BIND, which allows us to change only the + // flags even as an unprivileged user (i.e. user namespace) + // assuming we don't drop any security related flags (nodev, + // nosuid, etc.). So, let's use that case so that we can do + // this re-mount without failing in a userns. + flags |= unix.MS_REMOUNT | unix.MS_BIND | unix.MS_RDONLY + if err := unix.Mount("", dest, "", uintptr(flags), ""); err != nil { + switch err { + case unix.EBUSY: + time.Sleep(100 * time.Millisecond) + continue + default: + return err + } + } + return nil + } + return fmt.Errorf("unable to mount %s as readonly max retries reached", dest) +} + +// maskPath masks the top of the specified path inside a container to avoid +// security issues from processes reading information from non-namespace aware +// mounts ( proc/kcore ). +// For files, maskPath bind mounts /dev/null over the top of the specified path. +// For directories, maskPath mounts read-only tmpfs over the top of the specified path. +func maskPath(path string, mountLabel string) error { + if err := unix.Mount("/dev/null", path, "", unix.MS_BIND, ""); err != nil && !os.IsNotExist(err) { + if err == unix.ENOTDIR { + return unix.Mount("tmpfs", path, "tmpfs", unix.MS_RDONLY, label.FormatMountLabel("", mountLabel)) + } + return err + } + return nil +} + +// writeSystemProperty writes the value to a path under /proc/sys as determined from the key. +// For e.g. net.ipv4.ip_forward translated to /proc/sys/net/ipv4/ip_forward. +func writeSystemProperty(key, value string) error { + keyPath := strings.Replace(key, ".", "/", -1) + return ioutil.WriteFile(path.Join("/proc/sys", keyPath), []byte(value), 0644) +} + +func remount(m *configs.Mount, rootfs string) error { + var ( + dest = m.Destination + ) + if !strings.HasPrefix(dest, rootfs) { + dest = filepath.Join(rootfs, dest) + } + if err := unix.Mount(m.Source, dest, m.Device, uintptr(m.Flags|unix.MS_REMOUNT), ""); err != nil { + return err + } + return nil +} + +// Do the mount operation followed by additional mounts required to take care +// of propagation flags. +func mountPropagate(m *configs.Mount, rootfs string, mountLabel string) error { + var ( + dest = m.Destination + data = label.FormatMountLabel(m.Data, mountLabel) + flags = m.Flags + ) + if libcontainerUtils.CleanPath(dest) == "/dev" { + flags &= ^unix.MS_RDONLY + } + + copyUp := m.Extensions&configs.EXT_COPYUP == configs.EXT_COPYUP + if !(copyUp || strings.HasPrefix(dest, rootfs)) { + dest = filepath.Join(rootfs, dest) + } + + if err := unix.Mount(m.Source, dest, m.Device, uintptr(flags), data); err != nil { + return err + } + + for _, pflag := range m.PropagationFlags { + if err := unix.Mount("", dest, "", uintptr(pflag), ""); err != nil { + return err + } + } + return nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/seccomp/config.go b/vendor/github.com/opencontainers/runc/libcontainer/seccomp/config.go new file mode 100644 index 000000000..ded5a6bbc --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/seccomp/config.go @@ -0,0 +1,76 @@ +package seccomp + +import ( + "fmt" + + "github.com/opencontainers/runc/libcontainer/configs" +) + +var operators = map[string]configs.Operator{ + "SCMP_CMP_NE": configs.NotEqualTo, + "SCMP_CMP_LT": configs.LessThan, + "SCMP_CMP_LE": configs.LessThanOrEqualTo, + "SCMP_CMP_EQ": configs.EqualTo, + "SCMP_CMP_GE": configs.GreaterThanOrEqualTo, + "SCMP_CMP_GT": configs.GreaterThan, + "SCMP_CMP_MASKED_EQ": configs.MaskEqualTo, +} + +var actions = map[string]configs.Action{ + "SCMP_ACT_KILL": configs.Kill, + "SCMP_ACT_ERRNO": configs.Errno, + "SCMP_ACT_TRAP": configs.Trap, + "SCMP_ACT_ALLOW": configs.Allow, + "SCMP_ACT_TRACE": configs.Trace, +} + +var archs = map[string]string{ + "SCMP_ARCH_X86": "x86", + "SCMP_ARCH_X86_64": "amd64", + "SCMP_ARCH_X32": "x32", + "SCMP_ARCH_ARM": "arm", + "SCMP_ARCH_AARCH64": "arm64", + "SCMP_ARCH_MIPS": "mips", + "SCMP_ARCH_MIPS64": "mips64", + "SCMP_ARCH_MIPS64N32": "mips64n32", + "SCMP_ARCH_MIPSEL": "mipsel", + "SCMP_ARCH_MIPSEL64": "mipsel64", + "SCMP_ARCH_MIPSEL64N32": "mipsel64n32", + "SCMP_ARCH_PPC": "ppc", + "SCMP_ARCH_PPC64": "ppc64", + "SCMP_ARCH_PPC64LE": "ppc64le", + "SCMP_ARCH_S390": "s390", + "SCMP_ARCH_S390X": "s390x", +} + +// ConvertStringToOperator converts a string into a Seccomp comparison operator. +// Comparison operators use the names they are assigned by Libseccomp's header. +// Attempting to convert a string that is not a valid operator results in an +// error. +func ConvertStringToOperator(in string) (configs.Operator, error) { + if op, ok := operators[in]; ok == true { + return op, nil + } + return 0, fmt.Errorf("string %s is not a valid operator for seccomp", in) +} + +// ConvertStringToAction converts a string into a Seccomp rule match action. +// Actions use the names they are assigned in Libseccomp's header, though some +// (notable, SCMP_ACT_TRACE) are not available in this implementation and will +// return errors. +// Attempting to convert a string that is not a valid action results in an +// error. +func ConvertStringToAction(in string) (configs.Action, error) { + if act, ok := actions[in]; ok == true { + return act, nil + } + return 0, fmt.Errorf("string %s is not a valid action for seccomp", in) +} + +// ConvertStringToArch converts a string into a Seccomp comparison arch. +func ConvertStringToArch(in string) (string, error) { + if arch, ok := archs[in]; ok == true { + return arch, nil + } + return "", fmt.Errorf("string %s is not a valid arch for seccomp", in) +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/seccomp/seccomp_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/seccomp/seccomp_linux.go new file mode 100644 index 000000000..d99f3fe64 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/seccomp/seccomp_linux.go @@ -0,0 +1,258 @@ +// +build linux,cgo,seccomp + +package seccomp + +import ( + "bufio" + "fmt" + "os" + "strings" + + "github.com/opencontainers/runc/libcontainer/configs" + libseccomp "github.com/seccomp/libseccomp-golang" + + "golang.org/x/sys/unix" +) + +var ( + actAllow = libseccomp.ActAllow + actTrap = libseccomp.ActTrap + actKill = libseccomp.ActKill + actTrace = libseccomp.ActTrace.SetReturnCode(int16(unix.EPERM)) + actErrno = libseccomp.ActErrno.SetReturnCode(int16(unix.EPERM)) +) + +const ( + // Linux system calls can have at most 6 arguments + syscallMaxArguments int = 6 +) + +// Filters given syscalls in a container, preventing them from being used +// Started in the container init process, and carried over to all child processes +// Setns calls, however, require a separate invocation, as they are not children +// of the init until they join the namespace +func InitSeccomp(config *configs.Seccomp) error { + if config == nil { + return fmt.Errorf("cannot initialize Seccomp - nil config passed") + } + + defaultAction, err := getAction(config.DefaultAction) + if err != nil { + return fmt.Errorf("error initializing seccomp - invalid default action") + } + + filter, err := libseccomp.NewFilter(defaultAction) + if err != nil { + return fmt.Errorf("error creating filter: %s", err) + } + + // Add extra architectures + for _, arch := range config.Architectures { + scmpArch, err := libseccomp.GetArchFromString(arch) + if err != nil { + return fmt.Errorf("error validating Seccomp architecture: %s", err) + } + + if err := filter.AddArch(scmpArch); err != nil { + return fmt.Errorf("error adding architecture to seccomp filter: %s", err) + } + } + + // Unset no new privs bit + if err := filter.SetNoNewPrivsBit(false); err != nil { + return fmt.Errorf("error setting no new privileges: %s", err) + } + + // Add a rule for each syscall + for _, call := range config.Syscalls { + if call == nil { + return fmt.Errorf("encountered nil syscall while initializing Seccomp") + } + + if err = matchCall(filter, call); err != nil { + return err + } + } + + if err = filter.Load(); err != nil { + return fmt.Errorf("error loading seccomp filter into kernel: %s", err) + } + + return nil +} + +// IsEnabled returns if the kernel has been configured to support seccomp. +func IsEnabled() bool { + // Try to read from /proc/self/status for kernels > 3.8 + s, err := parseStatusFile("/proc/self/status") + if err != nil { + // Check if Seccomp is supported, via CONFIG_SECCOMP. + if err := unix.Prctl(unix.PR_GET_SECCOMP, 0, 0, 0, 0); err != unix.EINVAL { + // Make sure the kernel has CONFIG_SECCOMP_FILTER. + if err := unix.Prctl(unix.PR_SET_SECCOMP, unix.SECCOMP_MODE_FILTER, 0, 0, 0); err != unix.EINVAL { + return true + } + } + return false + } + _, ok := s["Seccomp"] + return ok +} + +// Convert Libcontainer Action to Libseccomp ScmpAction +func getAction(act configs.Action) (libseccomp.ScmpAction, error) { + switch act { + case configs.Kill: + return actKill, nil + case configs.Errno: + return actErrno, nil + case configs.Trap: + return actTrap, nil + case configs.Allow: + return actAllow, nil + case configs.Trace: + return actTrace, nil + default: + return libseccomp.ActInvalid, fmt.Errorf("invalid action, cannot use in rule") + } +} + +// Convert Libcontainer Operator to Libseccomp ScmpCompareOp +func getOperator(op configs.Operator) (libseccomp.ScmpCompareOp, error) { + switch op { + case configs.EqualTo: + return libseccomp.CompareEqual, nil + case configs.NotEqualTo: + return libseccomp.CompareNotEqual, nil + case configs.GreaterThan: + return libseccomp.CompareGreater, nil + case configs.GreaterThanOrEqualTo: + return libseccomp.CompareGreaterEqual, nil + case configs.LessThan: + return libseccomp.CompareLess, nil + case configs.LessThanOrEqualTo: + return libseccomp.CompareLessOrEqual, nil + case configs.MaskEqualTo: + return libseccomp.CompareMaskedEqual, nil + default: + return libseccomp.CompareInvalid, fmt.Errorf("invalid operator, cannot use in rule") + } +} + +// Convert Libcontainer Arg to Libseccomp ScmpCondition +func getCondition(arg *configs.Arg) (libseccomp.ScmpCondition, error) { + cond := libseccomp.ScmpCondition{} + + if arg == nil { + return cond, fmt.Errorf("cannot convert nil to syscall condition") + } + + op, err := getOperator(arg.Op) + if err != nil { + return cond, err + } + + return libseccomp.MakeCondition(arg.Index, op, arg.Value, arg.ValueTwo) +} + +// Add a rule to match a single syscall +func matchCall(filter *libseccomp.ScmpFilter, call *configs.Syscall) error { + if call == nil || filter == nil { + return fmt.Errorf("cannot use nil as syscall to block") + } + + if len(call.Name) == 0 { + return fmt.Errorf("empty string is not a valid syscall") + } + + // If we can't resolve the syscall, assume it's not supported on this kernel + // Ignore it, don't error out + callNum, err := libseccomp.GetSyscallFromName(call.Name) + if err != nil { + return nil + } + + // Convert the call's action to the libseccomp equivalent + callAct, err := getAction(call.Action) + if err != nil { + return fmt.Errorf("action in seccomp profile is invalid: %s", err) + } + + // Unconditional match - just add the rule + if len(call.Args) == 0 { + if err = filter.AddRule(callNum, callAct); err != nil { + return fmt.Errorf("error adding seccomp filter rule for syscall %s: %s", call.Name, err) + } + } else { + // If two or more arguments have the same condition, + // Revert to old behavior, adding each condition as a separate rule + argCounts := make([]uint, syscallMaxArguments) + conditions := []libseccomp.ScmpCondition{} + + for _, cond := range call.Args { + newCond, err := getCondition(cond) + if err != nil { + return fmt.Errorf("error creating seccomp syscall condition for syscall %s: %s", call.Name, err) + } + + argCounts[cond.Index] += 1 + + conditions = append(conditions, newCond) + } + + hasMultipleArgs := false + for _, count := range argCounts { + if count > 1 { + hasMultipleArgs = true + break + } + } + + if hasMultipleArgs { + // Revert to old behavior + // Add each condition attached to a separate rule + for _, cond := range conditions { + condArr := []libseccomp.ScmpCondition{cond} + + if err = filter.AddRuleConditional(callNum, callAct, condArr); err != nil { + return fmt.Errorf("error adding seccomp rule for syscall %s: %s", call.Name, err) + } + } + } else { + // No conditions share same argument + // Use new, proper behavior + if err = filter.AddRuleConditional(callNum, callAct, conditions); err != nil { + return fmt.Errorf("error adding seccomp rule for syscall %s: %s", call.Name, err) + } + } + } + + return nil +} + +func parseStatusFile(path string) (map[string]string, error) { + f, err := os.Open(path) + if err != nil { + return nil, err + } + defer f.Close() + + s := bufio.NewScanner(f) + status := make(map[string]string) + + for s.Scan() { + text := s.Text() + parts := strings.Split(text, ":") + + if len(parts) <= 1 { + continue + } + + status[parts[0]] = parts[1] + } + if err := s.Err(); err != nil { + return nil, err + } + + return status, nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/seccomp/seccomp_unsupported.go b/vendor/github.com/opencontainers/runc/libcontainer/seccomp/seccomp_unsupported.go new file mode 100644 index 000000000..44df1ad4c --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/seccomp/seccomp_unsupported.go @@ -0,0 +1,24 @@ +// +build !linux !cgo !seccomp + +package seccomp + +import ( + "errors" + + "github.com/opencontainers/runc/libcontainer/configs" +) + +var ErrSeccompNotEnabled = errors.New("seccomp: config provided but seccomp not supported") + +// InitSeccomp does nothing because seccomp is not supported. +func InitSeccomp(config *configs.Seccomp) error { + if config != nil { + return ErrSeccompNotEnabled + } + return nil +} + +// IsEnabled returns false, because it is not supported. +func IsEnabled() bool { + return false +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/setns_init_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/setns_init_linux.go new file mode 100644 index 000000000..096c601e7 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/setns_init_linux.go @@ -0,0 +1,76 @@ +// +build linux + +package libcontainer + +import ( + "fmt" + "os" + + "github.com/opencontainers/runc/libcontainer/apparmor" + "github.com/opencontainers/runc/libcontainer/keys" + "github.com/opencontainers/runc/libcontainer/seccomp" + "github.com/opencontainers/runc/libcontainer/system" + "github.com/opencontainers/selinux/go-selinux/label" + + "golang.org/x/sys/unix" +) + +// linuxSetnsInit performs the container's initialization for running a new process +// inside an existing container. +type linuxSetnsInit struct { + pipe *os.File + consoleSocket *os.File + config *initConfig +} + +func (l *linuxSetnsInit) getSessionRingName() string { + return fmt.Sprintf("_ses.%s", l.config.ContainerId) +} + +func (l *linuxSetnsInit) Init() error { + if !l.config.Config.NoNewKeyring { + // do not inherit the parent's session keyring + if _, err := keys.JoinSessionKeyring(l.getSessionRingName()); err != nil { + return err + } + } + if l.config.CreateConsole { + if err := setupConsole(l.consoleSocket, l.config, false); err != nil { + return err + } + if err := system.Setctty(); err != nil { + return err + } + } + if l.config.NoNewPrivileges { + if err := unix.Prctl(unix.PR_SET_NO_NEW_PRIVS, 1, 0, 0, 0); err != nil { + return err + } + } + // Without NoNewPrivileges seccomp is a privileged operation, so we need to + // do this before dropping capabilities; otherwise do it as late as possible + // just before execve so as few syscalls take place after it as possible. + if l.config.Config.Seccomp != nil && !l.config.NoNewPrivileges { + if err := seccomp.InitSeccomp(l.config.Config.Seccomp); err != nil { + return err + } + } + if err := finalizeNamespace(l.config); err != nil { + return err + } + if err := apparmor.ApplyProfile(l.config.AppArmorProfile); err != nil { + return err + } + if err := label.SetProcessLabel(l.config.ProcessLabel); err != nil { + return err + } + // Set seccomp as close to execve as possible, so as few syscalls take + // place afterward (reducing the amount of syscalls that users need to + // enable in their seccomp profiles). + if l.config.Config.Seccomp != nil && l.config.NoNewPrivileges { + if err := seccomp.InitSeccomp(l.config.Config.Seccomp); err != nil { + return newSystemErrorWithCause(err, "init seccomp") + } + } + return system.Execv(l.config.Args[0], l.config.Args[0:], os.Environ()) +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/stacktrace/capture.go b/vendor/github.com/opencontainers/runc/libcontainer/stacktrace/capture.go new file mode 100644 index 000000000..0bbe14950 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/stacktrace/capture.go @@ -0,0 +1,27 @@ +package stacktrace + +import "runtime" + +// Capture captures a stacktrace for the current calling go program +// +// skip is the number of frames to skip +func Capture(userSkip int) Stacktrace { + var ( + skip = userSkip + 1 // add one for our own function + frames []Frame + prevPc uintptr + ) + for i := skip; ; i++ { + pc, file, line, ok := runtime.Caller(i) + //detect if caller is repeated to avoid loop, gccgo + //currently runs into a loop without this check + if !ok || pc == prevPc { + break + } + frames = append(frames, NewFrame(pc, file, line)) + prevPc = pc + } + return Stacktrace{ + Frames: frames, + } +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/stacktrace/frame.go b/vendor/github.com/opencontainers/runc/libcontainer/stacktrace/frame.go new file mode 100644 index 000000000..0d590d9a5 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/stacktrace/frame.go @@ -0,0 +1,38 @@ +package stacktrace + +import ( + "path/filepath" + "runtime" + "strings" +) + +// NewFrame returns a new stack frame for the provided information +func NewFrame(pc uintptr, file string, line int) Frame { + fn := runtime.FuncForPC(pc) + if fn == nil { + return Frame{} + } + pack, name := parseFunctionName(fn.Name()) + return Frame{ + Line: line, + File: filepath.Base(file), + Package: pack, + Function: name, + } +} + +func parseFunctionName(name string) (string, string) { + i := strings.LastIndex(name, ".") + if i == -1 { + return "", name + } + return name[:i], name[i+1:] +} + +// Frame contains all the information for a stack frame within a go program +type Frame struct { + File string + Function string + Package string + Line int +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/stacktrace/stacktrace.go b/vendor/github.com/opencontainers/runc/libcontainer/stacktrace/stacktrace.go new file mode 100644 index 000000000..5e8b58d2d --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/stacktrace/stacktrace.go @@ -0,0 +1,5 @@ +package stacktrace + +type Stacktrace struct { + Frames []Frame +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/standard_init_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/standard_init_linux.go new file mode 100644 index 000000000..0b6530d68 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/standard_init_linux.go @@ -0,0 +1,194 @@ +// +build linux + +package libcontainer + +import ( + "fmt" + "os" + "os/exec" + "syscall" //only for Exec + + "github.com/opencontainers/runc/libcontainer/apparmor" + "github.com/opencontainers/runc/libcontainer/configs" + "github.com/opencontainers/runc/libcontainer/keys" + "github.com/opencontainers/runc/libcontainer/seccomp" + "github.com/opencontainers/runc/libcontainer/system" + "github.com/opencontainers/selinux/go-selinux/label" + "github.com/pkg/errors" + + "golang.org/x/sys/unix" +) + +type linuxStandardInit struct { + pipe *os.File + consoleSocket *os.File + parentPid int + fifoFd int + config *initConfig +} + +func (l *linuxStandardInit) getSessionRingParams() (string, uint32, uint32) { + var newperms uint32 + + if l.config.Config.Namespaces.Contains(configs.NEWUSER) { + // With user ns we need 'other' search permissions. + newperms = 0x8 + } else { + // Without user ns we need 'UID' search permissions. + newperms = 0x80000 + } + + // Create a unique per session container name that we can join in setns; + // However, other containers can also join it. + return fmt.Sprintf("_ses.%s", l.config.ContainerId), 0xffffffff, newperms +} + +func (l *linuxStandardInit) Init() error { + if !l.config.Config.NoNewKeyring { + ringname, keepperms, newperms := l.getSessionRingParams() + + // Do not inherit the parent's session keyring. + sessKeyId, err := keys.JoinSessionKeyring(ringname) + if err != nil { + return errors.Wrap(err, "join session keyring") + } + // Make session keyring searcheable. + if err := keys.ModKeyringPerm(sessKeyId, keepperms, newperms); err != nil { + return errors.Wrap(err, "mod keyring permissions") + } + } + + if err := setupNetwork(l.config); err != nil { + return err + } + if err := setupRoute(l.config.Config); err != nil { + return err + } + + label.Init() + if err := prepareRootfs(l.pipe, l.config); err != nil { + return err + } + // Set up the console. This has to be done *before* we finalize the rootfs, + // but *after* we've given the user the chance to set up all of the mounts + // they wanted. + if l.config.CreateConsole { + if err := setupConsole(l.consoleSocket, l.config, true); err != nil { + return err + } + if err := system.Setctty(); err != nil { + return errors.Wrap(err, "setctty") + } + } + + // Finish the rootfs setup. + if l.config.Config.Namespaces.Contains(configs.NEWNS) { + if err := finalizeRootfs(l.config.Config); err != nil { + return err + } + } + + if hostname := l.config.Config.Hostname; hostname != "" { + if err := unix.Sethostname([]byte(hostname)); err != nil { + return errors.Wrap(err, "sethostname") + } + } + if err := apparmor.ApplyProfile(l.config.AppArmorProfile); err != nil { + return errors.Wrap(err, "apply apparmor profile") + } + if err := label.SetProcessLabel(l.config.ProcessLabel); err != nil { + return errors.Wrap(err, "set process label") + } + + for key, value := range l.config.Config.Sysctl { + if err := writeSystemProperty(key, value); err != nil { + return errors.Wrapf(err, "write sysctl key %s", key) + } + } + for _, path := range l.config.Config.ReadonlyPaths { + if err := readonlyPath(path); err != nil { + return errors.Wrapf(err, "readonly path %s", path) + } + } + for _, path := range l.config.Config.MaskPaths { + if err := maskPath(path, l.config.Config.MountLabel); err != nil { + return errors.Wrapf(err, "mask path %s", path) + } + } + pdeath, err := system.GetParentDeathSignal() + if err != nil { + return errors.Wrap(err, "get pdeath signal") + } + if l.config.NoNewPrivileges { + if err := unix.Prctl(unix.PR_SET_NO_NEW_PRIVS, 1, 0, 0, 0); err != nil { + return errors.Wrap(err, "set nonewprivileges") + } + } + // Tell our parent that we're ready to Execv. This must be done before the + // Seccomp rules have been applied, because we need to be able to read and + // write to a socket. + if err := syncParentReady(l.pipe); err != nil { + return errors.Wrap(err, "sync ready") + } + // Without NoNewPrivileges seccomp is a privileged operation, so we need to + // do this before dropping capabilities; otherwise do it as late as possible + // just before execve so as few syscalls take place after it as possible. + if l.config.Config.Seccomp != nil && !l.config.NoNewPrivileges { + if err := seccomp.InitSeccomp(l.config.Config.Seccomp); err != nil { + return err + } + } + if err := finalizeNamespace(l.config); err != nil { + return err + } + // finalizeNamespace can change user/group which clears the parent death + // signal, so we restore it here. + if err := pdeath.Restore(); err != nil { + return errors.Wrap(err, "restore pdeath signal") + } + // Compare the parent from the initial start of the init process and make + // sure that it did not change. if the parent changes that means it died + // and we were reparented to something else so we should just kill ourself + // and not cause problems for someone else. + if unix.Getppid() != l.parentPid { + return unix.Kill(unix.Getpid(), unix.SIGKILL) + } + // Check for the arg before waiting to make sure it exists and it is + // returned as a create time error. + name, err := exec.LookPath(l.config.Args[0]) + if err != nil { + return err + } + // Close the pipe to signal that we have completed our init. + l.pipe.Close() + // Wait for the FIFO to be opened on the other side before exec-ing the + // user process. We open it through /proc/self/fd/$fd, because the fd that + // was given to us was an O_PATH fd to the fifo itself. Linux allows us to + // re-open an O_PATH fd through /proc. + fd, err := unix.Open(fmt.Sprintf("/proc/self/fd/%d", l.fifoFd), unix.O_WRONLY|unix.O_CLOEXEC, 0) + if err != nil { + return newSystemErrorWithCause(err, "open exec fifo") + } + if _, err := unix.Write(fd, []byte("0")); err != nil { + return newSystemErrorWithCause(err, "write 0 exec fifo") + } + // Close the O_PATH fifofd fd before exec because the kernel resets + // dumpable in the wrong order. This has been fixed in newer kernels, but + // we keep this to ensure CVE-2016-9962 doesn't re-emerge on older kernels. + // N.B. the core issue itself (passing dirfds to the host filesystem) has + // since been resolved. + // https://github.com/torvalds/linux/blob/v4.9/fs/exec.c#L1290-L1318 + unix.Close(l.fifoFd) + // Set seccomp as close to execve as possible, so as few syscalls take + // place afterward (reducing the amount of syscalls that users need to + // enable in their seccomp profiles). + if l.config.Config.Seccomp != nil && l.config.NoNewPrivileges { + if err := seccomp.InitSeccomp(l.config.Config.Seccomp); err != nil { + return newSystemErrorWithCause(err, "init seccomp") + } + } + if err := syscall.Exec(name, l.config.Args[0:], os.Environ()); err != nil { + return newSystemErrorWithCause(err, "exec user process") + } + return nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/state_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/state_linux.go new file mode 100644 index 000000000..b45ce23e4 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/state_linux.go @@ -0,0 +1,255 @@ +// +build linux + +package libcontainer + +import ( + "fmt" + "os" + "path/filepath" + + "github.com/opencontainers/runc/libcontainer/configs" + "github.com/opencontainers/runc/libcontainer/utils" + + "github.com/sirupsen/logrus" + "golang.org/x/sys/unix" +) + +func newStateTransitionError(from, to containerState) error { + return &stateTransitionError{ + From: from.status().String(), + To: to.status().String(), + } +} + +// stateTransitionError is returned when an invalid state transition happens from one +// state to another. +type stateTransitionError struct { + From string + To string +} + +func (s *stateTransitionError) Error() string { + return fmt.Sprintf("invalid state transition from %s to %s", s.From, s.To) +} + +type containerState interface { + transition(containerState) error + destroy() error + status() Status +} + +func destroy(c *linuxContainer) error { + if !c.config.Namespaces.Contains(configs.NEWPID) { + if err := signalAllProcesses(c.cgroupManager, unix.SIGKILL); err != nil { + logrus.Warn(err) + } + } + err := c.cgroupManager.Destroy() + if c.intelRdtManager != nil { + if ierr := c.intelRdtManager.Destroy(); err == nil { + err = ierr + } + } + if rerr := os.RemoveAll(c.root); err == nil { + err = rerr + } + c.initProcess = nil + if herr := runPoststopHooks(c); err == nil { + err = herr + } + c.state = &stoppedState{c: c} + return err +} + +func runPoststopHooks(c *linuxContainer) error { + if c.config.Hooks != nil { + bundle, annotations := utils.Annotations(c.config.Labels) + s := configs.HookState{ + Version: c.config.Version, + ID: c.id, + Bundle: bundle, + Annotations: annotations, + } + for _, hook := range c.config.Hooks.Poststop { + if err := hook.Run(s); err != nil { + return err + } + } + } + return nil +} + +// stoppedState represents a container is a stopped/destroyed state. +type stoppedState struct { + c *linuxContainer +} + +func (b *stoppedState) status() Status { + return Stopped +} + +func (b *stoppedState) transition(s containerState) error { + switch s.(type) { + case *runningState, *restoredState: + b.c.state = s + return nil + case *stoppedState: + return nil + } + return newStateTransitionError(b, s) +} + +func (b *stoppedState) destroy() error { + return destroy(b.c) +} + +// runningState represents a container that is currently running. +type runningState struct { + c *linuxContainer +} + +func (r *runningState) status() Status { + return Running +} + +func (r *runningState) transition(s containerState) error { + switch s.(type) { + case *stoppedState: + t, err := r.c.runType() + if err != nil { + return err + } + if t == Running { + return newGenericError(fmt.Errorf("container still running"), ContainerNotStopped) + } + r.c.state = s + return nil + case *pausedState: + r.c.state = s + return nil + case *runningState: + return nil + } + return newStateTransitionError(r, s) +} + +func (r *runningState) destroy() error { + t, err := r.c.runType() + if err != nil { + return err + } + if t == Running { + return newGenericError(fmt.Errorf("container is not destroyed"), ContainerNotStopped) + } + return destroy(r.c) +} + +type createdState struct { + c *linuxContainer +} + +func (i *createdState) status() Status { + return Created +} + +func (i *createdState) transition(s containerState) error { + switch s.(type) { + case *runningState, *pausedState, *stoppedState: + i.c.state = s + return nil + case *createdState: + return nil + } + return newStateTransitionError(i, s) +} + +func (i *createdState) destroy() error { + i.c.initProcess.signal(unix.SIGKILL) + return destroy(i.c) +} + +// pausedState represents a container that is currently pause. It cannot be destroyed in a +// paused state and must transition back to running first. +type pausedState struct { + c *linuxContainer +} + +func (p *pausedState) status() Status { + return Paused +} + +func (p *pausedState) transition(s containerState) error { + switch s.(type) { + case *runningState, *stoppedState: + p.c.state = s + return nil + case *pausedState: + return nil + } + return newStateTransitionError(p, s) +} + +func (p *pausedState) destroy() error { + t, err := p.c.runType() + if err != nil { + return err + } + if t != Running && t != Created { + if err := p.c.cgroupManager.Freeze(configs.Thawed); err != nil { + return err + } + return destroy(p.c) + } + return newGenericError(fmt.Errorf("container is paused"), ContainerPaused) +} + +// restoredState is the same as the running state but also has associated checkpoint +// information that maybe need destroyed when the container is stopped and destroy is called. +type restoredState struct { + imageDir string + c *linuxContainer +} + +func (r *restoredState) status() Status { + return Running +} + +func (r *restoredState) transition(s containerState) error { + switch s.(type) { + case *stoppedState, *runningState: + return nil + } + return newStateTransitionError(r, s) +} + +func (r *restoredState) destroy() error { + if _, err := os.Stat(filepath.Join(r.c.root, "checkpoint")); err != nil { + if !os.IsNotExist(err) { + return err + } + } + return destroy(r.c) +} + +// loadedState is used whenever a container is restored, loaded, or setting additional +// processes inside and it should not be destroyed when it is exiting. +type loadedState struct { + c *linuxContainer + s Status +} + +func (n *loadedState) status() Status { + return n.s +} + +func (n *loadedState) transition(s containerState) error { + n.c.state = s + return nil +} + +func (n *loadedState) destroy() error { + if err := n.c.refreshState(); err != nil { + return err + } + return n.c.state.destroy() +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/stats.go b/vendor/github.com/opencontainers/runc/libcontainer/stats.go new file mode 100644 index 000000000..303e4b94c --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/stats.go @@ -0,0 +1,15 @@ +package libcontainer + +type NetworkInterface struct { + // Name is the name of the network interface. + Name string + + RxBytes uint64 + RxPackets uint64 + RxErrors uint64 + RxDropped uint64 + TxBytes uint64 + TxPackets uint64 + TxErrors uint64 + TxDropped uint64 +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/stats_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/stats_linux.go new file mode 100644 index 000000000..29fd641e9 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/stats_linux.go @@ -0,0 +1,10 @@ +package libcontainer + +import "github.com/opencontainers/runc/libcontainer/cgroups" +import "github.com/opencontainers/runc/libcontainer/intelrdt" + +type Stats struct { + Interfaces []*NetworkInterface + CgroupStats *cgroups.Stats + IntelRdtStats *intelrdt.Stats +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/sync.go b/vendor/github.com/opencontainers/runc/libcontainer/sync.go new file mode 100644 index 000000000..cf7b45bc3 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/sync.go @@ -0,0 +1,107 @@ +package libcontainer + +import ( + "encoding/json" + "fmt" + "io" + + "github.com/opencontainers/runc/libcontainer/utils" +) + +type syncType string + +// Constants that are used for synchronisation between the parent and child +// during container setup. They come in pairs (with procError being a generic +// response which is followed by a &genericError). +// +// [ child ] <-> [ parent ] +// +// procHooks --> [run hooks] +// <-- procResume +// +// procConsole --> +// <-- procConsoleReq +// [send(fd)] --> [recv(fd)] +// <-- procConsoleAck +// +// procReady --> [final setup] +// <-- procRun +const ( + procError syncType = "procError" + procReady syncType = "procReady" + procRun syncType = "procRun" + procHooks syncType = "procHooks" + procResume syncType = "procResume" +) + +type syncT struct { + Type syncType `json:"type"` +} + +// writeSync is used to write to a synchronisation pipe. An error is returned +// if there was a problem writing the payload. +func writeSync(pipe io.Writer, sync syncType) error { + if err := utils.WriteJSON(pipe, syncT{sync}); err != nil { + return err + } + return nil +} + +// readSync is used to read from a synchronisation pipe. An error is returned +// if we got a genericError, the pipe was closed, or we got an unexpected flag. +func readSync(pipe io.Reader, expected syncType) error { + var procSync syncT + if err := json.NewDecoder(pipe).Decode(&procSync); err != nil { + if err == io.EOF { + return fmt.Errorf("parent closed synchronisation channel") + } + + if procSync.Type == procError { + var ierr genericError + + if err := json.NewDecoder(pipe).Decode(&ierr); err != nil { + return fmt.Errorf("failed reading error from parent: %v", err) + } + + return &ierr + } + + if procSync.Type != expected { + return fmt.Errorf("invalid synchronisation flag from parent") + } + } + return nil +} + +// parseSync runs the given callback function on each syncT received from the +// child. It will return once io.EOF is returned from the given pipe. +func parseSync(pipe io.Reader, fn func(*syncT) error) error { + dec := json.NewDecoder(pipe) + for { + var sync syncT + if err := dec.Decode(&sync); err != nil { + if err == io.EOF { + break + } + return err + } + + // We handle this case outside fn for cleanliness reasons. + var ierr *genericError + if sync.Type == procError { + if err := dec.Decode(&ierr); err != nil && err != io.EOF { + return newSystemErrorWithCause(err, "decoding proc error from init") + } + if ierr != nil { + return ierr + } + // Programmer error. + panic("No error following JSON procError payload.") + } + + if err := fn(&sync); err != nil { + return err + } + } + return nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/system/linux.go b/vendor/github.com/opencontainers/runc/libcontainer/system/linux.go index 8b199d92e..a4ae8901a 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/system/linux.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/system/linux.go @@ -3,12 +3,13 @@ package system import ( - "bufio" - "fmt" "os" "os/exec" - "syscall" + "syscall" // only for exec "unsafe" + + "github.com/opencontainers/runc/libcontainer/user" + "golang.org/x/sys/unix" ) // If arg2 is nonzero, set the "child subreaper" attribute of the @@ -53,8 +54,8 @@ func Execv(cmd string, args []string, env []string) error { return syscall.Exec(name, args, env) } -func Prlimit(pid, resource int, limit syscall.Rlimit) error { - _, _, err := syscall.RawSyscall6(syscall.SYS_PRLIMIT64, uintptr(pid), uintptr(resource), uintptr(unsafe.Pointer(&limit)), uintptr(unsafe.Pointer(&limit)), 0, 0) +func Prlimit(pid, resource int, limit unix.Rlimit) error { + _, _, err := unix.RawSyscall6(unix.SYS_PRLIMIT64, uintptr(pid), uintptr(resource), uintptr(unsafe.Pointer(&limit)), uintptr(unsafe.Pointer(&limit)), 0, 0) if err != 0 { return err } @@ -62,7 +63,7 @@ func Prlimit(pid, resource int, limit syscall.Rlimit) error { } func SetParentDeathSignal(sig uintptr) error { - if _, _, err := syscall.RawSyscall(syscall.SYS_PRCTL, syscall.PR_SET_PDEATHSIG, sig, 0); err != 0 { + if err := unix.Prctl(unix.PR_SET_PDEATHSIG, sig, 0, 0, 0); err != nil { return err } return nil @@ -70,15 +71,14 @@ func SetParentDeathSignal(sig uintptr) error { func GetParentDeathSignal() (ParentDeathSignal, error) { var sig int - _, _, err := syscall.RawSyscall(syscall.SYS_PRCTL, syscall.PR_GET_PDEATHSIG, uintptr(unsafe.Pointer(&sig)), 0) - if err != 0 { + if err := unix.Prctl(unix.PR_GET_PDEATHSIG, uintptr(unsafe.Pointer(&sig)), 0, 0, 0); err != nil { return -1, err } return ParentDeathSignal(sig), nil } func SetKeepCaps() error { - if _, _, err := syscall.RawSyscall(syscall.SYS_PRCTL, syscall.PR_SET_KEEPCAPS, 1, 0); err != 0 { + if err := unix.Prctl(unix.PR_SET_KEEPCAPS, 1, 0, 0, 0); err != nil { return err } @@ -86,7 +86,7 @@ func SetKeepCaps() error { } func ClearKeepCaps() error { - if _, _, err := syscall.RawSyscall(syscall.SYS_PRCTL, syscall.PR_SET_KEEPCAPS, 0, 0); err != 0 { + if err := unix.Prctl(unix.PR_SET_KEEPCAPS, 0, 0, 0, 0); err != nil { return err } @@ -94,55 +94,62 @@ func ClearKeepCaps() error { } func Setctty() error { - if _, _, err := syscall.RawSyscall(syscall.SYS_IOCTL, 0, uintptr(syscall.TIOCSCTTY), 0); err != 0 { + if err := unix.IoctlSetInt(0, unix.TIOCSCTTY, 0); err != nil { return err } return nil } -/* - * Detect whether we are currently running in a user namespace. - * Copied from github.com/lxc/lxd/shared/util.go - */ +// RunningInUserNS detects whether we are currently running in a user namespace. +// Originally copied from github.com/lxc/lxd/shared/util.go func RunningInUserNS() bool { - file, err := os.Open("/proc/self/uid_map") + uidmap, err := user.CurrentProcessUIDMap() if err != nil { - /* - * This kernel-provided file only exists if user namespaces are - * supported - */ + // This kernel-provided file only exists if user namespaces are supported return false } - defer file.Close() + return UIDMapInUserNS(uidmap) +} - buf := bufio.NewReader(file) - l, _, err := buf.ReadLine() - if err != nil { - return false - } - - line := string(l) - var a, b, c int64 - fmt.Sscanf(line, "%d %d %d", &a, &b, &c) +func UIDMapInUserNS(uidmap []user.IDMap) bool { /* * We assume we are in the initial user namespace if we have a full * range - 4294967295 uids starting at uid 0. */ - if a == 0 && b == 0 && c == 4294967295 { + if len(uidmap) == 1 && uidmap[0].ID == 0 && uidmap[0].ParentID == 0 && uidmap[0].Count == 4294967295 { return false } return true } -// SetSubreaper sets the value i as the subreaper setting for the calling process -func SetSubreaper(i int) error { - return Prctl(PR_SET_CHILD_SUBREAPER, uintptr(i), 0, 0, 0) +// GetParentNSeuid returns the euid within the parent user namespace +func GetParentNSeuid() int64 { + euid := int64(os.Geteuid()) + uidmap, err := user.CurrentProcessUIDMap() + if err != nil { + // This kernel-provided file only exists if user namespaces are supported + return euid + } + for _, um := range uidmap { + if um.ID <= euid && euid <= um.ID+um.Count-1 { + return um.ParentID + euid - um.ID + } + } + return euid } -func Prctl(option int, arg2, arg3, arg4, arg5 uintptr) (err error) { - _, _, e1 := syscall.Syscall6(syscall.SYS_PRCTL, uintptr(option), arg2, arg3, arg4, arg5, 0) - if e1 != 0 { - err = e1 - } - return +// SetSubreaper sets the value i as the subreaper setting for the calling process +func SetSubreaper(i int) error { + return unix.Prctl(PR_SET_CHILD_SUBREAPER, uintptr(i), 0, 0, 0) +} + +// GetSubreaper returns the subreaper setting for the calling process +func GetSubreaper() (int, error) { + var i uintptr + + if err := unix.Prctl(unix.PR_GET_CHILD_SUBREAPER, uintptr(unsafe.Pointer(&i)), 0, 0, 0); err != nil { + return -1, err + } + + return int(i), nil } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/system/proc.go b/vendor/github.com/opencontainers/runc/libcontainer/system/proc.go index 37808a29f..79232a437 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/system/proc.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/system/proc.go @@ -1,27 +1,113 @@ package system import ( + "fmt" "io/ioutil" "path/filepath" "strconv" "strings" ) -// look in /proc to find the process start time so that we can verify -// that this pid has started after ourself +// State is the status of a process. +type State rune + +const ( // Only values for Linux 3.14 and later are listed here + Dead State = 'X' + DiskSleep State = 'D' + Running State = 'R' + Sleeping State = 'S' + Stopped State = 'T' + TracingStop State = 't' + Zombie State = 'Z' +) + +// String forms of the state from proc(5)'s documentation for +// /proc/[pid]/status' "State" field. +func (s State) String() string { + switch s { + case Dead: + return "dead" + case DiskSleep: + return "disk sleep" + case Running: + return "running" + case Sleeping: + return "sleeping" + case Stopped: + return "stopped" + case TracingStop: + return "tracing stop" + case Zombie: + return "zombie" + default: + return fmt.Sprintf("unknown (%c)", s) + } +} + +// Stat_t represents the information from /proc/[pid]/stat, as +// described in proc(5) with names based on the /proc/[pid]/status +// fields. +type Stat_t struct { + // PID is the process ID. + PID uint + + // Name is the command run by the process. + Name string + + // State is the state of the process. + State State + + // StartTime is the number of clock ticks after system boot (since + // Linux 2.6). + StartTime uint64 +} + +// Stat returns a Stat_t instance for the specified process. +func Stat(pid int) (stat Stat_t, err error) { + bytes, err := ioutil.ReadFile(filepath.Join("/proc", strconv.Itoa(pid), "stat")) + if err != nil { + return stat, err + } + return parseStat(string(bytes)) +} + +// GetProcessStartTime is deprecated. Use Stat(pid) and +// Stat_t.StartTime instead. func GetProcessStartTime(pid int) (string, error) { - data, err := ioutil.ReadFile(filepath.Join("/proc", strconv.Itoa(pid), "stat")) + stat, err := Stat(pid) if err != nil { return "", err } - - parts := strings.Split(string(data), " ") - // the starttime is located at pos 22 - // from the man page - // - // starttime %llu (was %lu before Linux 2.6) - // (22) The time the process started after system boot. In kernels before Linux 2.6, this - // value was expressed in jiffies. Since Linux 2.6, the value is expressed in clock ticks - // (divide by sysconf(_SC_CLK_TCK)). - return parts[22-1], nil // starts at 1 + return fmt.Sprintf("%d", stat.StartTime), nil +} + +func parseStat(data string) (stat Stat_t, err error) { + // From proc(5), field 2 could contain space and is inside `(` and `)`. + // The following is an example: + // 89653 (gunicorn: maste) S 89630 89653 89653 0 -1 4194560 29689 28896 0 3 146 32 76 19 20 0 1 0 2971844 52965376 3920 18446744073709551615 1 1 0 0 0 0 0 16781312 137447943 0 0 0 17 1 0 0 0 0 0 0 0 0 0 0 0 0 0 + i := strings.LastIndex(data, ")") + if i <= 2 || i >= len(data)-1 { + return stat, fmt.Errorf("invalid stat data: %q", data) + } + + parts := strings.SplitN(data[:i], "(", 2) + if len(parts) != 2 { + return stat, fmt.Errorf("invalid stat data: %q", data) + } + + stat.Name = parts[1] + _, err = fmt.Sscanf(parts[0], "%d", &stat.PID) + if err != nil { + return stat, err + } + + // parts indexes should be offset by 3 from the field number given + // proc(5), because parts is zero-indexed and we've removed fields + // one (PID) and two (Name) in the paren-split. + parts = strings.Split(data[i+2:], " ") + var state int + fmt.Sscanf(parts[3-3], "%c", &state) + stat.State = State(state) + fmt.Sscanf(parts[22-3], "%d", &stat.StartTime) + return stat, nil } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/system/setns_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/system/setns_linux.go deleted file mode 100644 index 615ff4c82..000000000 --- a/vendor/github.com/opencontainers/runc/libcontainer/system/setns_linux.go +++ /dev/null @@ -1,40 +0,0 @@ -package system - -import ( - "fmt" - "runtime" - "syscall" -) - -// Via http://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git/commit/?id=7b21fddd087678a70ad64afc0f632e0f1071b092 -// -// We need different setns values for the different platforms and arch -// We are declaring the macro here because the SETNS syscall does not exist in th stdlib -var setNsMap = map[string]uintptr{ - "linux/386": 346, - "linux/arm64": 268, - "linux/amd64": 308, - "linux/arm": 375, - "linux/ppc": 350, - "linux/ppc64": 350, - "linux/ppc64le": 350, - "linux/s390x": 339, -} - -var sysSetns = setNsMap[fmt.Sprintf("%s/%s", runtime.GOOS, runtime.GOARCH)] - -func SysSetns() uint32 { - return uint32(sysSetns) -} - -func Setns(fd uintptr, flags uintptr) error { - ns, exists := setNsMap[fmt.Sprintf("%s/%s", runtime.GOOS, runtime.GOARCH)] - if !exists { - return fmt.Errorf("unsupported platform %s/%s", runtime.GOOS, runtime.GOARCH) - } - _, _, err := syscall.RawSyscall(ns, fd, flags, 0) - if err != 0 { - return err - } - return nil -} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_arm.go b/vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_32.go similarity index 61% rename from vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_arm.go rename to vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_32.go index 3f780f312..c5ca5d862 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_arm.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_32.go @@ -1,14 +1,15 @@ -// +build linux,arm +// +build linux +// +build 386 arm package system import ( - "syscall" + "golang.org/x/sys/unix" ) // Setuid sets the uid of the calling thread to the specified uid. func Setuid(uid int) (err error) { - _, _, e1 := syscall.RawSyscall(syscall.SYS_SETUID32, uintptr(uid), 0, 0) + _, _, e1 := unix.RawSyscall(unix.SYS_SETUID32, uintptr(uid), 0, 0) if e1 != 0 { err = e1 } @@ -17,7 +18,7 @@ func Setuid(uid int) (err error) { // Setgid sets the gid of the calling thread to the specified gid. func Setgid(gid int) (err error) { - _, _, e1 := syscall.RawSyscall(syscall.SYS_SETGID32, uintptr(gid), 0, 0) + _, _, e1 := unix.RawSyscall(unix.SYS_SETGID32, uintptr(gid), 0, 0) if e1 != 0 { err = e1 } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_386.go b/vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_386.go deleted file mode 100644 index c99006518..000000000 --- a/vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_386.go +++ /dev/null @@ -1,25 +0,0 @@ -// +build linux,386 - -package system - -import ( - "syscall" -) - -// Setuid sets the uid of the calling thread to the specified uid. -func Setuid(uid int) (err error) { - _, _, e1 := syscall.RawSyscall(syscall.SYS_SETUID, uintptr(uid), 0, 0) - if e1 != 0 { - err = e1 - } - return -} - -// Setgid sets the gid of the calling thread to the specified gid. -func Setgid(gid int) (err error) { - _, _, e1 := syscall.RawSyscall(syscall.SYS_SETGID32, uintptr(gid), 0, 0) - if e1 != 0 { - err = e1 - } - return -} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_64.go b/vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_64.go index 0816bf828..11c3faafb 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_64.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/system/syscall_linux_64.go @@ -1,14 +1,15 @@ -// +build linux,arm64 linux,amd64 linux,ppc linux,ppc64 linux,ppc64le linux,s390x +// +build linux +// +build arm64 amd64 mips mipsle mips64 mips64le ppc ppc64 ppc64le s390x package system import ( - "syscall" + "golang.org/x/sys/unix" ) // Setuid sets the uid of the calling thread to the specified uid. func Setuid(uid int) (err error) { - _, _, e1 := syscall.RawSyscall(syscall.SYS_SETUID, uintptr(uid), 0, 0) + _, _, e1 := unix.RawSyscall(unix.SYS_SETUID, uintptr(uid), 0, 0) if e1 != 0 { err = e1 } @@ -17,7 +18,7 @@ func Setuid(uid int) (err error) { // Setgid sets the gid of the calling thread to the specified gid. func Setgid(gid int) (err error) { - _, _, e1 := syscall.RawSyscall(syscall.SYS_SETGID, uintptr(gid), 0, 0) + _, _, e1 := unix.RawSyscall(unix.SYS_SETGID, uintptr(gid), 0, 0) if e1 != 0 { err = e1 } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/system/sysconfig.go b/vendor/github.com/opencontainers/runc/libcontainer/system/sysconfig.go index b3a07cba3..b8434f105 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/system/sysconfig.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/system/sysconfig.go @@ -1,4 +1,4 @@ -// +build cgo,linux cgo,freebsd +// +build cgo,linux package system diff --git a/vendor/github.com/opencontainers/runc/libcontainer/system/unsupported.go b/vendor/github.com/opencontainers/runc/libcontainer/system/unsupported.go index e7cfd62b2..b94be74a6 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/system/unsupported.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/system/unsupported.go @@ -2,8 +2,26 @@ package system +import ( + "os" + + "github.com/opencontainers/runc/libcontainer/user" +) + // RunningInUserNS is a stub for non-Linux systems // Always returns false func RunningInUserNS() bool { return false } + +// UIDMapInUserNS is a stub for non-Linux systems +// Always returns false +func UIDMapInUserNS(uidmap []user.IDMap) bool { + return false +} + +// GetParentNSeuid returns the euid within the parent user namespace +// Always returns os.Geteuid on non-linux +func GetParentNSeuid() int { + return os.Geteuid() +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/system/xattrs_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/system/xattrs_linux.go index 30f74dfb1..a6823fc99 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/system/xattrs_linux.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/system/xattrs_linux.go @@ -1,99 +1,35 @@ package system -import ( - "syscall" - "unsafe" -) - -var _zero uintptr - -// Returns the size of xattrs and nil error -// Requires path, takes allocated []byte or nil as last argument -func Llistxattr(path string, dest []byte) (size int, err error) { - pathBytes, err := syscall.BytePtrFromString(path) - if err != nil { - return -1, err - } - var newpathBytes unsafe.Pointer - if len(dest) > 0 { - newpathBytes = unsafe.Pointer(&dest[0]) - } else { - newpathBytes = unsafe.Pointer(&_zero) - } - - _size, _, errno := syscall.Syscall6(syscall.SYS_LLISTXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(newpathBytes), uintptr(len(dest)), 0, 0, 0) - size = int(_size) - if errno != 0 { - return -1, errno - } - - return size, nil -} +import "golang.org/x/sys/unix" // Returns a []byte slice if the xattr is set and nil otherwise // Requires path and its attribute as arguments func Lgetxattr(path string, attr string) ([]byte, error) { var sz int - pathBytes, err := syscall.BytePtrFromString(path) - if err != nil { - return nil, err - } - attrBytes, err := syscall.BytePtrFromString(attr) - if err != nil { - return nil, err - } - // Start with a 128 length byte array - sz = 128 - dest := make([]byte, sz) - destBytes := unsafe.Pointer(&dest[0]) - _sz, _, errno := syscall.Syscall6(syscall.SYS_LGETXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(attrBytes)), uintptr(destBytes), uintptr(len(dest)), 0, 0) + dest := make([]byte, 128) + sz, errno := unix.Lgetxattr(path, attr, dest) switch { - case errno == syscall.ENODATA: + case errno == unix.ENODATA: return nil, errno - case errno == syscall.ENOTSUP: + case errno == unix.ENOTSUP: return nil, errno - case errno == syscall.ERANGE: + case errno == unix.ERANGE: // 128 byte array might just not be good enough, - // A dummy buffer is used ``uintptr(0)`` to get real size + // A dummy buffer is used to get the real size // of the xattrs on disk - _sz, _, errno = syscall.Syscall6(syscall.SYS_LGETXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(attrBytes)), uintptr(unsafe.Pointer(nil)), uintptr(0), 0, 0) - sz = int(_sz) - if sz < 0 { + sz, errno = unix.Lgetxattr(path, attr, []byte{}) + if errno != nil { return nil, errno } dest = make([]byte, sz) - destBytes := unsafe.Pointer(&dest[0]) - _sz, _, errno = syscall.Syscall6(syscall.SYS_LGETXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(attrBytes)), uintptr(destBytes), uintptr(len(dest)), 0, 0) - if errno != 0 { + sz, errno = unix.Lgetxattr(path, attr, dest) + if errno != nil { return nil, errno } - case errno != 0: + case errno != nil: return nil, errno } - sz = int(_sz) return dest[:sz], nil } - -func Lsetxattr(path string, attr string, data []byte, flags int) error { - pathBytes, err := syscall.BytePtrFromString(path) - if err != nil { - return err - } - attrBytes, err := syscall.BytePtrFromString(attr) - if err != nil { - return err - } - var dataBytes unsafe.Pointer - if len(data) > 0 { - dataBytes = unsafe.Pointer(&data[0]) - } else { - dataBytes = unsafe.Pointer(&_zero) - } - _, _, errno := syscall.Syscall6(syscall.SYS_LSETXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(attrBytes)), uintptr(dataBytes), uintptr(len(data)), uintptr(flags), 0) - if errno != 0 { - return errno - } - return nil -} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/user/lookup.go b/vendor/github.com/opencontainers/runc/libcontainer/user/lookup.go index ab1439f36..6fd8dd0d4 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/user/lookup.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/user/lookup.go @@ -2,7 +2,6 @@ package user import ( "errors" - "syscall" ) var ( @@ -13,98 +12,30 @@ var ( ErrNoGroupEntries = errors.New("no matching entries in group file") ) -func lookupUser(filter func(u User) bool) (User, error) { - // Get operating system-specific passwd reader-closer. - passwd, err := GetPasswd() - if err != nil { - return User{}, err - } - defer passwd.Close() - - // Get the users. - users, err := ParsePasswdFilter(passwd, filter) - if err != nil { - return User{}, err - } - - // No user entries found. - if len(users) == 0 { - return User{}, ErrNoPasswdEntries - } - - // Assume the first entry is the "correct" one. - return users[0], nil -} - -// CurrentUser looks up the current user by their user id in /etc/passwd. If the -// user cannot be found (or there is no /etc/passwd file on the filesystem), -// then CurrentUser returns an error. -func CurrentUser() (User, error) { - return LookupUid(syscall.Getuid()) -} - // LookupUser looks up a user by their username in /etc/passwd. If the user // cannot be found (or there is no /etc/passwd file on the filesystem), then // LookupUser returns an error. func LookupUser(username string) (User, error) { - return lookupUser(func(u User) bool { - return u.Name == username - }) + return lookupUser(username) } // LookupUid looks up a user by their user id in /etc/passwd. If the user cannot // be found (or there is no /etc/passwd file on the filesystem), then LookupId // returns an error. func LookupUid(uid int) (User, error) { - return lookupUser(func(u User) bool { - return u.Uid == uid - }) -} - -func lookupGroup(filter func(g Group) bool) (Group, error) { - // Get operating system-specific group reader-closer. - group, err := GetGroup() - if err != nil { - return Group{}, err - } - defer group.Close() - - // Get the users. - groups, err := ParseGroupFilter(group, filter) - if err != nil { - return Group{}, err - } - - // No user entries found. - if len(groups) == 0 { - return Group{}, ErrNoGroupEntries - } - - // Assume the first entry is the "correct" one. - return groups[0], nil -} - -// CurrentGroup looks up the current user's group by their primary group id's -// entry in /etc/passwd. If the group cannot be found (or there is no -// /etc/group file on the filesystem), then CurrentGroup returns an error. -func CurrentGroup() (Group, error) { - return LookupGid(syscall.Getgid()) + return lookupUid(uid) } // LookupGroup looks up a group by its name in /etc/group. If the group cannot // be found (or there is no /etc/group file on the filesystem), then LookupGroup // returns an error. func LookupGroup(groupname string) (Group, error) { - return lookupGroup(func(g Group) bool { - return g.Name == groupname - }) + return lookupGroup(groupname) } // LookupGid looks up a group by its group id in /etc/group. If the group cannot // be found (or there is no /etc/group file on the filesystem), then LookupGid // returns an error. func LookupGid(gid int) (Group, error) { - return lookupGroup(func(g Group) bool { - return g.Gid == gid - }) + return lookupGid(gid) } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/user/lookup_unix.go b/vendor/github.com/opencontainers/runc/libcontainer/user/lookup_unix.go index 758b734c2..c1e634c94 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/user/lookup_unix.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/user/lookup_unix.go @@ -5,6 +5,8 @@ package user import ( "io" "os" + + "golang.org/x/sys/unix" ) // Unix-specific path to the passwd and group formatted files. @@ -13,6 +15,76 @@ const ( unixGroupPath = "/etc/group" ) +func lookupUser(username string) (User, error) { + return lookupUserFunc(func(u User) bool { + return u.Name == username + }) +} + +func lookupUid(uid int) (User, error) { + return lookupUserFunc(func(u User) bool { + return u.Uid == uid + }) +} + +func lookupUserFunc(filter func(u User) bool) (User, error) { + // Get operating system-specific passwd reader-closer. + passwd, err := GetPasswd() + if err != nil { + return User{}, err + } + defer passwd.Close() + + // Get the users. + users, err := ParsePasswdFilter(passwd, filter) + if err != nil { + return User{}, err + } + + // No user entries found. + if len(users) == 0 { + return User{}, ErrNoPasswdEntries + } + + // Assume the first entry is the "correct" one. + return users[0], nil +} + +func lookupGroup(groupname string) (Group, error) { + return lookupGroupFunc(func(g Group) bool { + return g.Name == groupname + }) +} + +func lookupGid(gid int) (Group, error) { + return lookupGroupFunc(func(g Group) bool { + return g.Gid == gid + }) +} + +func lookupGroupFunc(filter func(g Group) bool) (Group, error) { + // Get operating system-specific group reader-closer. + group, err := GetGroup() + if err != nil { + return Group{}, err + } + defer group.Close() + + // Get the users. + groups, err := ParseGroupFilter(group, filter) + if err != nil { + return Group{}, err + } + + // No user entries found. + if len(groups) == 0 { + return Group{}, ErrNoGroupEntries + } + + // Assume the first entry is the "correct" one. + return groups[0], nil +} + func GetPasswdPath() (string, error) { return unixPasswdPath, nil } @@ -28,3 +100,43 @@ func GetGroupPath() (string, error) { func GetGroup() (io.ReadCloser, error) { return os.Open(unixGroupPath) } + +// CurrentUser looks up the current user by their user id in /etc/passwd. If the +// user cannot be found (or there is no /etc/passwd file on the filesystem), +// then CurrentUser returns an error. +func CurrentUser() (User, error) { + return LookupUid(unix.Getuid()) +} + +// CurrentGroup looks up the current user's group by their primary group id's +// entry in /etc/passwd. If the group cannot be found (or there is no +// /etc/group file on the filesystem), then CurrentGroup returns an error. +func CurrentGroup() (Group, error) { + return LookupGid(unix.Getgid()) +} + +func CurrentUserSubUIDs() ([]SubID, error) { + u, err := CurrentUser() + if err != nil { + return nil, err + } + return ParseSubIDFileFilter("/etc/subuid", + func(entry SubID) bool { return entry.Name == u.Name }) +} + +func CurrentGroupSubGIDs() ([]SubID, error) { + g, err := CurrentGroup() + if err != nil { + return nil, err + } + return ParseSubIDFileFilter("/etc/subgid", + func(entry SubID) bool { return entry.Name == g.Name }) +} + +func CurrentProcessUIDMap() ([]IDMap, error) { + return ParseIDMapFile("/proc/self/uid_map") +} + +func CurrentProcessGIDMap() ([]IDMap, error) { + return ParseIDMapFile("/proc/self/gid_map") +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/user/lookup_unsupported.go b/vendor/github.com/opencontainers/runc/libcontainer/user/lookup_unsupported.go deleted file mode 100644 index 721794887..000000000 --- a/vendor/github.com/opencontainers/runc/libcontainer/user/lookup_unsupported.go +++ /dev/null @@ -1,21 +0,0 @@ -// +build !darwin,!dragonfly,!freebsd,!linux,!netbsd,!openbsd,!solaris - -package user - -import "io" - -func GetPasswdPath() (string, error) { - return "", ErrUnsupported -} - -func GetPasswd() (io.ReadCloser, error) { - return nil, ErrUnsupported -} - -func GetGroupPath() (string, error) { - return "", ErrUnsupported -} - -func GetGroup() (io.ReadCloser, error) { - return nil, ErrUnsupported -} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/user/lookup_windows.go b/vendor/github.com/opencontainers/runc/libcontainer/user/lookup_windows.go new file mode 100644 index 000000000..65cd40e92 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/user/lookup_windows.go @@ -0,0 +1,40 @@ +// +build windows + +package user + +import ( + "fmt" + "os/user" +) + +func lookupUser(username string) (User, error) { + u, err := user.Lookup(username) + if err != nil { + return User{}, err + } + return userFromOS(u) +} + +func lookupUid(uid int) (User, error) { + u, err := user.LookupId(fmt.Sprintf("%d", uid)) + if err != nil { + return User{}, err + } + return userFromOS(u) +} + +func lookupGroup(groupname string) (Group, error) { + g, err := user.LookupGroup(groupname) + if err != nil { + return Group{}, err + } + return groupFromOS(g) +} + +func lookupGid(gid int) (Group, error) { + g, err := user.LookupGroupId(fmt.Sprintf("%d", gid)) + if err != nil { + return Group{}, err + } + return groupFromOS(g) +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/user/user.go b/vendor/github.com/opencontainers/runc/libcontainer/user/user.go index 43fd39ef5..7b912bbf8 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/user/user.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/user/user.go @@ -5,6 +5,7 @@ import ( "fmt" "io" "os" + "os/user" "strconv" "strings" ) @@ -28,6 +29,28 @@ type User struct { Shell string } +// userFromOS converts an os/user.(*User) to local User +// +// (This does not include Pass, Shell or Gecos) +func userFromOS(u *user.User) (User, error) { + newUser := User{ + Name: u.Username, + Home: u.HomeDir, + } + id, err := strconv.Atoi(u.Uid) + if err != nil { + return newUser, err + } + newUser.Uid = id + + id, err = strconv.Atoi(u.Gid) + if err != nil { + return newUser, err + } + newUser.Gid = id + return newUser, nil +} + type Group struct { Name string Pass string @@ -35,12 +58,46 @@ type Group struct { List []string } +// groupFromOS converts an os/user.(*Group) to local Group +// +// (This does not include Pass, Shell or Gecos) +func groupFromOS(g *user.Group) (Group, error) { + newGroup := Group{ + Name: g.Name, + } + + id, err := strconv.Atoi(g.Gid) + if err != nil { + return newGroup, err + } + newGroup.Gid = id + + return newGroup, nil +} + +// SubID represents an entry in /etc/sub{u,g}id +type SubID struct { + Name string + SubID int64 + Count int64 +} + +// IDMap represents an entry in /proc/PID/{u,g}id_map +type IDMap struct { + ID int64 + ParentID int64 + Count int64 +} + func parseLine(line string, v ...interface{}) { - if line == "" { + parseParts(strings.Split(line, ":"), v...) +} + +func parseParts(parts []string, v ...interface{}) { + if len(parts) == 0 { return } - parts := strings.Split(line, ":") for i, p := range parts { // Ignore cases where we don't have enough fields to populate the arguments. // Some configuration files like to misbehave. @@ -56,6 +113,8 @@ func parseLine(line string, v ...interface{}) { case *int: // "numbers", with conversion errors ignored because of some misbehaving configuration files. *e, _ = strconv.Atoi(p) + case *int64: + *e, _ = strconv.ParseInt(p, 10, 64) case *[]string: // Comma-separated lists. if p != "" { @@ -65,7 +124,7 @@ func parseLine(line string, v ...interface{}) { } default: // Someone goof'd when writing code using this function. Scream so they can hear us. - panic(fmt.Sprintf("parseLine only accepts {*string, *int, *[]string} as arguments! %#v is not a pointer!", e)) + panic(fmt.Sprintf("parseLine only accepts {*string, *int, *int64, *[]string} as arguments! %#v is not a pointer!", e)) } } } @@ -199,18 +258,16 @@ type ExecUser struct { // files cannot be opened for any reason, the error is ignored and a nil // io.Reader is passed instead. func GetExecUserPath(userSpec string, defaults *ExecUser, passwdPath, groupPath string) (*ExecUser, error) { - passwd, err := os.Open(passwdPath) - if err != nil { - passwd = nil - } else { - defer passwd.Close() + var passwd, group io.Reader + + if passwdFile, err := os.Open(passwdPath); err == nil { + passwd = passwdFile + defer passwdFile.Close() } - group, err := os.Open(groupPath) - if err != nil { - group = nil - } else { - defer group.Close() + if groupFile, err := os.Open(groupPath); err == nil { + group = groupFile + defer groupFile.Close() } return GetExecUser(userSpec, defaults, passwd, group) @@ -343,7 +400,7 @@ func GetExecUser(userSpec string, defaults *ExecUser, passwd, group io.Reader) ( if len(groups) > 0 { // First match wins, even if there's more than one matching entry. user.Gid = groups[0].Gid - } else if groupArg != "" { + } else { // If we can't find a group with the given name, the only other valid // option is if it's a numeric group name with no associated entry in group. @@ -433,9 +490,119 @@ func GetAdditionalGroups(additionalGroups []string, group io.Reader) ([]int, err // that opens the groupPath given and gives it as an argument to // GetAdditionalGroups. func GetAdditionalGroupsPath(additionalGroups []string, groupPath string) ([]int, error) { - group, err := os.Open(groupPath) - if err == nil { - defer group.Close() + var group io.Reader + + if groupFile, err := os.Open(groupPath); err == nil { + group = groupFile + defer groupFile.Close() } return GetAdditionalGroups(additionalGroups, group) } + +func ParseSubIDFile(path string) ([]SubID, error) { + subid, err := os.Open(path) + if err != nil { + return nil, err + } + defer subid.Close() + return ParseSubID(subid) +} + +func ParseSubID(subid io.Reader) ([]SubID, error) { + return ParseSubIDFilter(subid, nil) +} + +func ParseSubIDFileFilter(path string, filter func(SubID) bool) ([]SubID, error) { + subid, err := os.Open(path) + if err != nil { + return nil, err + } + defer subid.Close() + return ParseSubIDFilter(subid, filter) +} + +func ParseSubIDFilter(r io.Reader, filter func(SubID) bool) ([]SubID, error) { + if r == nil { + return nil, fmt.Errorf("nil source for subid-formatted data") + } + + var ( + s = bufio.NewScanner(r) + out = []SubID{} + ) + + for s.Scan() { + if err := s.Err(); err != nil { + return nil, err + } + + line := strings.TrimSpace(s.Text()) + if line == "" { + continue + } + + // see: man 5 subuid + p := SubID{} + parseLine(line, &p.Name, &p.SubID, &p.Count) + + if filter == nil || filter(p) { + out = append(out, p) + } + } + + return out, nil +} + +func ParseIDMapFile(path string) ([]IDMap, error) { + r, err := os.Open(path) + if err != nil { + return nil, err + } + defer r.Close() + return ParseIDMap(r) +} + +func ParseIDMap(r io.Reader) ([]IDMap, error) { + return ParseIDMapFilter(r, nil) +} + +func ParseIDMapFileFilter(path string, filter func(IDMap) bool) ([]IDMap, error) { + r, err := os.Open(path) + if err != nil { + return nil, err + } + defer r.Close() + return ParseIDMapFilter(r, filter) +} + +func ParseIDMapFilter(r io.Reader, filter func(IDMap) bool) ([]IDMap, error) { + if r == nil { + return nil, fmt.Errorf("nil source for idmap-formatted data") + } + + var ( + s = bufio.NewScanner(r) + out = []IDMap{} + ) + + for s.Scan() { + if err := s.Err(); err != nil { + return nil, err + } + + line := strings.TrimSpace(s.Text()) + if line == "" { + continue + } + + // see: man 7 user_namespaces + p := IDMap{} + parseParts(strings.Fields(line), &p.ID, &p.ParentID, &p.Count) + + if filter == nil || filter(p) { + out = append(out, p) + } + } + + return out, nil +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/utils/cmsg.go b/vendor/github.com/opencontainers/runc/libcontainer/utils/cmsg.go new file mode 100644 index 000000000..c8a9364d5 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/utils/cmsg.go @@ -0,0 +1,93 @@ +// +build linux + +package utils + +/* + * Copyright 2016, 2017 SUSE LLC + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +import ( + "fmt" + "os" + + "golang.org/x/sys/unix" +) + +// MaxSendfdLen is the maximum length of the name of a file descriptor being +// sent using SendFd. The name of the file handle returned by RecvFd will never +// be larger than this value. +const MaxNameLen = 4096 + +// oobSpace is the size of the oob slice required to store a single FD. Note +// that unix.UnixRights appears to make the assumption that fd is always int32, +// so sizeof(fd) = 4. +var oobSpace = unix.CmsgSpace(4) + +// RecvFd waits for a file descriptor to be sent over the given AF_UNIX +// socket. The file name of the remote file descriptor will be recreated +// locally (it is sent as non-auxiliary data in the same payload). +func RecvFd(socket *os.File) (*os.File, error) { + // For some reason, unix.Recvmsg uses the length rather than the capacity + // when passing the msg_controllen and other attributes to recvmsg. So we + // have to actually set the length. + name := make([]byte, MaxNameLen) + oob := make([]byte, oobSpace) + + sockfd := socket.Fd() + n, oobn, _, _, err := unix.Recvmsg(int(sockfd), name, oob, 0) + if err != nil { + return nil, err + } + + if n >= MaxNameLen || oobn != oobSpace { + return nil, fmt.Errorf("recvfd: incorrect number of bytes read (n=%d oobn=%d)", n, oobn) + } + + // Truncate. + name = name[:n] + oob = oob[:oobn] + + scms, err := unix.ParseSocketControlMessage(oob) + if err != nil { + return nil, err + } + if len(scms) != 1 { + return nil, fmt.Errorf("recvfd: number of SCMs is not 1: %d", len(scms)) + } + scm := scms[0] + + fds, err := unix.ParseUnixRights(&scm) + if err != nil { + return nil, err + } + if len(fds) != 1 { + return nil, fmt.Errorf("recvfd: number of fds is not 1: %d", len(fds)) + } + fd := uintptr(fds[0]) + + return os.NewFile(fd, string(name)), nil +} + +// SendFd sends a file descriptor over the given AF_UNIX socket. In +// addition, the file.Name() of the given file will also be sent as +// non-auxiliary data in the same payload (allowing to send contextual +// information for a file descriptor). +func SendFd(socket *os.File, name string, fd uintptr) error { + if len(name) >= MaxNameLen { + return fmt.Errorf("sendfd: filename too long: %s", name) + } + oob := unix.UnixRights(int(fd)) + return unix.Sendmsg(int(socket.Fd()), []byte(name), oob, nil, 0) +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/utils/utils.go b/vendor/github.com/opencontainers/runc/libcontainer/utils/utils.go index 68ae3c477..baa54c9ba 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/utils/utils.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/utils/utils.go @@ -7,7 +7,10 @@ import ( "io" "os" "path/filepath" - "syscall" + "strings" + "unsafe" + + "golang.org/x/sys/unix" ) const ( @@ -39,7 +42,7 @@ func ResolveRootfs(uncleanRootfs string) (string, error) { // ExitStatus returns the correct exit status for a process based on if it // was signaled or exited cleanly -func ExitStatus(status syscall.WaitStatus) int { +func ExitStatus(status unix.WaitStatus) int { if status.Signaled() { return exitSignalOffset + int(status.Signal()) } @@ -84,3 +87,41 @@ func CleanPath(path string) string { // Clean the path again for good measure. return filepath.Clean(path) } + +// SearchLabels searches a list of key-value pairs for the provided key and +// returns the corresponding value. The pairs must be separated with '='. +func SearchLabels(labels []string, query string) string { + for _, l := range labels { + parts := strings.SplitN(l, "=", 2) + if len(parts) < 2 { + continue + } + if parts[0] == query { + return parts[1] + } + } + return "" +} + +// Annotations returns the bundle path and user defined annotations from the +// libcontainer state. We need to remove the bundle because that is a label +// added by libcontainer. +func Annotations(labels []string) (bundle string, userAnnotations map[string]string) { + userAnnotations = make(map[string]string) + for _, l := range labels { + parts := strings.SplitN(l, "=", 2) + if len(parts) < 2 { + continue + } + if parts[0] == "bundle" { + bundle = parts[1] + } else { + userAnnotations[parts[0]] = parts[1] + } + } + return +} + +func GetIntSize() int { + return int(unsafe.Sizeof(1)) +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/utils/utils_unix.go b/vendor/github.com/opencontainers/runc/libcontainer/utils/utils_unix.go index 408918f27..c96088988 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/utils/utils_unix.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/utils/utils_unix.go @@ -4,8 +4,10 @@ package utils import ( "io/ioutil" + "os" "strconv" - "syscall" + + "golang.org/x/sys/unix" ) func CloseExecFrom(minFd int) error { @@ -25,9 +27,18 @@ func CloseExecFrom(minFd int) error { continue } - // intentionally ignore errors from syscall.CloseOnExec - syscall.CloseOnExec(fd) + // intentionally ignore errors from unix.CloseOnExec + unix.CloseOnExec(fd) // the cases where this might fail are basically file descriptors that have already been closed (including and especially the one that was created when ioutil.ReadDir did the "opendir" syscall) } return nil } + +// NewSockPair returns a new unix socket pair +func NewSockPair(name string) (parent *os.File, child *os.File, err error) { + fds, err := unix.Socketpair(unix.AF_LOCAL, unix.SOCK_STREAM|unix.SOCK_CLOEXEC, 0) + if err != nil { + return nil, nil, err + } + return os.NewFile(uintptr(fds[1]), name+"-p"), os.NewFile(uintptr(fds[0]), name+"-c"), nil +} diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/opencontainers/specs/LICENSE b/vendor/github.com/opencontainers/runtime-spec/LICENSE similarity index 100% rename from vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/opencontainers/specs/LICENSE rename to vendor/github.com/opencontainers/runtime-spec/LICENSE diff --git a/vendor/github.com/opencontainers/runtime-spec/specs-go/config.go b/vendor/github.com/opencontainers/runtime-spec/specs-go/config.go new file mode 100644 index 000000000..f3f37d42d --- /dev/null +++ b/vendor/github.com/opencontainers/runtime-spec/specs-go/config.go @@ -0,0 +1,570 @@ +package specs + +import "os" + +// Spec is the base configuration for the container. +type Spec struct { + // Version of the Open Container Runtime Specification with which the bundle complies. + Version string `json:"ociVersion"` + // Process configures the container process. + Process *Process `json:"process,omitempty"` + // Root configures the container's root filesystem. + Root *Root `json:"root,omitempty"` + // Hostname configures the container's hostname. + Hostname string `json:"hostname,omitempty"` + // Mounts configures additional mounts (on top of Root). + Mounts []Mount `json:"mounts,omitempty"` + // Hooks configures callbacks for container lifecycle events. + Hooks *Hooks `json:"hooks,omitempty" platform:"linux,solaris"` + // Annotations contains arbitrary metadata for the container. + Annotations map[string]string `json:"annotations,omitempty"` + + // Linux is platform-specific configuration for Linux based containers. + Linux *Linux `json:"linux,omitempty" platform:"linux"` + // Solaris is platform-specific configuration for Solaris based containers. + Solaris *Solaris `json:"solaris,omitempty" platform:"solaris"` + // Windows is platform-specific configuration for Windows based containers. + Windows *Windows `json:"windows,omitempty" platform:"windows"` +} + +// Process contains information to start a specific application inside the container. +type Process struct { + // Terminal creates an interactive terminal for the container. + Terminal bool `json:"terminal,omitempty"` + // ConsoleSize specifies the size of the console. + ConsoleSize *Box `json:"consoleSize,omitempty"` + // User specifies user information for the process. + User User `json:"user"` + // Args specifies the binary and arguments for the application to execute. + Args []string `json:"args"` + // Env populates the process environment for the process. + Env []string `json:"env,omitempty"` + // Cwd is the current working directory for the process and must be + // relative to the container's root. + Cwd string `json:"cwd"` + // Capabilities are Linux capabilities that are kept for the process. + Capabilities *LinuxCapabilities `json:"capabilities,omitempty" platform:"linux"` + // Rlimits specifies rlimit options to apply to the process. + Rlimits []POSIXRlimit `json:"rlimits,omitempty" platform:"linux,solaris"` + // NoNewPrivileges controls whether additional privileges could be gained by processes in the container. + NoNewPrivileges bool `json:"noNewPrivileges,omitempty" platform:"linux"` + // ApparmorProfile specifies the apparmor profile for the container. + ApparmorProfile string `json:"apparmorProfile,omitempty" platform:"linux"` + // Specify an oom_score_adj for the container. + OOMScoreAdj *int `json:"oomScoreAdj,omitempty" platform:"linux"` + // SelinuxLabel specifies the selinux context that the container process is run as. + SelinuxLabel string `json:"selinuxLabel,omitempty" platform:"linux"` +} + +// LinuxCapabilities specifies the whitelist of capabilities that are kept for a process. +// http://man7.org/linux/man-pages/man7/capabilities.7.html +type LinuxCapabilities struct { + // Bounding is the set of capabilities checked by the kernel. + Bounding []string `json:"bounding,omitempty" platform:"linux"` + // Effective is the set of capabilities checked by the kernel. + Effective []string `json:"effective,omitempty" platform:"linux"` + // Inheritable is the capabilities preserved across execve. + Inheritable []string `json:"inheritable,omitempty" platform:"linux"` + // Permitted is the limiting superset for effective capabilities. + Permitted []string `json:"permitted,omitempty" platform:"linux"` + // Ambient is the ambient set of capabilities that are kept. + Ambient []string `json:"ambient,omitempty" platform:"linux"` +} + +// Box specifies dimensions of a rectangle. Used for specifying the size of a console. +type Box struct { + // Height is the vertical dimension of a box. + Height uint `json:"height"` + // Width is the horizontal dimension of a box. + Width uint `json:"width"` +} + +// User specifies specific user (and group) information for the container process. +type User struct { + // UID is the user id. + UID uint32 `json:"uid" platform:"linux,solaris"` + // GID is the group id. + GID uint32 `json:"gid" platform:"linux,solaris"` + // AdditionalGids are additional group ids set for the container's process. + AdditionalGids []uint32 `json:"additionalGids,omitempty" platform:"linux,solaris"` + // Username is the user name. + Username string `json:"username,omitempty" platform:"windows"` +} + +// Root contains information about the container's root filesystem on the host. +type Root struct { + // Path is the absolute path to the container's root filesystem. + Path string `json:"path"` + // Readonly makes the root filesystem for the container readonly before the process is executed. + Readonly bool `json:"readonly,omitempty"` +} + +// Mount specifies a mount for a container. +type Mount struct { + // Destination is the absolute path where the mount will be placed in the container. + Destination string `json:"destination"` + // Type specifies the mount kind. + Type string `json:"type,omitempty" platform:"linux,solaris"` + // Source specifies the source path of the mount. + Source string `json:"source,omitempty"` + // Options are fstab style mount options. + Options []string `json:"options,omitempty"` +} + +// Hook specifies a command that is run at a particular event in the lifecycle of a container +type Hook struct { + Path string `json:"path"` + Args []string `json:"args,omitempty"` + Env []string `json:"env,omitempty"` + Timeout *int `json:"timeout,omitempty"` +} + +// Hooks for container setup and teardown +type Hooks struct { + // Prestart is a list of hooks to be run before the container process is executed. + Prestart []Hook `json:"prestart,omitempty"` + // Poststart is a list of hooks to be run after the container process is started. + Poststart []Hook `json:"poststart,omitempty"` + // Poststop is a list of hooks to be run after the container process exits. + Poststop []Hook `json:"poststop,omitempty"` +} + +// Linux contains platform-specific configuration for Linux based containers. +type Linux struct { + // UIDMapping specifies user mappings for supporting user namespaces. + UIDMappings []LinuxIDMapping `json:"uidMappings,omitempty"` + // GIDMapping specifies group mappings for supporting user namespaces. + GIDMappings []LinuxIDMapping `json:"gidMappings,omitempty"` + // Sysctl are a set of key value pairs that are set for the container on start + Sysctl map[string]string `json:"sysctl,omitempty"` + // Resources contain cgroup information for handling resource constraints + // for the container + Resources *LinuxResources `json:"resources,omitempty"` + // CgroupsPath specifies the path to cgroups that are created and/or joined by the container. + // The path is expected to be relative to the cgroups mountpoint. + // If resources are specified, the cgroups at CgroupsPath will be updated based on resources. + CgroupsPath string `json:"cgroupsPath,omitempty"` + // Namespaces contains the namespaces that are created and/or joined by the container + Namespaces []LinuxNamespace `json:"namespaces,omitempty"` + // Devices are a list of device nodes that are created for the container + Devices []LinuxDevice `json:"devices,omitempty"` + // Seccomp specifies the seccomp security settings for the container. + Seccomp *LinuxSeccomp `json:"seccomp,omitempty"` + // RootfsPropagation is the rootfs mount propagation mode for the container. + RootfsPropagation string `json:"rootfsPropagation,omitempty"` + // MaskedPaths masks over the provided paths inside the container. + MaskedPaths []string `json:"maskedPaths,omitempty"` + // ReadonlyPaths sets the provided paths as RO inside the container. + ReadonlyPaths []string `json:"readonlyPaths,omitempty"` + // MountLabel specifies the selinux context for the mounts in the container. + MountLabel string `json:"mountLabel,omitempty"` + // IntelRdt contains Intel Resource Director Technology (RDT) information + // for handling resource constraints (e.g., L3 cache) for the container + IntelRdt *LinuxIntelRdt `json:"intelRdt,omitempty"` +} + +// LinuxNamespace is the configuration for a Linux namespace +type LinuxNamespace struct { + // Type is the type of namespace + Type LinuxNamespaceType `json:"type"` + // Path is a path to an existing namespace persisted on disk that can be joined + // and is of the same type + Path string `json:"path,omitempty"` +} + +// LinuxNamespaceType is one of the Linux namespaces +type LinuxNamespaceType string + +const ( + // PIDNamespace for isolating process IDs + PIDNamespace LinuxNamespaceType = "pid" + // NetworkNamespace for isolating network devices, stacks, ports, etc + NetworkNamespace = "network" + // MountNamespace for isolating mount points + MountNamespace = "mount" + // IPCNamespace for isolating System V IPC, POSIX message queues + IPCNamespace = "ipc" + // UTSNamespace for isolating hostname and NIS domain name + UTSNamespace = "uts" + // UserNamespace for isolating user and group IDs + UserNamespace = "user" + // CgroupNamespace for isolating cgroup hierarchies + CgroupNamespace = "cgroup" +) + +// LinuxIDMapping specifies UID/GID mappings +type LinuxIDMapping struct { + // HostID is the starting UID/GID on the host to be mapped to 'ContainerID' + HostID uint32 `json:"hostID"` + // ContainerID is the starting UID/GID in the container + ContainerID uint32 `json:"containerID"` + // Size is the number of IDs to be mapped + Size uint32 `json:"size"` +} + +// POSIXRlimit type and restrictions +type POSIXRlimit struct { + // Type of the rlimit to set + Type string `json:"type"` + // Hard is the hard limit for the specified type + Hard uint64 `json:"hard"` + // Soft is the soft limit for the specified type + Soft uint64 `json:"soft"` +} + +// LinuxHugepageLimit structure corresponds to limiting kernel hugepages +type LinuxHugepageLimit struct { + // Pagesize is the hugepage size + Pagesize string `json:"pageSize"` + // Limit is the limit of "hugepagesize" hugetlb usage + Limit uint64 `json:"limit"` +} + +// LinuxInterfacePriority for network interfaces +type LinuxInterfacePriority struct { + // Name is the name of the network interface + Name string `json:"name"` + // Priority for the interface + Priority uint32 `json:"priority"` +} + +// linuxBlockIODevice holds major:minor format supported in blkio cgroup +type linuxBlockIODevice struct { + // Major is the device's major number. + Major int64 `json:"major"` + // Minor is the device's minor number. + Minor int64 `json:"minor"` +} + +// LinuxWeightDevice struct holds a `major:minor weight` pair for weightDevice +type LinuxWeightDevice struct { + linuxBlockIODevice + // Weight is the bandwidth rate for the device. + Weight *uint16 `json:"weight,omitempty"` + // LeafWeight is the bandwidth rate for the device while competing with the cgroup's child cgroups, CFQ scheduler only + LeafWeight *uint16 `json:"leafWeight,omitempty"` +} + +// LinuxThrottleDevice struct holds a `major:minor rate_per_second` pair +type LinuxThrottleDevice struct { + linuxBlockIODevice + // Rate is the IO rate limit per cgroup per device + Rate uint64 `json:"rate"` +} + +// LinuxBlockIO for Linux cgroup 'blkio' resource management +type LinuxBlockIO struct { + // Specifies per cgroup weight + Weight *uint16 `json:"weight,omitempty"` + // Specifies tasks' weight in the given cgroup while competing with the cgroup's child cgroups, CFQ scheduler only + LeafWeight *uint16 `json:"leafWeight,omitempty"` + // Weight per cgroup per device, can override BlkioWeight + WeightDevice []LinuxWeightDevice `json:"weightDevice,omitempty"` + // IO read rate limit per cgroup per device, bytes per second + ThrottleReadBpsDevice []LinuxThrottleDevice `json:"throttleReadBpsDevice,omitempty"` + // IO write rate limit per cgroup per device, bytes per second + ThrottleWriteBpsDevice []LinuxThrottleDevice `json:"throttleWriteBpsDevice,omitempty"` + // IO read rate limit per cgroup per device, IO per second + ThrottleReadIOPSDevice []LinuxThrottleDevice `json:"throttleReadIOPSDevice,omitempty"` + // IO write rate limit per cgroup per device, IO per second + ThrottleWriteIOPSDevice []LinuxThrottleDevice `json:"throttleWriteIOPSDevice,omitempty"` +} + +// LinuxMemory for Linux cgroup 'memory' resource management +type LinuxMemory struct { + // Memory limit (in bytes). + Limit *int64 `json:"limit,omitempty"` + // Memory reservation or soft_limit (in bytes). + Reservation *int64 `json:"reservation,omitempty"` + // Total memory limit (memory + swap). + Swap *int64 `json:"swap,omitempty"` + // Kernel memory limit (in bytes). + Kernel *int64 `json:"kernel,omitempty"` + // Kernel memory limit for tcp (in bytes) + KernelTCP *int64 `json:"kernelTCP,omitempty"` + // How aggressive the kernel will swap memory pages. + Swappiness *uint64 `json:"swappiness,omitempty"` + // DisableOOMKiller disables the OOM killer for out of memory conditions + DisableOOMKiller *bool `json:"disableOOMKiller,omitempty"` +} + +// LinuxCPU for Linux cgroup 'cpu' resource management +type LinuxCPU struct { + // CPU shares (relative weight (ratio) vs. other cgroups with cpu shares). + Shares *uint64 `json:"shares,omitempty"` + // CPU hardcap limit (in usecs). Allowed cpu time in a given period. + Quota *int64 `json:"quota,omitempty"` + // CPU period to be used for hardcapping (in usecs). + Period *uint64 `json:"period,omitempty"` + // How much time realtime scheduling may use (in usecs). + RealtimeRuntime *int64 `json:"realtimeRuntime,omitempty"` + // CPU period to be used for realtime scheduling (in usecs). + RealtimePeriod *uint64 `json:"realtimePeriod,omitempty"` + // CPUs to use within the cpuset. Default is to use any CPU available. + Cpus string `json:"cpus,omitempty"` + // List of memory nodes in the cpuset. Default is to use any available memory node. + Mems string `json:"mems,omitempty"` +} + +// LinuxPids for Linux cgroup 'pids' resource management (Linux 4.3) +type LinuxPids struct { + // Maximum number of PIDs. Default is "no limit". + Limit int64 `json:"limit"` +} + +// LinuxNetwork identification and priority configuration +type LinuxNetwork struct { + // Set class identifier for container's network packets + ClassID *uint32 `json:"classID,omitempty"` + // Set priority of network traffic for container + Priorities []LinuxInterfacePriority `json:"priorities,omitempty"` +} + +// LinuxResources has container runtime resource constraints +type LinuxResources struct { + // Devices configures the device whitelist. + Devices []LinuxDeviceCgroup `json:"devices,omitempty"` + // Memory restriction configuration + Memory *LinuxMemory `json:"memory,omitempty"` + // CPU resource restriction configuration + CPU *LinuxCPU `json:"cpu,omitempty"` + // Task resource restriction configuration. + Pids *LinuxPids `json:"pids,omitempty"` + // BlockIO restriction configuration + BlockIO *LinuxBlockIO `json:"blockIO,omitempty"` + // Hugetlb limit (in bytes) + HugepageLimits []LinuxHugepageLimit `json:"hugepageLimits,omitempty"` + // Network restriction configuration + Network *LinuxNetwork `json:"network,omitempty"` +} + +// LinuxDevice represents the mknod information for a Linux special device file +type LinuxDevice struct { + // Path to the device. + Path string `json:"path"` + // Device type, block, char, etc. + Type string `json:"type"` + // Major is the device's major number. + Major int64 `json:"major"` + // Minor is the device's minor number. + Minor int64 `json:"minor"` + // FileMode permission bits for the device. + FileMode *os.FileMode `json:"fileMode,omitempty"` + // UID of the device. + UID *uint32 `json:"uid,omitempty"` + // Gid of the device. + GID *uint32 `json:"gid,omitempty"` +} + +// LinuxDeviceCgroup represents a device rule for the whitelist controller +type LinuxDeviceCgroup struct { + // Allow or deny + Allow bool `json:"allow"` + // Device type, block, char, etc. + Type string `json:"type,omitempty"` + // Major is the device's major number. + Major *int64 `json:"major,omitempty"` + // Minor is the device's minor number. + Minor *int64 `json:"minor,omitempty"` + // Cgroup access permissions format, rwm. + Access string `json:"access,omitempty"` +} + +// Solaris contains platform-specific configuration for Solaris application containers. +type Solaris struct { + // SMF FMRI which should go "online" before we start the container process. + Milestone string `json:"milestone,omitempty"` + // Maximum set of privileges any process in this container can obtain. + LimitPriv string `json:"limitpriv,omitempty"` + // The maximum amount of shared memory allowed for this container. + MaxShmMemory string `json:"maxShmMemory,omitempty"` + // Specification for automatic creation of network resources for this container. + Anet []SolarisAnet `json:"anet,omitempty"` + // Set limit on the amount of CPU time that can be used by container. + CappedCPU *SolarisCappedCPU `json:"cappedCPU,omitempty"` + // The physical and swap caps on the memory that can be used by this container. + CappedMemory *SolarisCappedMemory `json:"cappedMemory,omitempty"` +} + +// SolarisCappedCPU allows users to set limit on the amount of CPU time that can be used by container. +type SolarisCappedCPU struct { + Ncpus string `json:"ncpus,omitempty"` +} + +// SolarisCappedMemory allows users to set the physical and swap caps on the memory that can be used by this container. +type SolarisCappedMemory struct { + Physical string `json:"physical,omitempty"` + Swap string `json:"swap,omitempty"` +} + +// SolarisAnet provides the specification for automatic creation of network resources for this container. +type SolarisAnet struct { + // Specify a name for the automatically created VNIC datalink. + Linkname string `json:"linkname,omitempty"` + // Specify the link over which the VNIC will be created. + Lowerlink string `json:"lowerLink,omitempty"` + // The set of IP addresses that the container can use. + Allowedaddr string `json:"allowedAddress,omitempty"` + // Specifies whether allowedAddress limitation is to be applied to the VNIC. + Configallowedaddr string `json:"configureAllowedAddress,omitempty"` + // The value of the optional default router. + Defrouter string `json:"defrouter,omitempty"` + // Enable one or more types of link protection. + Linkprotection string `json:"linkProtection,omitempty"` + // Set the VNIC's macAddress + Macaddress string `json:"macAddress,omitempty"` +} + +// Windows defines the runtime configuration for Windows based containers, including Hyper-V containers. +type Windows struct { + // LayerFolders contains a list of absolute paths to directories containing image layers. + LayerFolders []string `json:"layerFolders"` + // Resources contains information for handling resource constraints for the container. + Resources *WindowsResources `json:"resources,omitempty"` + // CredentialSpec contains a JSON object describing a group Managed Service Account (gMSA) specification. + CredentialSpec interface{} `json:"credentialSpec,omitempty"` + // Servicing indicates if the container is being started in a mode to apply a Windows Update servicing operation. + Servicing bool `json:"servicing,omitempty"` + // IgnoreFlushesDuringBoot indicates if the container is being started in a mode where disk writes are not flushed during its boot process. + IgnoreFlushesDuringBoot bool `json:"ignoreFlushesDuringBoot,omitempty"` + // HyperV contains information for running a container with Hyper-V isolation. + HyperV *WindowsHyperV `json:"hyperv,omitempty"` + // Network restriction configuration. + Network *WindowsNetwork `json:"network,omitempty"` +} + +// WindowsResources has container runtime resource constraints for containers running on Windows. +type WindowsResources struct { + // Memory restriction configuration. + Memory *WindowsMemoryResources `json:"memory,omitempty"` + // CPU resource restriction configuration. + CPU *WindowsCPUResources `json:"cpu,omitempty"` + // Storage restriction configuration. + Storage *WindowsStorageResources `json:"storage,omitempty"` +} + +// WindowsMemoryResources contains memory resource management settings. +type WindowsMemoryResources struct { + // Memory limit in bytes. + Limit *uint64 `json:"limit,omitempty"` +} + +// WindowsCPUResources contains CPU resource management settings. +type WindowsCPUResources struct { + // Number of CPUs available to the container. + Count *uint64 `json:"count,omitempty"` + // CPU shares (relative weight to other containers with cpu shares). + Shares *uint16 `json:"shares,omitempty"` + // Specifies the portion of processor cycles that this container can use as a percentage times 100. + Maximum *uint16 `json:"maximum,omitempty"` +} + +// WindowsStorageResources contains storage resource management settings. +type WindowsStorageResources struct { + // Specifies maximum Iops for the system drive. + Iops *uint64 `json:"iops,omitempty"` + // Specifies maximum bytes per second for the system drive. + Bps *uint64 `json:"bps,omitempty"` + // Sandbox size specifies the minimum size of the system drive in bytes. + SandboxSize *uint64 `json:"sandboxSize,omitempty"` +} + +// WindowsNetwork contains network settings for Windows containers. +type WindowsNetwork struct { + // List of HNS endpoints that the container should connect to. + EndpointList []string `json:"endpointList,omitempty"` + // Specifies if unqualified DNS name resolution is allowed. + AllowUnqualifiedDNSQuery bool `json:"allowUnqualifiedDNSQuery,omitempty"` + // Comma separated list of DNS suffixes to use for name resolution. + DNSSearchList []string `json:"DNSSearchList,omitempty"` + // Name (ID) of the container that we will share with the network stack. + NetworkSharedContainerName string `json:"networkSharedContainerName,omitempty"` +} + +// WindowsHyperV contains information for configuring a container to run with Hyper-V isolation. +type WindowsHyperV struct { + // UtilityVMPath is an optional path to the image used for the Utility VM. + UtilityVMPath string `json:"utilityVMPath,omitempty"` +} + +// LinuxSeccomp represents syscall restrictions +type LinuxSeccomp struct { + DefaultAction LinuxSeccompAction `json:"defaultAction"` + Architectures []Arch `json:"architectures,omitempty"` + Syscalls []LinuxSyscall `json:"syscalls,omitempty"` +} + +// Arch used for additional architectures +type Arch string + +// Additional architectures permitted to be used for system calls +// By default only the native architecture of the kernel is permitted +const ( + ArchX86 Arch = "SCMP_ARCH_X86" + ArchX86_64 Arch = "SCMP_ARCH_X86_64" + ArchX32 Arch = "SCMP_ARCH_X32" + ArchARM Arch = "SCMP_ARCH_ARM" + ArchAARCH64 Arch = "SCMP_ARCH_AARCH64" + ArchMIPS Arch = "SCMP_ARCH_MIPS" + ArchMIPS64 Arch = "SCMP_ARCH_MIPS64" + ArchMIPS64N32 Arch = "SCMP_ARCH_MIPS64N32" + ArchMIPSEL Arch = "SCMP_ARCH_MIPSEL" + ArchMIPSEL64 Arch = "SCMP_ARCH_MIPSEL64" + ArchMIPSEL64N32 Arch = "SCMP_ARCH_MIPSEL64N32" + ArchPPC Arch = "SCMP_ARCH_PPC" + ArchPPC64 Arch = "SCMP_ARCH_PPC64" + ArchPPC64LE Arch = "SCMP_ARCH_PPC64LE" + ArchS390 Arch = "SCMP_ARCH_S390" + ArchS390X Arch = "SCMP_ARCH_S390X" + ArchPARISC Arch = "SCMP_ARCH_PARISC" + ArchPARISC64 Arch = "SCMP_ARCH_PARISC64" +) + +// LinuxSeccompAction taken upon Seccomp rule match +type LinuxSeccompAction string + +// Define actions for Seccomp rules +const ( + ActKill LinuxSeccompAction = "SCMP_ACT_KILL" + ActTrap LinuxSeccompAction = "SCMP_ACT_TRAP" + ActErrno LinuxSeccompAction = "SCMP_ACT_ERRNO" + ActTrace LinuxSeccompAction = "SCMP_ACT_TRACE" + ActAllow LinuxSeccompAction = "SCMP_ACT_ALLOW" +) + +// LinuxSeccompOperator used to match syscall arguments in Seccomp +type LinuxSeccompOperator string + +// Define operators for syscall arguments in Seccomp +const ( + OpNotEqual LinuxSeccompOperator = "SCMP_CMP_NE" + OpLessThan LinuxSeccompOperator = "SCMP_CMP_LT" + OpLessEqual LinuxSeccompOperator = "SCMP_CMP_LE" + OpEqualTo LinuxSeccompOperator = "SCMP_CMP_EQ" + OpGreaterEqual LinuxSeccompOperator = "SCMP_CMP_GE" + OpGreaterThan LinuxSeccompOperator = "SCMP_CMP_GT" + OpMaskedEqual LinuxSeccompOperator = "SCMP_CMP_MASKED_EQ" +) + +// LinuxSeccompArg used for matching specific syscall arguments in Seccomp +type LinuxSeccompArg struct { + Index uint `json:"index"` + Value uint64 `json:"value"` + ValueTwo uint64 `json:"valueTwo,omitempty"` + Op LinuxSeccompOperator `json:"op"` +} + +// LinuxSyscall is used to match a syscall in Seccomp +type LinuxSyscall struct { + Names []string `json:"names"` + Action LinuxSeccompAction `json:"action"` + Args []LinuxSeccompArg `json:"args,omitempty"` +} + +// LinuxIntelRdt has container runtime resource constraints +// for Intel RDT/CAT which introduced in Linux 4.10 kernel +type LinuxIntelRdt struct { + // The schema for L3 cache id and capacity bitmask (CBM) + // Format: "L3:=;=;..." + L3CacheSchema string `json:"l3CacheSchema,omitempty"` +} diff --git a/vendor/github.com/opencontainers/runtime-spec/specs-go/state.go b/vendor/github.com/opencontainers/runtime-spec/specs-go/state.go new file mode 100644 index 000000000..89dce34be --- /dev/null +++ b/vendor/github.com/opencontainers/runtime-spec/specs-go/state.go @@ -0,0 +1,17 @@ +package specs + +// State holds information about the runtime state of the container. +type State struct { + // Version is the version of the specification that is supported. + Version string `json:"ociVersion"` + // ID is the container ID + ID string `json:"id"` + // Status is the runtime status of the container. + Status string `json:"status"` + // Pid is the process ID for the container process. + Pid int `json:"pid,omitempty"` + // Bundle is the path to the container's bundle directory. + Bundle string `json:"bundle"` + // Annotations are key values associated with the container. + Annotations map[string]string `json:"annotations,omitempty"` +} diff --git a/vendor/github.com/opencontainers/runtime-spec/specs-go/version.go b/vendor/github.com/opencontainers/runtime-spec/specs-go/version.go new file mode 100644 index 000000000..926ce6650 --- /dev/null +++ b/vendor/github.com/opencontainers/runtime-spec/specs-go/version.go @@ -0,0 +1,18 @@ +package specs + +import "fmt" + +const ( + // VersionMajor is for an API incompatible changes + VersionMajor = 1 + // VersionMinor is for functionality in a backwards-compatible manner + VersionMinor = 0 + // VersionPatch is for backwards-compatible bug fixes + VersionPatch = 0 + + // VersionDev indicates development branch. Releases will be empty string. + VersionDev = "" +) + +// Version is the specification version that the package types support. +var Version = fmt.Sprintf("%d.%d.%d%s", VersionMajor, VersionMinor, VersionPatch, VersionDev) diff --git a/vendor/github.com/opencontainers/selinux/LICENSE b/vendor/github.com/opencontainers/selinux/LICENSE new file mode 100644 index 000000000..8dada3eda --- /dev/null +++ b/vendor/github.com/opencontainers/selinux/LICENSE @@ -0,0 +1,201 @@ + Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "{}" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright {yyyy} {name of copyright owner} + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/vendor/github.com/opencontainers/selinux/go-selinux/label/label.go b/vendor/github.com/opencontainers/selinux/go-selinux/label/label.go new file mode 100644 index 000000000..6cfc5fded --- /dev/null +++ b/vendor/github.com/opencontainers/selinux/go-selinux/label/label.go @@ -0,0 +1,84 @@ +// +build !selinux !linux + +package label + +// InitLabels returns the process label and file labels to be used within +// the container. A list of options can be passed into this function to alter +// the labels. +func InitLabels(options []string) (string, string, error) { + return "", "", nil +} + +func GetROMountLabel() string { + return "" +} + +func GenLabels(options string) (string, string, error) { + return "", "", nil +} + +func FormatMountLabel(src string, mountLabel string) string { + return src +} + +func SetProcessLabel(processLabel string) error { + return nil +} + +func GetFileLabel(path string) (string, error) { + return "", nil +} + +func SetFileLabel(path string, fileLabel string) error { + return nil +} + +func SetFileCreateLabel(fileLabel string) error { + return nil +} + +func Relabel(path string, fileLabel string, shared bool) error { + return nil +} + +func GetPidLabel(pid int) (string, error) { + return "", nil +} + +func Init() { +} + +func ReserveLabel(label string) error { + return nil +} + +func ReleaseLabel(label string) error { + return nil +} + +// DupSecOpt takes a process label and returns security options that +// can be used to set duplicate labels on future container processes +func DupSecOpt(src string) []string { + return nil +} + +// DisableSecOpt returns a security opt that can disable labeling +// support for future container processes +func DisableSecOpt() []string { + return nil +} + +// Validate checks that the label does not include unexpected options +func Validate(label string) error { + return nil +} + +// RelabelNeeded checks whether the user requested a relabel +func RelabelNeeded(label string) bool { + return false +} + +// IsShared checks that the label includes a "shared" mark +func IsShared(label string) bool { + return false +} diff --git a/vendor/github.com/opencontainers/selinux/go-selinux/label/label_selinux.go b/vendor/github.com/opencontainers/selinux/go-selinux/label/label_selinux.go new file mode 100644 index 000000000..569dcf084 --- /dev/null +++ b/vendor/github.com/opencontainers/selinux/go-selinux/label/label_selinux.go @@ -0,0 +1,204 @@ +// +build selinux,linux + +package label + +import ( + "fmt" + "strings" + + "github.com/opencontainers/selinux/go-selinux" +) + +// Valid Label Options +var validOptions = map[string]bool{ + "disable": true, + "type": true, + "user": true, + "role": true, + "level": true, +} + +var ErrIncompatibleLabel = fmt.Errorf("Bad SELinux option z and Z can not be used together") + +// InitLabels returns the process label and file labels to be used within +// the container. A list of options can be passed into this function to alter +// the labels. The labels returned will include a random MCS String, that is +// guaranteed to be unique. +func InitLabels(options []string) (string, string, error) { + if !selinux.GetEnabled() { + return "", "", nil + } + processLabel, mountLabel := selinux.ContainerLabels() + if processLabel != "" { + pcon := selinux.NewContext(processLabel) + mcon := selinux.NewContext(mountLabel) + for _, opt := range options { + if opt == "disable" { + return "", "", nil + } + if i := strings.Index(opt, ":"); i == -1 { + return "", "", fmt.Errorf("Bad label option %q, valid options 'disable' or \n'user, role, level, type' followed by ':' and a value", opt) + } + con := strings.SplitN(opt, ":", 2) + if !validOptions[con[0]] { + return "", "", fmt.Errorf("Bad label option %q, valid options 'disable, user, role, level, type'", con[0]) + + } + pcon[con[0]] = con[1] + if con[0] == "level" || con[0] == "user" { + mcon[con[0]] = con[1] + } + } + processLabel = pcon.Get() + mountLabel = mcon.Get() + } + return processLabel, mountLabel, nil +} + +func ROMountLabel() string { + return selinux.ROFileLabel() +} + +// DEPRECATED: The GenLabels function is only to be used during the transition to the official API. +func GenLabels(options string) (string, string, error) { + return InitLabels(strings.Fields(options)) +} + +// FormatMountLabel returns a string to be used by the mount command. +// The format of this string will be used to alter the labeling of the mountpoint. +// The string returned is suitable to be used as the options field of the mount command. +// If you need to have additional mount point options, you can pass them in as +// the first parameter. Second parameter is the label that you wish to apply +// to all content in the mount point. +func FormatMountLabel(src, mountLabel string) string { + if mountLabel != "" { + switch src { + case "": + src = fmt.Sprintf("context=%q", mountLabel) + default: + src = fmt.Sprintf("%s,context=%q", src, mountLabel) + } + } + return src +} + +// SetProcessLabel takes a process label and tells the kernel to assign the +// label to the next program executed by the current process. +func SetProcessLabel(processLabel string) error { + if processLabel == "" { + return nil + } + return selinux.SetExecLabel(processLabel) +} + +// ProcessLabel returns the process label that the kernel will assign +// to the next program executed by the current process. If "" is returned +// this indicates that the default labeling will happen for the process. +func ProcessLabel() (string, error) { + return selinux.ExecLabel() +} + +// GetFileLabel returns the label for specified path +func FileLabel(path string) (string, error) { + return selinux.FileLabel(path) +} + +// SetFileLabel modifies the "path" label to the specified file label +func SetFileLabel(path string, fileLabel string) error { + if selinux.GetEnabled() && fileLabel != "" { + return selinux.SetFileLabel(path, fileLabel) + } + return nil +} + +// SetFileCreateLabel tells the kernel the label for all files to be created +func SetFileCreateLabel(fileLabel string) error { + if selinux.GetEnabled() { + return selinux.SetFSCreateLabel(fileLabel) + } + return nil +} + +// Relabel changes the label of path to the filelabel string. +// It changes the MCS label to s0 if shared is true. +// This will allow all containers to share the content. +func Relabel(path string, fileLabel string, shared bool) error { + if !selinux.GetEnabled() { + return nil + } + + if fileLabel == "" { + return nil + } + + exclude_paths := map[string]bool{"/": true, "/usr": true, "/etc": true} + if exclude_paths[path] { + return fmt.Errorf("SELinux relabeling of %s is not allowed", path) + } + + if shared { + c := selinux.NewContext(fileLabel) + c["level"] = "s0" + fileLabel = c.Get() + } + if err := selinux.Chcon(path, fileLabel, true); err != nil { + return err + } + return nil +} + +// PidLabel will return the label of the process running with the specified pid +func PidLabel(pid int) (string, error) { + return selinux.PidLabel(pid) +} + +// Init initialises the labeling system +func Init() { + selinux.GetEnabled() +} + +// ReserveLabel will record the fact that the MCS label has already been used. +// This will prevent InitLabels from using the MCS label in a newly created +// container +func ReserveLabel(label string) error { + selinux.ReserveLabel(label) + return nil +} + +// ReleaseLabel will remove the reservation of the MCS label. +// This will allow InitLabels to use the MCS label in a newly created +// containers +func ReleaseLabel(label string) error { + selinux.ReleaseLabel(label) + return nil +} + +// DupSecOpt takes a process label and returns security options that +// can be used to set duplicate labels on future container processes +func DupSecOpt(src string) []string { + return selinux.DupSecOpt(src) +} + +// DisableSecOpt returns a security opt that can disable labeling +// support for future container processes +func DisableSecOpt() []string { + return selinux.DisableSecOpt() +} + +// Validate checks that the label does not include unexpected options +func Validate(label string) error { + if strings.Contains(label, "z") && strings.Contains(label, "Z") { + return ErrIncompatibleLabel + } + return nil +} + +// RelabelNeeded checks whether the user requested a relabel +func RelabelNeeded(label string) bool { + return strings.Contains(label, "z") || strings.Contains(label, "Z") +} + +// IsShared checks that the label includes a "shared" mark +func IsShared(label string) bool { + return strings.Contains(label, "z") +} diff --git a/vendor/github.com/opencontainers/selinux/go-selinux/selinux.go b/vendor/github.com/opencontainers/selinux/go-selinux/selinux.go new file mode 100644 index 000000000..4cf2c45de --- /dev/null +++ b/vendor/github.com/opencontainers/selinux/go-selinux/selinux.go @@ -0,0 +1,593 @@ +// +build linux + +package selinux + +import ( + "bufio" + "crypto/rand" + "encoding/binary" + "fmt" + "io" + "os" + "path/filepath" + "regexp" + "strconv" + "strings" + "sync" + "syscall" +) + +const ( + // Enforcing constant indicate SELinux is in enforcing mode + Enforcing = 1 + // Permissive constant to indicate SELinux is in permissive mode + Permissive = 0 + // Disabled constant to indicate SELinux is disabled + Disabled = -1 + selinuxDir = "/etc/selinux/" + selinuxConfig = selinuxDir + "config" + selinuxTypeTag = "SELINUXTYPE" + selinuxTag = "SELINUX" + selinuxPath = "/sys/fs/selinux" + xattrNameSelinux = "security.selinux" + stRdOnly = 0x01 +) + +type selinuxState struct { + enabledSet bool + enabled bool + selinuxfsSet bool + selinuxfs string + mcsList map[string]bool + sync.Mutex +} + +var ( + assignRegex = regexp.MustCompile(`^([^=]+)=(.*)$`) + state = selinuxState{ + mcsList: make(map[string]bool), + } +) + +// Context is a representation of the SELinux label broken into 4 parts +type Context map[string]string + +func (s *selinuxState) setEnable(enabled bool) bool { + s.Lock() + defer s.Unlock() + s.enabledSet = true + s.enabled = enabled + return s.enabled +} + +func (s *selinuxState) getEnabled() bool { + s.Lock() + enabled := s.enabled + enabledSet := s.enabledSet + s.Unlock() + if enabledSet { + return enabled + } + + enabled = false + if fs := getSelinuxMountPoint(); fs != "" { + if con, _ := CurrentLabel(); con != "kernel" { + enabled = true + } + } + return s.setEnable(enabled) +} + +// SetDisabled disables selinux support for the package +func SetDisabled() { + state.setEnable(false) +} + +func (s *selinuxState) setSELinuxfs(selinuxfs string) string { + s.Lock() + defer s.Unlock() + s.selinuxfsSet = true + s.selinuxfs = selinuxfs + return s.selinuxfs +} + +func (s *selinuxState) getSELinuxfs() string { + s.Lock() + selinuxfs := s.selinuxfs + selinuxfsSet := s.selinuxfsSet + s.Unlock() + if selinuxfsSet { + return selinuxfs + } + + selinuxfs = "" + f, err := os.Open("/proc/self/mountinfo") + if err != nil { + return selinuxfs + } + defer f.Close() + + scanner := bufio.NewScanner(f) + for scanner.Scan() { + txt := scanner.Text() + // Safe as mountinfo encodes mountpoints with spaces as \040. + sepIdx := strings.Index(txt, " - ") + if sepIdx == -1 { + continue + } + if !strings.Contains(txt[sepIdx:], "selinuxfs") { + continue + } + fields := strings.Split(txt, " ") + if len(fields) < 5 { + continue + } + selinuxfs = fields[4] + break + } + + if selinuxfs != "" { + var buf syscall.Statfs_t + syscall.Statfs(selinuxfs, &buf) + if (buf.Flags & stRdOnly) == 1 { + selinuxfs = "" + } + } + return s.setSELinuxfs(selinuxfs) +} + +// getSelinuxMountPoint returns the path to the mountpoint of an selinuxfs +// filesystem or an empty string if no mountpoint is found. Selinuxfs is +// a proc-like pseudo-filesystem that exposes the selinux policy API to +// processes. The existence of an selinuxfs mount is used to determine +// whether selinux is currently enabled or not. +func getSelinuxMountPoint() string { + return state.getSELinuxfs() +} + +// GetEnabled returns whether selinux is currently enabled. +func GetEnabled() bool { + return state.getEnabled() +} + +func readConfig(target string) (value string) { + var ( + val, key string + bufin *bufio.Reader + ) + + in, err := os.Open(selinuxConfig) + if err != nil { + return "" + } + defer in.Close() + + bufin = bufio.NewReader(in) + + for done := false; !done; { + var line string + if line, err = bufin.ReadString('\n'); err != nil { + if err != io.EOF { + return "" + } + done = true + } + line = strings.TrimSpace(line) + if len(line) == 0 { + // Skip blank lines + continue + } + if line[0] == ';' || line[0] == '#' { + // Skip comments + continue + } + if groups := assignRegex.FindStringSubmatch(line); groups != nil { + key, val = strings.TrimSpace(groups[1]), strings.TrimSpace(groups[2]) + if key == target { + return strings.Trim(val, "\"") + } + } + } + return "" +} + +func getSELinuxPolicyRoot() string { + return selinuxDir + readConfig(selinuxTypeTag) +} + +func readCon(name string) (string, error) { + var val string + + in, err := os.Open(name) + if err != nil { + return "", err + } + defer in.Close() + + _, err = fmt.Fscanf(in, "%s", &val) + return val, err +} + +// SetFileLabel sets the SELinux label for this path or returns an error. +func SetFileLabel(path string, label string) error { + return lsetxattr(path, xattrNameSelinux, []byte(label), 0) +} + +// Filecon returns the SELinux label for this path or returns an error. +func FileLabel(path string) (string, error) { + label, err := lgetxattr(path, xattrNameSelinux) + if err != nil { + return "", err + } + // Trim the NUL byte at the end of the byte buffer, if present. + if len(label) > 0 && label[len(label)-1] == '\x00' { + label = label[:len(label)-1] + } + return string(label), nil +} + +/* +SetFSCreateLabel tells kernel the label to create all file system objects +created by this task. Setting label="" to return to default. +*/ +func SetFSCreateLabel(label string) error { + return writeCon(fmt.Sprintf("/proc/self/task/%d/attr/fscreate", syscall.Gettid()), label) +} + +/* +FSCreateLabel returns the default label the kernel which the kernel is using +for file system objects created by this task. "" indicates default. +*/ +func FSCreateLabel() (string, error) { + return readCon(fmt.Sprintf("/proc/self/task/%d/attr/fscreate", syscall.Gettid())) +} + +// CurrentLabel returns the SELinux label of the current process thread, or an error. +func CurrentLabel() (string, error) { + return readCon(fmt.Sprintf("/proc/self/task/%d/attr/current", syscall.Gettid())) +} + +// PidLabel returns the SELinux label of the given pid, or an error. +func PidLabel(pid int) (string, error) { + return readCon(fmt.Sprintf("/proc/%d/attr/current", pid)) +} + +/* +ExecLabel returns the SELinux label that the kernel will use for any programs +that are executed by the current process thread, or an error. +*/ +func ExecLabel() (string, error) { + return readCon(fmt.Sprintf("/proc/self/task/%d/attr/exec", syscall.Gettid())) +} + +func writeCon(name string, val string) error { + out, err := os.OpenFile(name, os.O_WRONLY, 0) + if err != nil { + return err + } + defer out.Close() + + if val != "" { + _, err = out.Write([]byte(val)) + } else { + _, err = out.Write(nil) + } + return err +} + +/* +SetExecLabel sets the SELinux label that the kernel will use for any programs +that are executed by the current process thread, or an error. +*/ +func SetExecLabel(label string) error { + return writeCon(fmt.Sprintf("/proc/self/task/%d/attr/exec", syscall.Gettid()), label) +} + +// Get returns the Context as a string +func (c Context) Get() string { + return fmt.Sprintf("%s:%s:%s:%s", c["user"], c["role"], c["type"], c["level"]) +} + +// NewContext creates a new Context struct from the specified label +func NewContext(label string) Context { + c := make(Context) + + if len(label) != 0 { + con := strings.SplitN(label, ":", 4) + c["user"] = con[0] + c["role"] = con[1] + c["type"] = con[2] + c["level"] = con[3] + } + return c +} + +// ReserveLabel reserves the MLS/MCS level component of the specified label +func ReserveLabel(label string) { + if len(label) != 0 { + con := strings.SplitN(label, ":", 4) + mcsAdd(con[3]) + } +} + +func selinuxEnforcePath() string { + return fmt.Sprintf("%s/enforce", selinuxPath) +} + +// EnforceMode returns the current SELinux mode Enforcing, Permissive, Disabled +func EnforceMode() int { + var enforce int + + enforceS, err := readCon(selinuxEnforcePath()) + if err != nil { + return -1 + } + + enforce, err = strconv.Atoi(string(enforceS)) + if err != nil { + return -1 + } + return enforce +} + +/* +SetEnforce sets the current SELinux mode Enforcing, Permissive. +Disabled is not valid, since this needs to be set at boot time. +*/ +func SetEnforceMode(mode int) error { + return writeCon(selinuxEnforcePath(), fmt.Sprintf("%d", mode)) +} + +/* +DefaultEnforceMode returns the systems default SELinux mode Enforcing, +Permissive or Disabled. Note this is is just the default at boot time. +EnforceMode tells you the systems current mode. +*/ +func DefaultEnforceMode() int { + switch readConfig(selinuxTag) { + case "enforcing": + return Enforcing + case "permissive": + return Permissive + } + return Disabled +} + +func mcsAdd(mcs string) error { + state.Lock() + defer state.Unlock() + if state.mcsList[mcs] { + return fmt.Errorf("MCS Label already exists") + } + state.mcsList[mcs] = true + return nil +} + +func mcsDelete(mcs string) { + state.Lock() + defer state.Unlock() + state.mcsList[mcs] = false +} + +func intToMcs(id int, catRange uint32) string { + var ( + SETSIZE = int(catRange) + TIER = SETSIZE + ORD = id + ) + + if id < 1 || id > 523776 { + return "" + } + + for ORD > TIER { + ORD = ORD - TIER + TIER-- + } + TIER = SETSIZE - TIER + ORD = ORD + TIER + return fmt.Sprintf("s0:c%d,c%d", TIER, ORD) +} + +func uniqMcs(catRange uint32) string { + var ( + n uint32 + c1, c2 uint32 + mcs string + ) + + for { + binary.Read(rand.Reader, binary.LittleEndian, &n) + c1 = n % catRange + binary.Read(rand.Reader, binary.LittleEndian, &n) + c2 = n % catRange + if c1 == c2 { + continue + } else { + if c1 > c2 { + c1, c2 = c2, c1 + } + } + mcs = fmt.Sprintf("s0:c%d,c%d", c1, c2) + if err := mcsAdd(mcs); err != nil { + continue + } + break + } + return mcs +} + +/* +ReleaseLabel will unreserve the MLS/MCS Level field of the specified label. +Allowing it to be used by another process. +*/ +func ReleaseLabel(label string) { + if len(label) != 0 { + con := strings.SplitN(label, ":", 4) + mcsDelete(con[3]) + } +} + +var roFileLabel string + +// ROFileLabel returns the specified SELinux readonly file label +func ROFileLabel() (fileLabel string) { + return roFileLabel +} + +/* +ContainerLabels returns an allocated processLabel and fileLabel to be used for +container labeling by the calling process. +*/ +func ContainerLabels() (processLabel string, fileLabel string) { + var ( + val, key string + bufin *bufio.Reader + ) + + if !GetEnabled() { + return "", "" + } + lxcPath := fmt.Sprintf("%s/contexts/lxc_contexts", getSELinuxPolicyRoot()) + in, err := os.Open(lxcPath) + if err != nil { + return "", "" + } + defer in.Close() + + bufin = bufio.NewReader(in) + + for done := false; !done; { + var line string + if line, err = bufin.ReadString('\n'); err != nil { + if err == io.EOF { + done = true + } else { + goto exit + } + } + line = strings.TrimSpace(line) + if len(line) == 0 { + // Skip blank lines + continue + } + if line[0] == ';' || line[0] == '#' { + // Skip comments + continue + } + if groups := assignRegex.FindStringSubmatch(line); groups != nil { + key, val = strings.TrimSpace(groups[1]), strings.TrimSpace(groups[2]) + if key == "process" { + processLabel = strings.Trim(val, "\"") + } + if key == "file" { + fileLabel = strings.Trim(val, "\"") + } + if key == "ro_file" { + roFileLabel = strings.Trim(val, "\"") + } + } + } + + if processLabel == "" || fileLabel == "" { + return "", "" + } + + if roFileLabel == "" { + roFileLabel = fileLabel + } +exit: + mcs := uniqMcs(1024) + scon := NewContext(processLabel) + scon["level"] = mcs + processLabel = scon.Get() + scon = NewContext(fileLabel) + scon["level"] = mcs + fileLabel = scon.Get() + return processLabel, fileLabel +} + +// SecurityCheckContext validates that the SELinux label is understood by the kernel +func SecurityCheckContext(val string) error { + return writeCon(fmt.Sprintf("%s.context", selinuxPath), val) +} + +/* +CopyLevel returns a label with the MLS/MCS level from src label replaces on +the dest label. +*/ +func CopyLevel(src, dest string) (string, error) { + if src == "" { + return "", nil + } + if err := SecurityCheckContext(src); err != nil { + return "", err + } + if err := SecurityCheckContext(dest); err != nil { + return "", err + } + scon := NewContext(src) + tcon := NewContext(dest) + mcsDelete(tcon["level"]) + mcsAdd(scon["level"]) + tcon["level"] = scon["level"] + return tcon.Get(), nil +} + +// Prevent users from relabing system files +func badPrefix(fpath string) error { + var badprefixes = []string{"/usr"} + + for _, prefix := range badprefixes { + if fpath == prefix || strings.HasPrefix(fpath, fmt.Sprintf("%s/", prefix)) { + return fmt.Errorf("relabeling content in %s is not allowed", prefix) + } + } + return nil +} + +// Chcon changes the fpath file object to the SELinux label label. +// If the fpath is a directory and recurse is true Chcon will walk the +// directory tree setting the label +func Chcon(fpath string, label string, recurse bool) error { + if label == "" { + return nil + } + if err := badPrefix(fpath); err != nil { + return err + } + callback := func(p string, info os.FileInfo, err error) error { + return SetFileLabel(p, label) + } + + if recurse { + return filepath.Walk(fpath, callback) + } + + return SetFileLabel(fpath, label) +} + +// DupSecOpt takes an SELinux process label and returns security options that +// can will set the SELinux Type and Level for future container processes +func DupSecOpt(src string) []string { + if src == "" { + return nil + } + con := NewContext(src) + if con["user"] == "" || + con["role"] == "" || + con["type"] == "" || + con["level"] == "" { + return nil + } + return []string{"user:" + con["user"], + "role:" + con["role"], + "type:" + con["type"], + "level:" + con["level"]} +} + +// DisableSecOpt returns a security opt that can be used to disabling SELinux +// labeling support for future container processes +func DisableSecOpt() []string { + return []string{"disable"} +} diff --git a/vendor/github.com/opencontainers/selinux/go-selinux/xattrs.go b/vendor/github.com/opencontainers/selinux/go-selinux/xattrs.go new file mode 100644 index 000000000..7f2ef8504 --- /dev/null +++ b/vendor/github.com/opencontainers/selinux/go-selinux/xattrs.go @@ -0,0 +1,78 @@ +// +build linux + +package selinux + +import ( + "syscall" + "unsafe" +) + +var _zero uintptr + +// Returns a []byte slice if the xattr is set and nil otherwise +// Requires path and its attribute as arguments +func lgetxattr(path string, attr string) ([]byte, error) { + var sz int + pathBytes, err := syscall.BytePtrFromString(path) + if err != nil { + return nil, err + } + attrBytes, err := syscall.BytePtrFromString(attr) + if err != nil { + return nil, err + } + + // Start with a 128 length byte array + sz = 128 + dest := make([]byte, sz) + destBytes := unsafe.Pointer(&dest[0]) + _sz, _, errno := syscall.Syscall6(syscall.SYS_LGETXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(attrBytes)), uintptr(destBytes), uintptr(len(dest)), 0, 0) + + switch { + case errno == syscall.ENODATA: + return nil, errno + case errno == syscall.ENOTSUP: + return nil, errno + case errno == syscall.ERANGE: + // 128 byte array might just not be good enough, + // A dummy buffer is used ``uintptr(0)`` to get real size + // of the xattrs on disk + _sz, _, errno = syscall.Syscall6(syscall.SYS_LGETXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(attrBytes)), uintptr(unsafe.Pointer(nil)), uintptr(0), 0, 0) + sz = int(_sz) + if sz < 0 { + return nil, errno + } + dest = make([]byte, sz) + destBytes := unsafe.Pointer(&dest[0]) + _sz, _, errno = syscall.Syscall6(syscall.SYS_LGETXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(attrBytes)), uintptr(destBytes), uintptr(len(dest)), 0, 0) + if errno != 0 { + return nil, errno + } + case errno != 0: + return nil, errno + } + sz = int(_sz) + return dest[:sz], nil +} + +func lsetxattr(path string, attr string, data []byte, flags int) error { + pathBytes, err := syscall.BytePtrFromString(path) + if err != nil { + return err + } + attrBytes, err := syscall.BytePtrFromString(attr) + if err != nil { + return err + } + var dataBytes unsafe.Pointer + if len(data) > 0 { + dataBytes = unsafe.Pointer(&data[0]) + } else { + dataBytes = unsafe.Pointer(&_zero) + } + _, _, errno := syscall.Syscall6(syscall.SYS_LSETXATTR, uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(attrBytes)), uintptr(dataBytes), uintptr(len(data)), uintptr(flags), 0) + if errno != 0 { + return errno + } + return nil +} diff --git a/vendor/github.com/ryanuber/columnize/LICENSE b/vendor/github.com/ryanuber/columnize/LICENSE new file mode 100644 index 000000000..b9c0e2b68 --- /dev/null +++ b/vendor/github.com/ryanuber/columnize/LICENSE @@ -0,0 +1,20 @@ +Copyright (c) 2016 Ryan Uber + +Permission is hereby granted, free of charge, to any person obtaining +a copy of this software and associated documentation files (the +"Software"), to deal in the Software without restriction, including +without limitation the rights to use, copy, modify, merge, publish, +distribute, sublicense, and/or sell copies of the Software, and to +permit persons to whom the Software is furnished to do so, subject to +the following conditions: + +The above copyright notice and this permission notice shall be +included in all copies or substantial portions of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, +EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF +MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND +NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE +LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION +OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION +WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/seccomp/libseccomp-golang/LICENSE b/vendor/github.com/seccomp/libseccomp-golang/LICENSE similarity index 100% rename from vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/seccomp/libseccomp-golang/LICENSE rename to vendor/github.com/seccomp/libseccomp-golang/LICENSE diff --git a/vendor/github.com/seccomp/libseccomp-golang/README b/vendor/github.com/seccomp/libseccomp-golang/README new file mode 100644 index 000000000..920c6ec9c --- /dev/null +++ b/vendor/github.com/seccomp/libseccomp-golang/README @@ -0,0 +1,50 @@ +libseccomp-golang: Go Language Bindings for the libseccomp Project +=============================================================================== +https://github.com/seccomp/libseccomp-golang +https://github.com/seccomp/libseccomp + +The libseccomp library provides an easy to use, platform independent, interface +to the Linux Kernel's syscall filtering mechanism. The libseccomp API is +designed to abstract away the underlying BPF based syscall filter language and +present a more conventional function-call based filtering interface that should +be familiar to, and easily adopted by, application developers. + +The libseccomp-golang library provides a Go based interface to the libseccomp +library. + +* Online Resources + +The library source repository currently lives on GitHub at the following URLs: + + -> https://github.com/seccomp/libseccomp-golang + -> https://github.com/seccomp/libseccomp + +The project mailing list is currently hosted on Google Groups at the URL below, +please note that a Google account is not required to subscribe to the mailing +list. + + -> https://groups.google.com/d/forum/libseccomp + +Documentation is also available at: + + -> https://godoc.org/github.com/seccomp/libseccomp-golang + +* Installing the package + +The libseccomp-golang bindings require at least Go v1.2.1 and GCC v4.8.4; +earlier versions may yield unpredictable results. If you meet these +requirements you can install this package using the command below: + + $ go get github.com/seccomp/libseccomp-golang + +* Testing the Library + +A number of tests and lint related recipes are provided in the Makefile, if +you want to run the standard regression tests, you can excute the following: + + $ make check + +In order to execute the 'make lint' recipe the 'golint' tool is needed, it +can be found at: + + -> https://github.com/golang/lint diff --git a/vendor/github.com/seccomp/libseccomp-golang/seccomp.go b/vendor/github.com/seccomp/libseccomp-golang/seccomp.go new file mode 100644 index 000000000..d3d9e6bfb --- /dev/null +++ b/vendor/github.com/seccomp/libseccomp-golang/seccomp.go @@ -0,0 +1,864 @@ +// +build linux + +// Public API specification for libseccomp Go bindings +// Contains public API for the bindings + +// Package seccomp provides bindings for libseccomp, a library wrapping the Linux +// seccomp syscall. Seccomp enables an application to restrict system call use +// for itself and its children. +package seccomp + +import ( + "fmt" + "os" + "runtime" + "strings" + "sync" + "syscall" + "unsafe" +) + +// C wrapping code + +// #cgo pkg-config: libseccomp +// #include +// #include +import "C" + +// Exported types + +// VersionError denotes that the system libseccomp version is incompatible +// with this package. +type VersionError struct { + message string + minimum string +} + +func (e VersionError) Error() string { + format := "Libseccomp version too low: " + if e.message != "" { + format += e.message + ": " + } + format += "minimum supported is " + if e.minimum != "" { + format += e.minimum + ": " + } else { + format += "2.2.0: " + } + format += "detected %d.%d.%d" + return fmt.Sprintf(format, verMajor, verMinor, verMicro) +} + +// ScmpArch represents a CPU architecture. Seccomp can restrict syscalls on a +// per-architecture basis. +type ScmpArch uint + +// ScmpAction represents an action to be taken on a filter rule match in +// libseccomp +type ScmpAction uint + +// ScmpCompareOp represents a comparison operator which can be used in a filter +// rule +type ScmpCompareOp uint + +// ScmpCondition represents a rule in a libseccomp filter context +type ScmpCondition struct { + Argument uint `json:"argument,omitempty"` + Op ScmpCompareOp `json:"operator,omitempty"` + Operand1 uint64 `json:"operand_one,omitempty"` + Operand2 uint64 `json:"operand_two,omitempty"` +} + +// ScmpSyscall represents a Linux System Call +type ScmpSyscall int32 + +// Exported Constants + +const ( + // Valid architectures recognized by libseccomp + // PowerPC and S390(x) architectures are unavailable below library version + // v2.3.0 and will returns errors if used with incompatible libraries + + // ArchInvalid is a placeholder to ensure uninitialized ScmpArch + // variables are invalid + ArchInvalid ScmpArch = iota + // ArchNative is the native architecture of the kernel + ArchNative ScmpArch = iota + // ArchX86 represents 32-bit x86 syscalls + ArchX86 ScmpArch = iota + // ArchAMD64 represents 64-bit x86-64 syscalls + ArchAMD64 ScmpArch = iota + // ArchX32 represents 64-bit x86-64 syscalls (32-bit pointers) + ArchX32 ScmpArch = iota + // ArchARM represents 32-bit ARM syscalls + ArchARM ScmpArch = iota + // ArchARM64 represents 64-bit ARM syscalls + ArchARM64 ScmpArch = iota + // ArchMIPS represents 32-bit MIPS syscalls + ArchMIPS ScmpArch = iota + // ArchMIPS64 represents 64-bit MIPS syscalls + ArchMIPS64 ScmpArch = iota + // ArchMIPS64N32 represents 64-bit MIPS syscalls (32-bit pointers) + ArchMIPS64N32 ScmpArch = iota + // ArchMIPSEL represents 32-bit MIPS syscalls (little endian) + ArchMIPSEL ScmpArch = iota + // ArchMIPSEL64 represents 64-bit MIPS syscalls (little endian) + ArchMIPSEL64 ScmpArch = iota + // ArchMIPSEL64N32 represents 64-bit MIPS syscalls (little endian, + // 32-bit pointers) + ArchMIPSEL64N32 ScmpArch = iota + // ArchPPC represents 32-bit POWERPC syscalls + ArchPPC ScmpArch = iota + // ArchPPC64 represents 64-bit POWER syscalls (big endian) + ArchPPC64 ScmpArch = iota + // ArchPPC64LE represents 64-bit POWER syscalls (little endian) + ArchPPC64LE ScmpArch = iota + // ArchS390 represents 31-bit System z/390 syscalls + ArchS390 ScmpArch = iota + // ArchS390X represents 64-bit System z/390 syscalls + ArchS390X ScmpArch = iota +) + +const ( + // Supported actions on filter match + + // ActInvalid is a placeholder to ensure uninitialized ScmpAction + // variables are invalid + ActInvalid ScmpAction = iota + // ActKill kills the process + ActKill ScmpAction = iota + // ActTrap throws SIGSYS + ActTrap ScmpAction = iota + // ActErrno causes the syscall to return a negative error code. This + // code can be set with the SetReturnCode method + ActErrno ScmpAction = iota + // ActTrace causes the syscall to notify tracing processes with the + // given error code. This code can be set with the SetReturnCode method + ActTrace ScmpAction = iota + // ActAllow permits the syscall to continue execution + ActAllow ScmpAction = iota +) + +const ( + // These are comparison operators used in conditional seccomp rules + // They are used to compare the value of a single argument of a syscall + // against a user-defined constant + + // CompareInvalid is a placeholder to ensure uninitialized ScmpCompareOp + // variables are invalid + CompareInvalid ScmpCompareOp = iota + // CompareNotEqual returns true if the argument is not equal to the + // given value + CompareNotEqual ScmpCompareOp = iota + // CompareLess returns true if the argument is less than the given value + CompareLess ScmpCompareOp = iota + // CompareLessOrEqual returns true if the argument is less than or equal + // to the given value + CompareLessOrEqual ScmpCompareOp = iota + // CompareEqual returns true if the argument is equal to the given value + CompareEqual ScmpCompareOp = iota + // CompareGreaterEqual returns true if the argument is greater than or + // equal to the given value + CompareGreaterEqual ScmpCompareOp = iota + // CompareGreater returns true if the argument is greater than the given + // value + CompareGreater ScmpCompareOp = iota + // CompareMaskedEqual returns true if the argument is equal to the given + // value, when masked (bitwise &) against the second given value + CompareMaskedEqual ScmpCompareOp = iota +) + +// Helpers for types + +// GetArchFromString returns an ScmpArch constant from a string representing an +// architecture +func GetArchFromString(arch string) (ScmpArch, error) { + if err := ensureSupportedVersion(); err != nil { + return ArchInvalid, err + } + + switch strings.ToLower(arch) { + case "x86": + return ArchX86, nil + case "amd64", "x86-64", "x86_64", "x64": + return ArchAMD64, nil + case "x32": + return ArchX32, nil + case "arm": + return ArchARM, nil + case "arm64", "aarch64": + return ArchARM64, nil + case "mips": + return ArchMIPS, nil + case "mips64": + return ArchMIPS64, nil + case "mips64n32": + return ArchMIPS64N32, nil + case "mipsel": + return ArchMIPSEL, nil + case "mipsel64": + return ArchMIPSEL64, nil + case "mipsel64n32": + return ArchMIPSEL64N32, nil + case "ppc": + return ArchPPC, nil + case "ppc64": + return ArchPPC64, nil + case "ppc64le": + return ArchPPC64LE, nil + case "s390": + return ArchS390, nil + case "s390x": + return ArchS390X, nil + default: + return ArchInvalid, fmt.Errorf("cannot convert unrecognized string %s", arch) + } +} + +// String returns a string representation of an architecture constant +func (a ScmpArch) String() string { + switch a { + case ArchX86: + return "x86" + case ArchAMD64: + return "amd64" + case ArchX32: + return "x32" + case ArchARM: + return "arm" + case ArchARM64: + return "arm64" + case ArchMIPS: + return "mips" + case ArchMIPS64: + return "mips64" + case ArchMIPS64N32: + return "mips64n32" + case ArchMIPSEL: + return "mipsel" + case ArchMIPSEL64: + return "mipsel64" + case ArchMIPSEL64N32: + return "mipsel64n32" + case ArchPPC: + return "ppc" + case ArchPPC64: + return "ppc64" + case ArchPPC64LE: + return "ppc64le" + case ArchS390: + return "s390" + case ArchS390X: + return "s390x" + case ArchNative: + return "native" + case ArchInvalid: + return "Invalid architecture" + default: + return "Unknown architecture" + } +} + +// String returns a string representation of a comparison operator constant +func (a ScmpCompareOp) String() string { + switch a { + case CompareNotEqual: + return "Not equal" + case CompareLess: + return "Less than" + case CompareLessOrEqual: + return "Less than or equal to" + case CompareEqual: + return "Equal" + case CompareGreaterEqual: + return "Greater than or equal to" + case CompareGreater: + return "Greater than" + case CompareMaskedEqual: + return "Masked equality" + case CompareInvalid: + return "Invalid comparison operator" + default: + return "Unrecognized comparison operator" + } +} + +// String returns a string representation of a seccomp match action +func (a ScmpAction) String() string { + switch a & 0xFFFF { + case ActKill: + return "Action: Kill Process" + case ActTrap: + return "Action: Send SIGSYS" + case ActErrno: + return fmt.Sprintf("Action: Return error code %d", (a >> 16)) + case ActTrace: + return fmt.Sprintf("Action: Notify tracing processes with code %d", + (a >> 16)) + case ActAllow: + return "Action: Allow system call" + default: + return "Unrecognized Action" + } +} + +// SetReturnCode adds a return code to a supporting ScmpAction, clearing any +// existing code Only valid on ActErrno and ActTrace. Takes no action otherwise. +// Accepts 16-bit return code as argument. +// Returns a valid ScmpAction of the original type with the new error code set. +func (a ScmpAction) SetReturnCode(code int16) ScmpAction { + aTmp := a & 0x0000FFFF + if aTmp == ActErrno || aTmp == ActTrace { + return (aTmp | (ScmpAction(code)&0xFFFF)<<16) + } + return a +} + +// GetReturnCode returns the return code of an ScmpAction +func (a ScmpAction) GetReturnCode() int16 { + return int16(a >> 16) +} + +// General utility functions + +// GetLibraryVersion returns the version of the library the bindings are built +// against. +// The version is formatted as follows: Major.Minor.Micro +func GetLibraryVersion() (major, minor, micro uint) { + return verMajor, verMinor, verMicro +} + +// Syscall functions + +// GetName retrieves the name of a syscall from its number. +// Acts on any syscall number. +// Returns either a string containing the name of the syscall, or an error. +func (s ScmpSyscall) GetName() (string, error) { + return s.GetNameByArch(ArchNative) +} + +// GetNameByArch retrieves the name of a syscall from its number for a given +// architecture. +// Acts on any syscall number. +// Accepts a valid architecture constant. +// Returns either a string containing the name of the syscall, or an error. +// if the syscall is unrecognized or an issue occurred. +func (s ScmpSyscall) GetNameByArch(arch ScmpArch) (string, error) { + if err := sanitizeArch(arch); err != nil { + return "", err + } + + cString := C.seccomp_syscall_resolve_num_arch(arch.toNative(), C.int(s)) + if cString == nil { + return "", fmt.Errorf("could not resolve syscall name") + } + defer C.free(unsafe.Pointer(cString)) + + finalStr := C.GoString(cString) + return finalStr, nil +} + +// GetSyscallFromName returns the number of a syscall by name on the kernel's +// native architecture. +// Accepts a string containing the name of a syscall. +// Returns the number of the syscall, or an error if no syscall with that name +// was found. +func GetSyscallFromName(name string) (ScmpSyscall, error) { + if err := ensureSupportedVersion(); err != nil { + return 0, err + } + + cString := C.CString(name) + defer C.free(unsafe.Pointer(cString)) + + result := C.seccomp_syscall_resolve_name(cString) + if result == scmpError { + return 0, fmt.Errorf("could not resolve name to syscall") + } + + return ScmpSyscall(result), nil +} + +// GetSyscallFromNameByArch returns the number of a syscall by name for a given +// architecture's ABI. +// Accepts the name of a syscall and an architecture constant. +// Returns the number of the syscall, or an error if an invalid architecture is +// passed or a syscall with that name was not found. +func GetSyscallFromNameByArch(name string, arch ScmpArch) (ScmpSyscall, error) { + if err := ensureSupportedVersion(); err != nil { + return 0, err + } + if err := sanitizeArch(arch); err != nil { + return 0, err + } + + cString := C.CString(name) + defer C.free(unsafe.Pointer(cString)) + + result := C.seccomp_syscall_resolve_name_arch(arch.toNative(), cString) + if result == scmpError { + return 0, fmt.Errorf("could not resolve name to syscall") + } + + return ScmpSyscall(result), nil +} + +// MakeCondition creates and returns a new condition to attach to a filter rule. +// Associated rules will only match if this condition is true. +// Accepts the number the argument we are checking, and a comparison operator +// and value to compare to. +// The rule will match if argument $arg (zero-indexed) of the syscall is +// $COMPARE_OP the provided comparison value. +// Some comparison operators accept two values. Masked equals, for example, +// will mask $arg of the syscall with the second value provided (via bitwise +// AND) and then compare against the first value provided. +// For example, in the less than or equal case, if the syscall argument was +// 0 and the value provided was 1, the condition would match, as 0 is less +// than or equal to 1. +// Return either an error on bad argument or a valid ScmpCondition struct. +func MakeCondition(arg uint, comparison ScmpCompareOp, values ...uint64) (ScmpCondition, error) { + var condStruct ScmpCondition + + if err := ensureSupportedVersion(); err != nil { + return condStruct, err + } + + if comparison == CompareInvalid { + return condStruct, fmt.Errorf("invalid comparison operator") + } else if arg > 5 { + return condStruct, fmt.Errorf("syscalls only have up to 6 arguments") + } else if len(values) > 2 { + return condStruct, fmt.Errorf("conditions can have at most 2 arguments") + } else if len(values) == 0 { + return condStruct, fmt.Errorf("must provide at least one value to compare against") + } + + condStruct.Argument = arg + condStruct.Op = comparison + condStruct.Operand1 = values[0] + if len(values) == 2 { + condStruct.Operand2 = values[1] + } else { + condStruct.Operand2 = 0 // Unused + } + + return condStruct, nil +} + +// Utility Functions + +// GetNativeArch returns architecture token representing the native kernel +// architecture +func GetNativeArch() (ScmpArch, error) { + if err := ensureSupportedVersion(); err != nil { + return ArchInvalid, err + } + + arch := C.seccomp_arch_native() + + return archFromNative(arch) +} + +// Public Filter API + +// ScmpFilter represents a filter context in libseccomp. +// A filter context is initially empty. Rules can be added to it, and it can +// then be loaded into the kernel. +type ScmpFilter struct { + filterCtx C.scmp_filter_ctx + valid bool + lock sync.Mutex +} + +// NewFilter creates and returns a new filter context. +// Accepts a default action to be taken for syscalls which match no rules in +// the filter. +// Returns a reference to a valid filter context, or nil and an error if the +// filter context could not be created or an invalid default action was given. +func NewFilter(defaultAction ScmpAction) (*ScmpFilter, error) { + if err := ensureSupportedVersion(); err != nil { + return nil, err + } + + if err := sanitizeAction(defaultAction); err != nil { + return nil, err + } + + fPtr := C.seccomp_init(defaultAction.toNative()) + if fPtr == nil { + return nil, fmt.Errorf("could not create filter") + } + + filter := new(ScmpFilter) + filter.filterCtx = fPtr + filter.valid = true + runtime.SetFinalizer(filter, filterFinalizer) + + // Enable TSync so all goroutines will receive the same rules + // If the kernel does not support TSYNC, allow us to continue without error + if err := filter.setFilterAttr(filterAttrTsync, 0x1); err != nil && err != syscall.ENOTSUP { + filter.Release() + return nil, fmt.Errorf("could not create filter - error setting tsync bit: %v", err) + } + + return filter, nil +} + +// IsValid determines whether a filter context is valid to use. +// Some operations (Release and Merge) render filter contexts invalid and +// consequently prevent further use. +func (f *ScmpFilter) IsValid() bool { + f.lock.Lock() + defer f.lock.Unlock() + + return f.valid +} + +// Reset resets a filter context, removing all its existing state. +// Accepts a new default action to be taken for syscalls which do not match. +// Returns an error if the filter or action provided are invalid. +func (f *ScmpFilter) Reset(defaultAction ScmpAction) error { + f.lock.Lock() + defer f.lock.Unlock() + + if err := sanitizeAction(defaultAction); err != nil { + return err + } else if !f.valid { + return errBadFilter + } + + retCode := C.seccomp_reset(f.filterCtx, defaultAction.toNative()) + if retCode != 0 { + return syscall.Errno(-1 * retCode) + } + + return nil +} + +// Release releases a filter context, freeing its memory. Should be called after +// loading into the kernel, when the filter is no longer needed. +// After calling this function, the given filter is no longer valid and cannot +// be used. +// Release() will be invoked automatically when a filter context is garbage +// collected, but can also be called manually to free memory. +func (f *ScmpFilter) Release() { + f.lock.Lock() + defer f.lock.Unlock() + + if !f.valid { + return + } + + f.valid = false + C.seccomp_release(f.filterCtx) +} + +// Merge merges two filter contexts. +// The source filter src will be released as part of the process, and will no +// longer be usable or valid after this call. +// To be merged, filters must NOT share any architectures, and all their +// attributes (Default Action, Bad Arch Action, and No New Privs bools) +// must match. +// The filter src will be merged into the filter this is called on. +// The architectures of the src filter not present in the destination, and all +// associated rules, will be added to the destination. +// Returns an error if merging the filters failed. +func (f *ScmpFilter) Merge(src *ScmpFilter) error { + f.lock.Lock() + defer f.lock.Unlock() + + src.lock.Lock() + defer src.lock.Unlock() + + if !src.valid || !f.valid { + return fmt.Errorf("one or more of the filter contexts is invalid or uninitialized") + } + + // Merge the filters + retCode := C.seccomp_merge(f.filterCtx, src.filterCtx) + if syscall.Errno(-1*retCode) == syscall.EINVAL { + return fmt.Errorf("filters could not be merged due to a mismatch in attributes or invalid filter") + } else if retCode != 0 { + return syscall.Errno(-1 * retCode) + } + + src.valid = false + + return nil +} + +// IsArchPresent checks if an architecture is present in a filter. +// If a filter contains an architecture, it uses its default action for +// syscalls which do not match rules in it, and its rules can match syscalls +// for that ABI. +// If a filter does not contain an architecture, all syscalls made to that +// kernel ABI will fail with the filter's default Bad Architecture Action +// (by default, killing the process). +// Accepts an architecture constant. +// Returns true if the architecture is present in the filter, false otherwise, +// and an error on an invalid filter context, architecture constant, or an +// issue with the call to libseccomp. +func (f *ScmpFilter) IsArchPresent(arch ScmpArch) (bool, error) { + f.lock.Lock() + defer f.lock.Unlock() + + if err := sanitizeArch(arch); err != nil { + return false, err + } else if !f.valid { + return false, errBadFilter + } + + retCode := C.seccomp_arch_exist(f.filterCtx, arch.toNative()) + if syscall.Errno(-1*retCode) == syscall.EEXIST { + // -EEXIST is "arch not present" + return false, nil + } else if retCode != 0 { + return false, syscall.Errno(-1 * retCode) + } + + return true, nil +} + +// AddArch adds an architecture to the filter. +// Accepts an architecture constant. +// Returns an error on invalid filter context or architecture token, or an +// issue with the call to libseccomp. +func (f *ScmpFilter) AddArch(arch ScmpArch) error { + f.lock.Lock() + defer f.lock.Unlock() + + if err := sanitizeArch(arch); err != nil { + return err + } else if !f.valid { + return errBadFilter + } + + // Libseccomp returns -EEXIST if the specified architecture is already + // present. Succeed silently in this case, as it's not fatal, and the + // architecture is present already. + retCode := C.seccomp_arch_add(f.filterCtx, arch.toNative()) + if retCode != 0 && syscall.Errno(-1*retCode) != syscall.EEXIST { + return syscall.Errno(-1 * retCode) + } + + return nil +} + +// RemoveArch removes an architecture from the filter. +// Accepts an architecture constant. +// Returns an error on invalid filter context or architecture token, or an +// issue with the call to libseccomp. +func (f *ScmpFilter) RemoveArch(arch ScmpArch) error { + f.lock.Lock() + defer f.lock.Unlock() + + if err := sanitizeArch(arch); err != nil { + return err + } else if !f.valid { + return errBadFilter + } + + // Similar to AddArch, -EEXIST is returned if the arch is not present + // Succeed silently in that case, this is not fatal and the architecture + // is not present in the filter after RemoveArch + retCode := C.seccomp_arch_remove(f.filterCtx, arch.toNative()) + if retCode != 0 && syscall.Errno(-1*retCode) != syscall.EEXIST { + return syscall.Errno(-1 * retCode) + } + + return nil +} + +// Load loads a filter context into the kernel. +// Returns an error if the filter context is invalid or the syscall failed. +func (f *ScmpFilter) Load() error { + f.lock.Lock() + defer f.lock.Unlock() + + if !f.valid { + return errBadFilter + } + + if retCode := C.seccomp_load(f.filterCtx); retCode != 0 { + return syscall.Errno(-1 * retCode) + } + + return nil +} + +// GetDefaultAction returns the default action taken on a syscall which does not +// match a rule in the filter, or an error if an issue was encountered +// retrieving the value. +func (f *ScmpFilter) GetDefaultAction() (ScmpAction, error) { + action, err := f.getFilterAttr(filterAttrActDefault) + if err != nil { + return 0x0, err + } + + return actionFromNative(action) +} + +// GetBadArchAction returns the default action taken on a syscall for an +// architecture not in the filter, or an error if an issue was encountered +// retrieving the value. +func (f *ScmpFilter) GetBadArchAction() (ScmpAction, error) { + action, err := f.getFilterAttr(filterAttrActBadArch) + if err != nil { + return 0x0, err + } + + return actionFromNative(action) +} + +// GetNoNewPrivsBit returns the current state the No New Privileges bit will be set +// to on the filter being loaded, or an error if an issue was encountered +// retrieving the value. +// The No New Privileges bit tells the kernel that new processes run with exec() +// cannot gain more privileges than the process that ran exec(). +// For example, a process with No New Privileges set would be unable to exec +// setuid/setgid executables. +func (f *ScmpFilter) GetNoNewPrivsBit() (bool, error) { + noNewPrivs, err := f.getFilterAttr(filterAttrNNP) + if err != nil { + return false, err + } + + if noNewPrivs == 0 { + return false, nil + } + + return true, nil +} + +// SetBadArchAction sets the default action taken on a syscall for an +// architecture not in the filter, or an error if an issue was encountered +// setting the value. +func (f *ScmpFilter) SetBadArchAction(action ScmpAction) error { + if err := sanitizeAction(action); err != nil { + return err + } + + return f.setFilterAttr(filterAttrActBadArch, action.toNative()) +} + +// SetNoNewPrivsBit sets the state of the No New Privileges bit, which will be +// applied on filter load, or an error if an issue was encountered setting the +// value. +// Filters with No New Privileges set to 0 can only be loaded if the process +// has the CAP_SYS_ADMIN capability. +func (f *ScmpFilter) SetNoNewPrivsBit(state bool) error { + var toSet C.uint32_t = 0x0 + + if state { + toSet = 0x1 + } + + return f.setFilterAttr(filterAttrNNP, toSet) +} + +// SetSyscallPriority sets a syscall's priority. +// This provides a hint to the filter generator in libseccomp about the +// importance of this syscall. High-priority syscalls are placed +// first in the filter code, and incur less overhead (at the expense of +// lower-priority syscalls). +func (f *ScmpFilter) SetSyscallPriority(call ScmpSyscall, priority uint8) error { + f.lock.Lock() + defer f.lock.Unlock() + + if !f.valid { + return errBadFilter + } + + if retCode := C.seccomp_syscall_priority(f.filterCtx, C.int(call), + C.uint8_t(priority)); retCode != 0 { + return syscall.Errno(-1 * retCode) + } + + return nil +} + +// AddRule adds a single rule for an unconditional action on a syscall. +// Accepts the number of the syscall and the action to be taken on the call +// being made. +// Returns an error if an issue was encountered adding the rule. +func (f *ScmpFilter) AddRule(call ScmpSyscall, action ScmpAction) error { + return f.addRuleGeneric(call, action, false, nil) +} + +// AddRuleExact adds a single rule for an unconditional action on a syscall. +// Accepts the number of the syscall and the action to be taken on the call +// being made. +// No modifications will be made to the rule, and it will fail to add if it +// cannot be applied to the current architecture without modification. +// The rule will function exactly as described, but it may not function identically +// (or be able to be applied to) all architectures. +// Returns an error if an issue was encountered adding the rule. +func (f *ScmpFilter) AddRuleExact(call ScmpSyscall, action ScmpAction) error { + return f.addRuleGeneric(call, action, true, nil) +} + +// AddRuleConditional adds a single rule for a conditional action on a syscall. +// Returns an error if an issue was encountered adding the rule. +// All conditions must match for the rule to match. +// There is a bug in library versions below v2.2.1 which can, in some cases, +// cause conditions to be lost when more than one are used. Consequently, +// AddRuleConditional is disabled on library versions lower than v2.2.1 +func (f *ScmpFilter) AddRuleConditional(call ScmpSyscall, action ScmpAction, conds []ScmpCondition) error { + return f.addRuleGeneric(call, action, false, conds) +} + +// AddRuleConditionalExact adds a single rule for a conditional action on a +// syscall. +// No modifications will be made to the rule, and it will fail to add if it +// cannot be applied to the current architecture without modification. +// The rule will function exactly as described, but it may not function identically +// (or be able to be applied to) all architectures. +// Returns an error if an issue was encountered adding the rule. +// There is a bug in library versions below v2.2.1 which can, in some cases, +// cause conditions to be lost when more than one are used. Consequently, +// AddRuleConditionalExact is disabled on library versions lower than v2.2.1 +func (f *ScmpFilter) AddRuleConditionalExact(call ScmpSyscall, action ScmpAction, conds []ScmpCondition) error { + return f.addRuleGeneric(call, action, true, conds) +} + +// ExportPFC output PFC-formatted, human-readable dump of a filter context's +// rules to a file. +// Accepts file to write to (must be open for writing). +// Returns an error if writing to the file fails. +func (f *ScmpFilter) ExportPFC(file *os.File) error { + f.lock.Lock() + defer f.lock.Unlock() + + fd := file.Fd() + + if !f.valid { + return errBadFilter + } + + if retCode := C.seccomp_export_pfc(f.filterCtx, C.int(fd)); retCode != 0 { + return syscall.Errno(-1 * retCode) + } + + return nil +} + +// ExportBPF outputs Berkeley Packet Filter-formatted, kernel-readable dump of a +// filter context's rules to a file. +// Accepts file to write to (must be open for writing). +// Returns an error if writing to the file fails. +func (f *ScmpFilter) ExportBPF(file *os.File) error { + f.lock.Lock() + defer f.lock.Unlock() + + fd := file.Fd() + + if !f.valid { + return errBadFilter + } + + if retCode := C.seccomp_export_bpf(f.filterCtx, C.int(fd)); retCode != 0 { + return syscall.Errno(-1 * retCode) + } + + return nil +} diff --git a/vendor/github.com/seccomp/libseccomp-golang/seccomp_internal.go b/vendor/github.com/seccomp/libseccomp-golang/seccomp_internal.go new file mode 100644 index 000000000..5b6a79ada --- /dev/null +++ b/vendor/github.com/seccomp/libseccomp-golang/seccomp_internal.go @@ -0,0 +1,508 @@ +// +build linux + +// Internal functions for libseccomp Go bindings +// No exported functions + +package seccomp + +import ( + "fmt" + "syscall" +) + +// Unexported C wrapping code - provides the C-Golang interface +// Get the seccomp header in scope +// Need stdlib.h for free() on cstrings + +// #cgo pkg-config: libseccomp +/* +#include +#include + +#if SCMP_VER_MAJOR < 2 +#error Minimum supported version of Libseccomp is v2.2.0 +#elif SCMP_VER_MAJOR == 2 && SCMP_VER_MINOR < 2 +#error Minimum supported version of Libseccomp is v2.2.0 +#endif + +#define ARCH_BAD ~0 + +const uint32_t C_ARCH_BAD = ARCH_BAD; + +#ifndef SCMP_ARCH_PPC +#define SCMP_ARCH_PPC ARCH_BAD +#endif + +#ifndef SCMP_ARCH_PPC64 +#define SCMP_ARCH_PPC64 ARCH_BAD +#endif + +#ifndef SCMP_ARCH_PPC64LE +#define SCMP_ARCH_PPC64LE ARCH_BAD +#endif + +#ifndef SCMP_ARCH_S390 +#define SCMP_ARCH_S390 ARCH_BAD +#endif + +#ifndef SCMP_ARCH_S390X +#define SCMP_ARCH_S390X ARCH_BAD +#endif + +const uint32_t C_ARCH_NATIVE = SCMP_ARCH_NATIVE; +const uint32_t C_ARCH_X86 = SCMP_ARCH_X86; +const uint32_t C_ARCH_X86_64 = SCMP_ARCH_X86_64; +const uint32_t C_ARCH_X32 = SCMP_ARCH_X32; +const uint32_t C_ARCH_ARM = SCMP_ARCH_ARM; +const uint32_t C_ARCH_AARCH64 = SCMP_ARCH_AARCH64; +const uint32_t C_ARCH_MIPS = SCMP_ARCH_MIPS; +const uint32_t C_ARCH_MIPS64 = SCMP_ARCH_MIPS64; +const uint32_t C_ARCH_MIPS64N32 = SCMP_ARCH_MIPS64N32; +const uint32_t C_ARCH_MIPSEL = SCMP_ARCH_MIPSEL; +const uint32_t C_ARCH_MIPSEL64 = SCMP_ARCH_MIPSEL64; +const uint32_t C_ARCH_MIPSEL64N32 = SCMP_ARCH_MIPSEL64N32; +const uint32_t C_ARCH_PPC = SCMP_ARCH_PPC; +const uint32_t C_ARCH_PPC64 = SCMP_ARCH_PPC64; +const uint32_t C_ARCH_PPC64LE = SCMP_ARCH_PPC64LE; +const uint32_t C_ARCH_S390 = SCMP_ARCH_S390; +const uint32_t C_ARCH_S390X = SCMP_ARCH_S390X; + +const uint32_t C_ACT_KILL = SCMP_ACT_KILL; +const uint32_t C_ACT_TRAP = SCMP_ACT_TRAP; +const uint32_t C_ACT_ERRNO = SCMP_ACT_ERRNO(0); +const uint32_t C_ACT_TRACE = SCMP_ACT_TRACE(0); +const uint32_t C_ACT_ALLOW = SCMP_ACT_ALLOW; + +const uint32_t C_ATTRIBUTE_DEFAULT = (uint32_t)SCMP_FLTATR_ACT_DEFAULT; +const uint32_t C_ATTRIBUTE_BADARCH = (uint32_t)SCMP_FLTATR_ACT_BADARCH; +const uint32_t C_ATTRIBUTE_NNP = (uint32_t)SCMP_FLTATR_CTL_NNP; +const uint32_t C_ATTRIBUTE_TSYNC = (uint32_t)SCMP_FLTATR_CTL_TSYNC; + +const int C_CMP_NE = (int)SCMP_CMP_NE; +const int C_CMP_LT = (int)SCMP_CMP_LT; +const int C_CMP_LE = (int)SCMP_CMP_LE; +const int C_CMP_EQ = (int)SCMP_CMP_EQ; +const int C_CMP_GE = (int)SCMP_CMP_GE; +const int C_CMP_GT = (int)SCMP_CMP_GT; +const int C_CMP_MASKED_EQ = (int)SCMP_CMP_MASKED_EQ; + +const int C_VERSION_MAJOR = SCMP_VER_MAJOR; +const int C_VERSION_MINOR = SCMP_VER_MINOR; +const int C_VERSION_MICRO = SCMP_VER_MICRO; + +#if SCMP_VER_MAJOR == 2 && SCMP_VER_MINOR >= 3 +unsigned int get_major_version() +{ + return seccomp_version()->major; +} + +unsigned int get_minor_version() +{ + return seccomp_version()->minor; +} + +unsigned int get_micro_version() +{ + return seccomp_version()->micro; +} +#else +unsigned int get_major_version() +{ + return (unsigned int)C_VERSION_MAJOR; +} + +unsigned int get_minor_version() +{ + return (unsigned int)C_VERSION_MINOR; +} + +unsigned int get_micro_version() +{ + return (unsigned int)C_VERSION_MICRO; +} +#endif + +typedef struct scmp_arg_cmp* scmp_cast_t; + +void* make_arg_cmp_array(unsigned int length) +{ + return calloc(length, sizeof(struct scmp_arg_cmp)); +} + +// Wrapper to add an scmp_arg_cmp struct to an existing arg_cmp array +void add_struct_arg_cmp( + struct scmp_arg_cmp* arr, + unsigned int pos, + unsigned int arg, + int compare, + uint64_t a, + uint64_t b + ) +{ + arr[pos].arg = arg; + arr[pos].op = compare; + arr[pos].datum_a = a; + arr[pos].datum_b = b; + + return; +} +*/ +import "C" + +// Nonexported types +type scmpFilterAttr uint32 + +// Nonexported constants + +const ( + filterAttrActDefault scmpFilterAttr = iota + filterAttrActBadArch scmpFilterAttr = iota + filterAttrNNP scmpFilterAttr = iota + filterAttrTsync scmpFilterAttr = iota +) + +const ( + // An error return from certain libseccomp functions + scmpError C.int = -1 + // Comparison boundaries to check for architecture validity + archStart ScmpArch = ArchNative + archEnd ScmpArch = ArchS390X + // Comparison boundaries to check for action validity + actionStart ScmpAction = ActKill + actionEnd ScmpAction = ActAllow + // Comparison boundaries to check for comparison operator validity + compareOpStart ScmpCompareOp = CompareNotEqual + compareOpEnd ScmpCompareOp = CompareMaskedEqual +) + +var ( + // Error thrown on bad filter context + errBadFilter = fmt.Errorf("filter is invalid or uninitialized") + // Constants representing library major, minor, and micro versions + verMajor = uint(C.get_major_version()) + verMinor = uint(C.get_minor_version()) + verMicro = uint(C.get_micro_version()) +) + +// Nonexported functions + +// Check if library version is greater than or equal to the given one +func checkVersionAbove(major, minor, micro uint) bool { + return (verMajor > major) || + (verMajor == major && verMinor > minor) || + (verMajor == major && verMinor == minor && verMicro >= micro) +} + +// Ensure that the library is supported, i.e. >= 2.2.0. +func ensureSupportedVersion() error { + if !checkVersionAbove(2, 2, 0) { + return VersionError{} + } + return nil +} + +// Filter helpers + +// Filter finalizer - ensure that kernel context for filters is freed +func filterFinalizer(f *ScmpFilter) { + f.Release() +} + +// Get a raw filter attribute +func (f *ScmpFilter) getFilterAttr(attr scmpFilterAttr) (C.uint32_t, error) { + f.lock.Lock() + defer f.lock.Unlock() + + if !f.valid { + return 0x0, errBadFilter + } + + var attribute C.uint32_t + + retCode := C.seccomp_attr_get(f.filterCtx, attr.toNative(), &attribute) + if retCode != 0 { + return 0x0, syscall.Errno(-1 * retCode) + } + + return attribute, nil +} + +// Set a raw filter attribute +func (f *ScmpFilter) setFilterAttr(attr scmpFilterAttr, value C.uint32_t) error { + f.lock.Lock() + defer f.lock.Unlock() + + if !f.valid { + return errBadFilter + } + + retCode := C.seccomp_attr_set(f.filterCtx, attr.toNative(), value) + if retCode != 0 { + return syscall.Errno(-1 * retCode) + } + + return nil +} + +// DOES NOT LOCK OR CHECK VALIDITY +// Assumes caller has already done this +// Wrapper for seccomp_rule_add_... functions +func (f *ScmpFilter) addRuleWrapper(call ScmpSyscall, action ScmpAction, exact bool, length C.uint, cond C.scmp_cast_t) error { + if length != 0 && cond == nil { + return fmt.Errorf("null conditions list, but length is nonzero") + } + + var retCode C.int + if exact { + retCode = C.seccomp_rule_add_exact_array(f.filterCtx, action.toNative(), C.int(call), length, cond) + } else { + retCode = C.seccomp_rule_add_array(f.filterCtx, action.toNative(), C.int(call), length, cond) + } + + if syscall.Errno(-1*retCode) == syscall.EFAULT { + return fmt.Errorf("unrecognized syscall") + } else if syscall.Errno(-1*retCode) == syscall.EPERM { + return fmt.Errorf("requested action matches default action of filter") + } else if syscall.Errno(-1*retCode) == syscall.EINVAL { + return fmt.Errorf("two checks on same syscall argument") + } else if retCode != 0 { + return syscall.Errno(-1 * retCode) + } + + return nil +} + +// Generic add function for filter rules +func (f *ScmpFilter) addRuleGeneric(call ScmpSyscall, action ScmpAction, exact bool, conds []ScmpCondition) error { + f.lock.Lock() + defer f.lock.Unlock() + + if !f.valid { + return errBadFilter + } + + if len(conds) == 0 { + if err := f.addRuleWrapper(call, action, exact, 0, nil); err != nil { + return err + } + } else { + // We don't support conditional filtering in library version v2.1 + if !checkVersionAbove(2, 2, 1) { + return VersionError{ + message: "conditional filtering is not supported", + minimum: "2.2.1", + } + } + + argsArr := C.make_arg_cmp_array(C.uint(len(conds))) + if argsArr == nil { + return fmt.Errorf("error allocating memory for conditions") + } + defer C.free(argsArr) + + for i, cond := range conds { + C.add_struct_arg_cmp(C.scmp_cast_t(argsArr), C.uint(i), + C.uint(cond.Argument), cond.Op.toNative(), + C.uint64_t(cond.Operand1), C.uint64_t(cond.Operand2)) + } + + if err := f.addRuleWrapper(call, action, exact, C.uint(len(conds)), C.scmp_cast_t(argsArr)); err != nil { + return err + } + } + + return nil +} + +// Generic Helpers + +// Helper - Sanitize Arch token input +func sanitizeArch(in ScmpArch) error { + if in < archStart || in > archEnd { + return fmt.Errorf("unrecognized architecture") + } + + if in.toNative() == C.C_ARCH_BAD { + return fmt.Errorf("architecture is not supported on this version of the library") + } + + return nil +} + +func sanitizeAction(in ScmpAction) error { + inTmp := in & 0x0000FFFF + if inTmp < actionStart || inTmp > actionEnd { + return fmt.Errorf("unrecognized action") + } + + if inTmp != ActTrace && inTmp != ActErrno && (in&0xFFFF0000) != 0 { + return fmt.Errorf("highest 16 bits must be zeroed except for Trace and Errno") + } + + return nil +} + +func sanitizeCompareOp(in ScmpCompareOp) error { + if in < compareOpStart || in > compareOpEnd { + return fmt.Errorf("unrecognized comparison operator") + } + + return nil +} + +func archFromNative(a C.uint32_t) (ScmpArch, error) { + switch a { + case C.C_ARCH_X86: + return ArchX86, nil + case C.C_ARCH_X86_64: + return ArchAMD64, nil + case C.C_ARCH_X32: + return ArchX32, nil + case C.C_ARCH_ARM: + return ArchARM, nil + case C.C_ARCH_NATIVE: + return ArchNative, nil + case C.C_ARCH_AARCH64: + return ArchARM64, nil + case C.C_ARCH_MIPS: + return ArchMIPS, nil + case C.C_ARCH_MIPS64: + return ArchMIPS64, nil + case C.C_ARCH_MIPS64N32: + return ArchMIPS64N32, nil + case C.C_ARCH_MIPSEL: + return ArchMIPSEL, nil + case C.C_ARCH_MIPSEL64: + return ArchMIPSEL64, nil + case C.C_ARCH_MIPSEL64N32: + return ArchMIPSEL64N32, nil + case C.C_ARCH_PPC: + return ArchPPC, nil + case C.C_ARCH_PPC64: + return ArchPPC64, nil + case C.C_ARCH_PPC64LE: + return ArchPPC64LE, nil + case C.C_ARCH_S390: + return ArchS390, nil + case C.C_ARCH_S390X: + return ArchS390X, nil + default: + return 0x0, fmt.Errorf("unrecognized architecture") + } +} + +// Only use with sanitized arches, no error handling +func (a ScmpArch) toNative() C.uint32_t { + switch a { + case ArchX86: + return C.C_ARCH_X86 + case ArchAMD64: + return C.C_ARCH_X86_64 + case ArchX32: + return C.C_ARCH_X32 + case ArchARM: + return C.C_ARCH_ARM + case ArchARM64: + return C.C_ARCH_AARCH64 + case ArchMIPS: + return C.C_ARCH_MIPS + case ArchMIPS64: + return C.C_ARCH_MIPS64 + case ArchMIPS64N32: + return C.C_ARCH_MIPS64N32 + case ArchMIPSEL: + return C.C_ARCH_MIPSEL + case ArchMIPSEL64: + return C.C_ARCH_MIPSEL64 + case ArchMIPSEL64N32: + return C.C_ARCH_MIPSEL64N32 + case ArchPPC: + return C.C_ARCH_PPC + case ArchPPC64: + return C.C_ARCH_PPC64 + case ArchPPC64LE: + return C.C_ARCH_PPC64LE + case ArchS390: + return C.C_ARCH_S390 + case ArchS390X: + return C.C_ARCH_S390X + case ArchNative: + return C.C_ARCH_NATIVE + default: + return 0x0 + } +} + +// Only use with sanitized ops, no error handling +func (a ScmpCompareOp) toNative() C.int { + switch a { + case CompareNotEqual: + return C.C_CMP_NE + case CompareLess: + return C.C_CMP_LT + case CompareLessOrEqual: + return C.C_CMP_LE + case CompareEqual: + return C.C_CMP_EQ + case CompareGreaterEqual: + return C.C_CMP_GE + case CompareGreater: + return C.C_CMP_GT + case CompareMaskedEqual: + return C.C_CMP_MASKED_EQ + default: + return 0x0 + } +} + +func actionFromNative(a C.uint32_t) (ScmpAction, error) { + aTmp := a & 0xFFFF + switch a & 0xFFFF0000 { + case C.C_ACT_KILL: + return ActKill, nil + case C.C_ACT_TRAP: + return ActTrap, nil + case C.C_ACT_ERRNO: + return ActErrno.SetReturnCode(int16(aTmp)), nil + case C.C_ACT_TRACE: + return ActTrace.SetReturnCode(int16(aTmp)), nil + case C.C_ACT_ALLOW: + return ActAllow, nil + default: + return 0x0, fmt.Errorf("unrecognized action") + } +} + +// Only use with sanitized actions, no error handling +func (a ScmpAction) toNative() C.uint32_t { + switch a & 0xFFFF { + case ActKill: + return C.C_ACT_KILL + case ActTrap: + return C.C_ACT_TRAP + case ActErrno: + return C.C_ACT_ERRNO | (C.uint32_t(a) >> 16) + case ActTrace: + return C.C_ACT_TRACE | (C.uint32_t(a) >> 16) + case ActAllow: + return C.C_ACT_ALLOW + default: + return 0x0 + } +} + +// Internal only, assumes safe attribute +func (a scmpFilterAttr) toNative() uint32 { + switch a { + case filterAttrActDefault: + return uint32(C.C_ATTRIBUTE_DEFAULT) + case filterAttrActBadArch: + return uint32(C.C_ATTRIBUTE_BADARCH) + case filterAttrNNP: + return uint32(C.C_ATTRIBUTE_NNP) + case filterAttrTsync: + return uint32(C.C_ATTRIBUTE_TSYNC) + default: + return 0x0 + } +} diff --git a/vendor/github.com/Sirupsen/logrus/CHANGELOG.md b/vendor/github.com/sirupsen/logrus/CHANGELOG.md similarity index 59% rename from vendor/github.com/Sirupsen/logrus/CHANGELOG.md rename to vendor/github.com/sirupsen/logrus/CHANGELOG.md index f2c2bc211..1bd1deb29 100644 --- a/vendor/github.com/Sirupsen/logrus/CHANGELOG.md +++ b/vendor/github.com/sirupsen/logrus/CHANGELOG.md @@ -1,3 +1,60 @@ +# 1.0.5 + +* Fix hooks race (#707) +* Fix panic deadlock (#695) + +# 1.0.4 + +* Fix race when adding hooks (#612) +* Fix terminal check in AppEngine (#635) + +# 1.0.3 + +* Replace example files with testable examples + +# 1.0.2 + +* bug: quote non-string values in text formatter (#583) +* Make (*Logger) SetLevel a public method + +# 1.0.1 + +* bug: fix escaping in text formatter (#575) + +# 1.0.0 + +* Officially changed name to lower-case +* bug: colors on Windows 10 (#541) +* bug: fix race in accessing level (#512) + +# 0.11.5 + +* feature: add writer and writerlevel to entry (#372) + +# 0.11.4 + +* bug: fix undefined variable on solaris (#493) + +# 0.11.3 + +* formatter: configure quoting of empty values (#484) +* formatter: configure quoting character (default is `"`) (#484) +* bug: fix not importing io correctly in non-linux environments (#481) + +# 0.11.2 + +* bug: fix windows terminal detection (#476) + +# 0.11.1 + +* bug: fix tty detection with custom out (#471) + +# 0.11.0 + +* performance: Use bufferpool to allocate (#370) +* terminal: terminal detection for app-engine (#343) +* feature: exit handler (#375) + # 0.10.0 * feature: Add a test hook (#180) diff --git a/vendor/github.com/Sirupsen/logrus/LICENSE b/vendor/github.com/sirupsen/logrus/LICENSE similarity index 100% rename from vendor/github.com/Sirupsen/logrus/LICENSE rename to vendor/github.com/sirupsen/logrus/LICENSE diff --git a/vendor/github.com/Sirupsen/logrus/README.md b/vendor/github.com/sirupsen/logrus/README.md similarity index 63% rename from vendor/github.com/Sirupsen/logrus/README.md rename to vendor/github.com/sirupsen/logrus/README.md index 6e1721a74..072e99be3 100644 --- a/vendor/github.com/Sirupsen/logrus/README.md +++ b/vendor/github.com/sirupsen/logrus/README.md @@ -1,11 +1,24 @@ -# Logrus :walrus: [![Build Status](https://travis-ci.org/Sirupsen/logrus.svg?branch=master)](https://travis-ci.org/Sirupsen/logrus) [![GoDoc](https://godoc.org/github.com/Sirupsen/logrus?status.svg)](https://godoc.org/github.com/Sirupsen/logrus) +# Logrus :walrus: [![Build Status](https://travis-ci.org/sirupsen/logrus.svg?branch=master)](https://travis-ci.org/sirupsen/logrus) [![GoDoc](https://godoc.org/github.com/sirupsen/logrus?status.svg)](https://godoc.org/github.com/sirupsen/logrus) Logrus is a structured logger for Go (golang), completely API compatible with -the standard library logger. [Godoc][godoc]. **Please note the Logrus API is not -yet stable (pre 1.0). Logrus itself is completely stable and has been used in -many large deployments. The core API is unlikely to change much but please -version control your Logrus to make sure you aren't fetching latest `master` on -every build.** +the standard library logger. + +**Seeing weird case-sensitive problems?** It's in the past been possible to +import Logrus as both upper- and lower-case. Due to the Go package environment, +this caused issues in the community and we needed a standard. Some environments +experienced problems with the upper-case variant, so the lower-case was decided. +Everything using `logrus` will need to use the lower-case: +`github.com/sirupsen/logrus`. Any package that isn't, should be changed. + +To fix Glide, see [these +comments](https://github.com/sirupsen/logrus/issues/553#issuecomment-306591437). +For an in-depth explanation of the casing issue, see [this +comment](https://github.com/sirupsen/logrus/issues/570#issuecomment-313933276). + +**Are you interested in assisting in maintaining Logrus?** Currently I have a +lot of obligations, and I am unable to provide Logrus with the maintainership it +needs. If you'd like to help, please reach out to me at `simon at author's +username dot com`. Nicely color-coded in development (when a TTY is attached, otherwise just plain text): @@ -46,6 +59,12 @@ time="2015-03-26T01:27:38-04:00" level=fatal msg="The ice breaks!" err=&{0x20822 exit status 1 ``` +#### Case-sensitivity + +The organization's name was changed to lower-case--and this will not be changed +back. If you are getting import conflicts due to case sensitivity, please use +the lower-case import: `github.com/sirupsen/logrus`. + #### Example The simplest way to use Logrus is simply the package-level exported logger: @@ -54,7 +73,7 @@ The simplest way to use Logrus is simply the package-level exported logger: package main import ( - log "github.com/Sirupsen/logrus" + log "github.com/sirupsen/logrus" ) func main() { @@ -65,7 +84,7 @@ func main() { ``` Note that it's completely api-compatible with the stdlib logger, so you can -replace your `log` imports everywhere with `log "github.com/Sirupsen/logrus"` +replace your `log` imports everywhere with `log "github.com/sirupsen/logrus"` and you'll now have the flexibility of Logrus. You can customize it all you want: @@ -74,15 +93,16 @@ package main import ( "os" - log "github.com/Sirupsen/logrus" + log "github.com/sirupsen/logrus" ) func init() { // Log as JSON instead of the default ASCII formatter. log.SetFormatter(&log.JSONFormatter{}) - // Output to stderr instead of stdout, could also be a file. - log.SetOutput(os.Stderr) + // Output to stdout instead of the default stderr + // Can be any io.Writer, see below for File example + log.SetOutput(os.Stdout) // Only log the warning severity or above. log.SetLevel(log.WarnLevel) @@ -123,7 +143,8 @@ application, you can also create an instance of the `logrus` Logger: package main import ( - "github.com/Sirupsen/logrus" + "os" + "github.com/sirupsen/logrus" ) // Create a new instance of the logger. You can have any number of instances. @@ -132,7 +153,15 @@ var log = logrus.New() func main() { // The API for setting attributes is a little different than the package level // exported logger. See Godoc. - log.Out = os.Stderr + log.Out = os.Stdout + + // You could set this to any `io.Writer` such as a file + // file, err := os.OpenFile("logrus.log", os.O_CREATE|os.O_WRONLY, 0666) + // if err == nil { + // log.Out = file + // } else { + // log.Info("Failed to log to file, using default stderr") + // } log.WithFields(logrus.Fields{ "animal": "walrus", @@ -143,7 +172,7 @@ func main() { #### Fields -Logrus encourages careful, structured logging though logging fields instead of +Logrus encourages careful, structured logging through logging fields instead of long, unparseable error messages. For example, instead of: `log.Fatalf("Failed to send event %s to topic %s with key %d")`, you should log the much more discoverable: @@ -165,6 +194,20 @@ In general, with Logrus using any of the `printf`-family functions should be seen as a hint you should add a field, however, you can still use the `printf`-family functions with Logrus. +#### Default Fields + +Often it's helpful to have fields _always_ attached to log statements in an +application or parts of one. For example, you may want to always log the +`request_id` and `user_ip` in the context of a request. Instead of writing +`log.WithFields(log.Fields{"request_id": request_id, "user_ip": user_ip})` on +every line, you can create a `logrus.Entry` to pass around instead: + +```go +requestLogger := log.WithFields(log.Fields{"request_id": request_id, "user_ip": user_ip}) +requestLogger.Info("something happened on that request") # will log request_id and user_ip +requestLogger.Warn("something not great happened") +``` + #### Hooks You can add hooks for logging levels. For example to send errors to an exception @@ -176,9 +219,9 @@ Logrus comes with [built-in hooks](hooks/). Add those, or your custom hook, in ```go import ( - log "github.com/Sirupsen/logrus" - "gopkg.in/gemnasium/logrus-airbrake-hook.v2" // the package is named "aibrake" - logrus_syslog "github.com/Sirupsen/logrus/hooks/syslog" + log "github.com/sirupsen/logrus" + "gopkg.in/gemnasium/logrus-airbrake-hook.v2" // the package is named "airbrake" + logrus_syslog "github.com/sirupsen/logrus/hooks/syslog" "log/syslog" ) @@ -198,34 +241,7 @@ func init() { ``` Note: Syslog hook also support connecting to local syslog (Ex. "/dev/log" or "/var/run/syslog" or "/var/run/log"). For the detail, please check the [syslog hook README](hooks/syslog/README.md). -| Hook | Description | -| ----- | ----------- | -| [Airbrake](https://github.com/gemnasium/logrus-airbrake-hook) | Send errors to the Airbrake API V3. Uses the official [`gobrake`](https://github.com/airbrake/gobrake) behind the scenes. | -| [Airbrake "legacy"](https://github.com/gemnasium/logrus-airbrake-legacy-hook) | Send errors to an exception tracking service compatible with the Airbrake API V2. Uses [`airbrake-go`](https://github.com/tobi/airbrake-go) behind the scenes. | -| [Papertrail](https://github.com/polds/logrus-papertrail-hook) | Send errors to the [Papertrail](https://papertrailapp.com) hosted logging service via UDP. | -| [Syslog](https://github.com/Sirupsen/logrus/blob/master/hooks/syslog/syslog.go) | Send errors to remote syslog server. Uses standard library `log/syslog` behind the scenes. | -| [Bugsnag](https://github.com/Shopify/logrus-bugsnag/blob/master/bugsnag.go) | Send errors to the Bugsnag exception tracking service. | -| [Sentry](https://github.com/evalphobia/logrus_sentry) | Send errors to the Sentry error logging and aggregation service. | -| [Hiprus](https://github.com/nubo/hiprus) | Send errors to a channel in hipchat. | -| [Logrusly](https://github.com/sebest/logrusly) | Send logs to [Loggly](https://www.loggly.com/) | -| [Slackrus](https://github.com/johntdyer/slackrus) | Hook for Slack chat. | -| [Journalhook](https://github.com/wercker/journalhook) | Hook for logging to `systemd-journald` | -| [Graylog](https://github.com/gemnasium/logrus-graylog-hook) | Hook for logging to [Graylog](http://graylog2.org/) | -| [Raygun](https://github.com/squirkle/logrus-raygun-hook) | Hook for logging to [Raygun.io](http://raygun.io/) | -| [LFShook](https://github.com/rifflock/lfshook) | Hook for logging to the local filesystem | -| [Honeybadger](https://github.com/agonzalezro/logrus_honeybadger) | Hook for sending exceptions to Honeybadger | -| [Mail](https://github.com/zbindenren/logrus_mail) | Hook for sending exceptions via mail | -| [Rollrus](https://github.com/heroku/rollrus) | Hook for sending errors to rollbar | -| [Fluentd](https://github.com/evalphobia/logrus_fluent) | Hook for logging to fluentd | -| [Mongodb](https://github.com/weekface/mgorus) | Hook for logging to mongodb | -| [InfluxDB](https://github.com/Abramovic/logrus_influxdb) | Hook for logging to influxdb | -| [Octokit](https://github.com/dorajistyle/logrus-octokit-hook) | Hook for logging to github via octokit | -| [DeferPanic](https://github.com/deferpanic/dp-logrus) | Hook for logging to DeferPanic | -| [Redis-Hook](https://github.com/rogierlommers/logrus-redis-hook) | Hook for logging to a ELK stack (through Redis) | -| [Amqp-Hook](https://github.com/vladoatanasov/logrus_amqp) | Hook for logging to Amqp broker (Like RabbitMQ) | -| [KafkaLogrus](https://github.com/goibibo/KafkaLogrus) | Hook for logging to kafka | -| [Typetalk](https://github.com/dragon3/logrus-typetalk-hook) | Hook for logging to [Typetalk](https://www.typetalk.in/) | -| [ElasticSearch](https://github.com/sohlich/elogrus) | Hook for logging to ElasticSearch| +A list of currently known of service hook can be found in this wiki [page](https://github.com/sirupsen/logrus/wiki/Hooks) #### Level logging @@ -275,7 +291,7 @@ could do: ```go import ( - log "github.com/Sirupsen/logrus" + log "github.com/sirupsen/logrus" ) init() { @@ -302,16 +318,18 @@ The built-in logging formatters are: without colors. * *Note:* to force colored output when there is no TTY, set the `ForceColors` field to `true`. To force no colored output even if there is a TTY set the - `DisableColors` field to `true` + `DisableColors` field to `true`. For Windows, see + [github.com/mattn/go-colorable](https://github.com/mattn/go-colorable). + * When colors are enabled, levels are truncated to 4 characters by default. To disable + truncation set the `DisableLevelTruncation` field to `true`. + * All options are listed in the [generated docs](https://godoc.org/github.com/sirupsen/logrus#TextFormatter). * `logrus.JSONFormatter`. Logs fields as JSON. -* `logrus/formatters/logstash.LogstashFormatter`. Logs fields as [Logstash](http://logstash.net) Events. - - ```go - logrus.SetFormatter(&logstash.LogstashFormatter{Type: "application_name"}) - ``` + * All options are listed in the [generated docs](https://godoc.org/github.com/sirupsen/logrus#JSONFormatter). Third party logging formatters: +* [`FluentdFormatter`](https://github.com/joonix/log). Formats entries that can be parsed by Kubernetes and Google Container Engine. +* [`logstash`](https://github.com/bshuster-repo/logrus-logstash-hook). Logs fields as [Logstash](http://logstash.net) Events. * [`prefixed`](https://github.com/x-cray/logrus-prefixed-formatter). Displays log entry source along with alternative layout. * [`zalgo`](https://github.com/aybabtme/logzalgo). Invoking the P͉̫o̳̼̊w̖͈̰͎e̬͔̭͂r͚̼̹̲ ̫͓͉̳͈ō̠͕͖̚f̝͍̠ ͕̲̞͖͑Z̖̫̤̫ͪa͉̬͈̗l͖͎g̳̥o̰̥̅!̣͔̲̻͊̄ ̙̘̦̹̦. @@ -356,6 +374,18 @@ srv := http.Server{ Each line written to that writer will be printed the usual way, using formatters and hooks. The level for those entries is `info`. +This means that we can override the standard library logger easily: + +```go +logger := logrus.New() +logger.Formatter = &logrus.JSONFormatter{} + +// Use logrus for standard log output +// Note that `log` here references stdlib's log +// Not logrus imported under the name `log`. +log.SetOutput(logger.Writer()) +``` + #### Rotation Log rotation is not provided with Logrus. Log rotation should be done by an @@ -367,6 +397,7 @@ entries. It should not be a feature of the application-level logger. | Tool | Description | | ---- | ----------- | |[Logrus Mate](https://github.com/gogap/logrus_mate)|Logrus mate is a tool for Logrus to manage loggers, you can initial logger's level, hook and formatter by config file, the logger will generated with different config at different environment.| +|[Logrus Viper Helper](https://github.com/heirko/go-contrib/tree/master/logrusHelper)|An Helper around Logrus to wrap with spf13/Viper to load configuration with fangs! And to simplify Logrus configuration use some behavior of [Logrus Mate](https://github.com/gogap/logrus_mate). [sample](https://github.com/heirko/iris-contrib/blob/master/middleware/logrus-logger/example) | #### Testing @@ -376,13 +407,55 @@ Logrus has a built in facility for asserting the presence of log messages. This * a test logger (`test.NewNullLogger`) that just records log messages (and does not output any): ```go -logger, hook := NewNullLogger() -logger.Error("Hello error") +import( + "github.com/sirupsen/logrus" + "github.com/sirupsen/logrus/hooks/test" + "github.com/stretchr/testify/assert" + "testing" +) -assert.Equal(1, len(hook.Entries)) -assert.Equal(logrus.ErrorLevel, hook.LastEntry().Level) -assert.Equal("Hello error", hook.LastEntry().Message) +func TestSomething(t*testing.T){ + logger, hook := test.NewNullLogger() + logger.Error("Helloerror") -hook.Reset() -assert.Nil(hook.LastEntry()) + assert.Equal(t, 1, len(hook.Entries)) + assert.Equal(t, logrus.ErrorLevel, hook.LastEntry().Level) + assert.Equal(t, "Helloerror", hook.LastEntry().Message) + + hook.Reset() + assert.Nil(t, hook.LastEntry()) +} ``` + +#### Fatal handlers + +Logrus can register one or more functions that will be called when any `fatal` +level message is logged. The registered handlers will be executed before +logrus performs a `os.Exit(1)`. This behavior may be helpful if callers need +to gracefully shutdown. Unlike a `panic("Something went wrong...")` call which can be intercepted with a deferred `recover` a call to `os.Exit(1)` can not be intercepted. + +``` +... +handler := func() { + // gracefully shutdown something... +} +logrus.RegisterExitHandler(handler) +... +``` + +#### Thread safety + +By default, Logger is protected by a mutex for concurrent writes. The mutex is held when calling hooks and writing logs. +If you are sure such locking is not needed, you can call logger.SetNoLock() to disable the locking. + +Situation when locking is not needed includes: + +* You have no hooks registered, or hooks calling is already thread-safe. + +* Writing to logger.Out is already thread-safe, for example: + + 1) logger.Out is protected by locks. + + 2) logger.Out is a os.File handler opened with `O_APPEND` flag, and every write is smaller than 4k. (This allow multi-thread/multi-process writing) + + (Refer to http://www.notthewizard.com/2014/06/17/are-files-appends-really-atomic/) diff --git a/vendor/github.com/sirupsen/logrus/alt_exit.go b/vendor/github.com/sirupsen/logrus/alt_exit.go new file mode 100644 index 000000000..8af90637a --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/alt_exit.go @@ -0,0 +1,64 @@ +package logrus + +// The following code was sourced and modified from the +// https://github.com/tebeka/atexit package governed by the following license: +// +// Copyright (c) 2012 Miki Tebeka . +// +// Permission is hereby granted, free of charge, to any person obtaining a copy of +// this software and associated documentation files (the "Software"), to deal in +// the Software without restriction, including without limitation the rights to +// use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of +// the Software, and to permit persons to whom the Software is furnished to do so, +// subject to the following conditions: +// +// The above copyright notice and this permission notice shall be included in all +// copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS +// FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR +// COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER +// IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN +// CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. + +import ( + "fmt" + "os" +) + +var handlers = []func(){} + +func runHandler(handler func()) { + defer func() { + if err := recover(); err != nil { + fmt.Fprintln(os.Stderr, "Error: Logrus exit handler error:", err) + } + }() + + handler() +} + +func runHandlers() { + for _, handler := range handlers { + runHandler(handler) + } +} + +// Exit runs all the Logrus atexit handlers and then terminates the program using os.Exit(code) +func Exit(code int) { + runHandlers() + os.Exit(code) +} + +// RegisterExitHandler adds a Logrus Exit handler, call logrus.Exit to invoke +// all handlers. The handlers will also be invoked when any Fatal log entry is +// made. +// +// This method is useful when a caller wishes to use logrus to log a fatal +// message but also needs to gracefully shutdown. An example usecase could be +// closing database connections, or sending a alert that the application is +// closing. +func RegisterExitHandler(handler func()) { + handlers = append(handlers, handler) +} diff --git a/vendor/github.com/sirupsen/logrus/appveyor.yml b/vendor/github.com/sirupsen/logrus/appveyor.yml new file mode 100644 index 000000000..df9d65c3a --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/appveyor.yml @@ -0,0 +1,14 @@ +version: "{build}" +platform: x64 +clone_folder: c:\gopath\src\github.com\sirupsen\logrus +environment: + GOPATH: c:\gopath +branches: + only: + - master +install: + - set PATH=%GOPATH%\bin;c:\go\bin;%PATH% + - go version +build_script: + - go get -t + - go test diff --git a/vendor/github.com/Sirupsen/logrus/doc.go b/vendor/github.com/sirupsen/logrus/doc.go similarity index 83% rename from vendor/github.com/Sirupsen/logrus/doc.go rename to vendor/github.com/sirupsen/logrus/doc.go index dddd5f877..da67aba06 100644 --- a/vendor/github.com/Sirupsen/logrus/doc.go +++ b/vendor/github.com/sirupsen/logrus/doc.go @@ -7,7 +7,7 @@ The simplest way to use Logrus is simply the package-level exported logger: package main import ( - log "github.com/Sirupsen/logrus" + log "github.com/sirupsen/logrus" ) func main() { @@ -21,6 +21,6 @@ The simplest way to use Logrus is simply the package-level exported logger: Output: time="2015-09-07T08:48:33Z" level=info msg="A walrus appears" animal=walrus number=1 size=10 -For a full guide visit https://github.com/Sirupsen/logrus +For a full guide visit https://github.com/sirupsen/logrus */ package logrus diff --git a/vendor/github.com/Sirupsen/logrus/entry.go b/vendor/github.com/sirupsen/logrus/entry.go similarity index 68% rename from vendor/github.com/Sirupsen/logrus/entry.go rename to vendor/github.com/sirupsen/logrus/entry.go index 89e966e7b..8c6a389ee 100644 --- a/vendor/github.com/Sirupsen/logrus/entry.go +++ b/vendor/github.com/sirupsen/logrus/entry.go @@ -3,11 +3,21 @@ package logrus import ( "bytes" "fmt" - "io" "os" + "sync" "time" ) +var bufferPool *sync.Pool + +func init() { + bufferPool = &sync.Pool{ + New: func() interface{} { + return new(bytes.Buffer) + }, + } +} + // Defines the key when adding errors using WithError. var ErrorKey = "error" @@ -25,35 +35,33 @@ type Entry struct { Time time.Time // Level the log entry was logged at: Debug, Info, Warn, Error, Fatal or Panic + // This field will be set on entry firing and the value will be equal to the one in Logger struct field. Level Level // Message passed to Debug, Info, Warn, Error, Fatal or Panic Message string + + // When formatter is called in entry.log(), an Buffer may be set to entry + Buffer *bytes.Buffer } func NewEntry(logger *Logger) *Entry { return &Entry{ Logger: logger, - // Default is three fields, give a little extra room + // Default is five fields, give a little extra room Data: make(Fields, 5), } } -// Returns a reader for the entry, which is a proxy to the formatter. -func (entry *Entry) Reader() (*bytes.Buffer, error) { - serialized, err := entry.Logger.Formatter.Format(entry) - return bytes.NewBuffer(serialized), err -} - // Returns the string representation from the reader and ultimately the // formatter. func (entry *Entry) String() (string, error) { - reader, err := entry.Reader() + serialized, err := entry.Logger.Formatter.Format(entry) if err != nil { return "", err } - - return reader.String(), err + str := string(serialized) + return str, nil } // Add an error as single field (using the key defined in ErrorKey) to the Entry. @@ -75,36 +83,41 @@ func (entry *Entry) WithFields(fields Fields) *Entry { for k, v := range fields { data[k] = v } - return &Entry{Logger: entry.Logger, Data: data} + return &Entry{Logger: entry.Logger, Data: data, Time: entry.Time} +} + +// Overrides the time of the Entry. +func (entry *Entry) WithTime(t time.Time) *Entry { + return &Entry{Logger: entry.Logger, Data: entry.Data, Time: t} } // This function is not declared with a pointer value because otherwise // race conditions will occur when using multiple goroutines func (entry Entry) log(level Level, msg string) { - entry.Time = time.Now() + var buffer *bytes.Buffer + + // Default to now, but allow users to override if they want. + // + // We don't have to worry about polluting future calls to Entry#log() + // with this assignment because this function is declared with a + // non-pointer receiver. + if entry.Time.IsZero() { + entry.Time = time.Now() + } + entry.Level = level entry.Message = msg - if err := entry.Logger.Hooks.Fire(level, &entry); err != nil { - entry.Logger.mu.Lock() - fmt.Fprintf(os.Stderr, "Failed to fire hook: %v\n", err) - entry.Logger.mu.Unlock() - } + entry.fireHooks() - reader, err := entry.Reader() - if err != nil { - entry.Logger.mu.Lock() - fmt.Fprintf(os.Stderr, "Failed to obtain reader, %v\n", err) - entry.Logger.mu.Unlock() - } + buffer = bufferPool.Get().(*bytes.Buffer) + buffer.Reset() + defer bufferPool.Put(buffer) + entry.Buffer = buffer - entry.Logger.mu.Lock() - defer entry.Logger.mu.Unlock() + entry.write() - _, err = io.Copy(entry.Logger.Out, reader) - if err != nil { - fmt.Fprintf(os.Stderr, "Failed to write to log, %v\n", err) - } + entry.Buffer = nil // To avoid Entry#log() returning a value that only would make sense for // panic() to use in Entry#Panic(), we avoid the allocation by checking @@ -114,8 +127,31 @@ func (entry Entry) log(level Level, msg string) { } } +func (entry *Entry) fireHooks() { + entry.Logger.mu.Lock() + defer entry.Logger.mu.Unlock() + err := entry.Logger.Hooks.Fire(entry.Level, entry) + if err != nil { + fmt.Fprintf(os.Stderr, "Failed to fire hook: %v\n", err) + } +} + +func (entry *Entry) write() { + entry.Logger.mu.Lock() + defer entry.Logger.mu.Unlock() + serialized, err := entry.Logger.Formatter.Format(entry) + if err != nil { + fmt.Fprintf(os.Stderr, "Failed to obtain reader, %v\n", err) + } else { + _, err = entry.Logger.Out.Write(serialized) + if err != nil { + fmt.Fprintf(os.Stderr, "Failed to write to log, %v\n", err) + } + } +} + func (entry *Entry) Debug(args ...interface{}) { - if entry.Logger.Level >= DebugLevel { + if entry.Logger.IsLevelEnabled(DebugLevel) { entry.log(DebugLevel, fmt.Sprint(args...)) } } @@ -125,13 +161,13 @@ func (entry *Entry) Print(args ...interface{}) { } func (entry *Entry) Info(args ...interface{}) { - if entry.Logger.Level >= InfoLevel { + if entry.Logger.IsLevelEnabled(InfoLevel) { entry.log(InfoLevel, fmt.Sprint(args...)) } } func (entry *Entry) Warn(args ...interface{}) { - if entry.Logger.Level >= WarnLevel { + if entry.Logger.IsLevelEnabled(WarnLevel) { entry.log(WarnLevel, fmt.Sprint(args...)) } } @@ -141,20 +177,20 @@ func (entry *Entry) Warning(args ...interface{}) { } func (entry *Entry) Error(args ...interface{}) { - if entry.Logger.Level >= ErrorLevel { + if entry.Logger.IsLevelEnabled(ErrorLevel) { entry.log(ErrorLevel, fmt.Sprint(args...)) } } func (entry *Entry) Fatal(args ...interface{}) { - if entry.Logger.Level >= FatalLevel { + if entry.Logger.IsLevelEnabled(FatalLevel) { entry.log(FatalLevel, fmt.Sprint(args...)) } - os.Exit(1) + Exit(1) } func (entry *Entry) Panic(args ...interface{}) { - if entry.Logger.Level >= PanicLevel { + if entry.Logger.IsLevelEnabled(PanicLevel) { entry.log(PanicLevel, fmt.Sprint(args...)) } panic(fmt.Sprint(args...)) @@ -163,13 +199,13 @@ func (entry *Entry) Panic(args ...interface{}) { // Entry Printf family functions func (entry *Entry) Debugf(format string, args ...interface{}) { - if entry.Logger.Level >= DebugLevel { + if entry.Logger.IsLevelEnabled(DebugLevel) { entry.Debug(fmt.Sprintf(format, args...)) } } func (entry *Entry) Infof(format string, args ...interface{}) { - if entry.Logger.Level >= InfoLevel { + if entry.Logger.IsLevelEnabled(InfoLevel) { entry.Info(fmt.Sprintf(format, args...)) } } @@ -179,7 +215,7 @@ func (entry *Entry) Printf(format string, args ...interface{}) { } func (entry *Entry) Warnf(format string, args ...interface{}) { - if entry.Logger.Level >= WarnLevel { + if entry.Logger.IsLevelEnabled(WarnLevel) { entry.Warn(fmt.Sprintf(format, args...)) } } @@ -189,20 +225,20 @@ func (entry *Entry) Warningf(format string, args ...interface{}) { } func (entry *Entry) Errorf(format string, args ...interface{}) { - if entry.Logger.Level >= ErrorLevel { + if entry.Logger.IsLevelEnabled(ErrorLevel) { entry.Error(fmt.Sprintf(format, args...)) } } func (entry *Entry) Fatalf(format string, args ...interface{}) { - if entry.Logger.Level >= FatalLevel { + if entry.Logger.IsLevelEnabled(FatalLevel) { entry.Fatal(fmt.Sprintf(format, args...)) } - os.Exit(1) + Exit(1) } func (entry *Entry) Panicf(format string, args ...interface{}) { - if entry.Logger.Level >= PanicLevel { + if entry.Logger.IsLevelEnabled(PanicLevel) { entry.Panic(fmt.Sprintf(format, args...)) } } @@ -210,13 +246,13 @@ func (entry *Entry) Panicf(format string, args ...interface{}) { // Entry Println family functions func (entry *Entry) Debugln(args ...interface{}) { - if entry.Logger.Level >= DebugLevel { + if entry.Logger.IsLevelEnabled(DebugLevel) { entry.Debug(entry.sprintlnn(args...)) } } func (entry *Entry) Infoln(args ...interface{}) { - if entry.Logger.Level >= InfoLevel { + if entry.Logger.IsLevelEnabled(InfoLevel) { entry.Info(entry.sprintlnn(args...)) } } @@ -226,7 +262,7 @@ func (entry *Entry) Println(args ...interface{}) { } func (entry *Entry) Warnln(args ...interface{}) { - if entry.Logger.Level >= WarnLevel { + if entry.Logger.IsLevelEnabled(WarnLevel) { entry.Warn(entry.sprintlnn(args...)) } } @@ -236,20 +272,20 @@ func (entry *Entry) Warningln(args ...interface{}) { } func (entry *Entry) Errorln(args ...interface{}) { - if entry.Logger.Level >= ErrorLevel { + if entry.Logger.IsLevelEnabled(ErrorLevel) { entry.Error(entry.sprintlnn(args...)) } } func (entry *Entry) Fatalln(args ...interface{}) { - if entry.Logger.Level >= FatalLevel { + if entry.Logger.IsLevelEnabled(FatalLevel) { entry.Fatal(entry.sprintlnn(args...)) } - os.Exit(1) + Exit(1) } func (entry *Entry) Panicln(args ...interface{}) { - if entry.Logger.Level >= PanicLevel { + if entry.Logger.IsLevelEnabled(PanicLevel) { entry.Panic(entry.sprintlnn(args...)) } } diff --git a/vendor/github.com/Sirupsen/logrus/exported.go b/vendor/github.com/sirupsen/logrus/exported.go similarity index 82% rename from vendor/github.com/Sirupsen/logrus/exported.go rename to vendor/github.com/sirupsen/logrus/exported.go index 9a0120ac1..fb2a7a1f0 100644 --- a/vendor/github.com/Sirupsen/logrus/exported.go +++ b/vendor/github.com/sirupsen/logrus/exported.go @@ -2,6 +2,7 @@ package logrus import ( "io" + "time" ) var ( @@ -15,37 +16,32 @@ func StandardLogger() *Logger { // SetOutput sets the standard logger output. func SetOutput(out io.Writer) { - std.mu.Lock() - defer std.mu.Unlock() - std.Out = out + std.SetOutput(out) } // SetFormatter sets the standard logger formatter. func SetFormatter(formatter Formatter) { - std.mu.Lock() - defer std.mu.Unlock() - std.Formatter = formatter + std.SetFormatter(formatter) } // SetLevel sets the standard logger level. func SetLevel(level Level) { - std.mu.Lock() - defer std.mu.Unlock() - std.Level = level + std.SetLevel(level) } // GetLevel returns the standard logger level. func GetLevel() Level { - std.mu.Lock() - defer std.mu.Unlock() - return std.Level + return std.GetLevel() +} + +// IsLevelEnabled checks if the log level of the standard logger is greater than the level param +func IsLevelEnabled(level Level) bool { + return std.IsLevelEnabled(level) } // AddHook adds a hook to the standard logger hooks. func AddHook(hook Hook) { - std.mu.Lock() - defer std.mu.Unlock() - std.Hooks.Add(hook) + std.AddHook(hook) } // WithError creates an entry from the standard logger and adds an error to it, using the value defined in ErrorKey as key. @@ -72,6 +68,15 @@ func WithFields(fields Fields) *Entry { return std.WithFields(fields) } +// WithTime creats an entry from the standard logger and overrides the time of +// logs generated with it. +// +// Note that it doesn't log until you call Debug, Print, Info, Warn, Fatal +// or Panic on the Entry it returns. +func WithTime(t time.Time) *Entry { + return std.WithTime(t) +} + // Debug logs a message at level Debug on the standard logger. func Debug(args ...interface{}) { std.Debug(args...) @@ -107,7 +112,7 @@ func Panic(args ...interface{}) { std.Panic(args...) } -// Fatal logs a message at level Fatal on the standard logger. +// Fatal logs a message at level Fatal on the standard logger then the process will exit with status set to 1. func Fatal(args ...interface{}) { std.Fatal(args...) } @@ -147,7 +152,7 @@ func Panicf(format string, args ...interface{}) { std.Panicf(format, args...) } -// Fatalf logs a message at level Fatal on the standard logger. +// Fatalf logs a message at level Fatal on the standard logger then the process will exit with status set to 1. func Fatalf(format string, args ...interface{}) { std.Fatalf(format, args...) } @@ -187,7 +192,7 @@ func Panicln(args ...interface{}) { std.Panicln(args...) } -// Fatalln logs a message at level Fatal on the standard logger. +// Fatalln logs a message at level Fatal on the standard logger then the process will exit with status set to 1. func Fatalln(args ...interface{}) { std.Fatalln(args...) } diff --git a/vendor/github.com/Sirupsen/logrus/formatter.go b/vendor/github.com/sirupsen/logrus/formatter.go similarity index 69% rename from vendor/github.com/Sirupsen/logrus/formatter.go rename to vendor/github.com/sirupsen/logrus/formatter.go index 104d689f1..83c74947b 100644 --- a/vendor/github.com/Sirupsen/logrus/formatter.go +++ b/vendor/github.com/sirupsen/logrus/formatter.go @@ -2,7 +2,7 @@ package logrus import "time" -const DefaultTimestampFormat = time.RFC3339 +const defaultTimestampFormat = time.RFC3339 // The Formatter interface is used to implement a custom Formatter. It takes an // `Entry`. It exposes all the fields, including the default ones: @@ -30,19 +30,22 @@ type Formatter interface { // // It's not exported because it's still using Data in an opinionated way. It's to // avoid code duplication between the two default formatters. -func prefixFieldClashes(data Fields) { - _, ok := data["time"] - if ok { - data["fields.time"] = data["time"] +func prefixFieldClashes(data Fields, fieldMap FieldMap) { + timeKey := fieldMap.resolve(FieldKeyTime) + if t, ok := data[timeKey]; ok { + data["fields."+timeKey] = t + delete(data, timeKey) } - _, ok = data["msg"] - if ok { - data["fields.msg"] = data["msg"] + msgKey := fieldMap.resolve(FieldKeyMsg) + if m, ok := data[msgKey]; ok { + data["fields."+msgKey] = m + delete(data, msgKey) } - _, ok = data["level"] - if ok { - data["fields.level"] = data["level"] + levelKey := fieldMap.resolve(FieldKeyLevel) + if l, ok := data[levelKey]; ok { + data["fields."+levelKey] = l + delete(data, levelKey) } } diff --git a/vendor/github.com/Sirupsen/logrus/hooks.go b/vendor/github.com/sirupsen/logrus/hooks.go similarity index 100% rename from vendor/github.com/Sirupsen/logrus/hooks.go rename to vendor/github.com/sirupsen/logrus/hooks.go diff --git a/vendor/github.com/sirupsen/logrus/json_formatter.go b/vendor/github.com/sirupsen/logrus/json_formatter.go new file mode 100644 index 000000000..b13f0cda3 --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/json_formatter.go @@ -0,0 +1,96 @@ +package logrus + +import ( + "bytes" + "encoding/json" + "fmt" +) + +type fieldKey string + +// FieldMap allows customization of the key names for default fields. +type FieldMap map[fieldKey]string + +// Default key names for the default fields +const ( + FieldKeyMsg = "msg" + FieldKeyLevel = "level" + FieldKeyTime = "time" +) + +func (f FieldMap) resolve(key fieldKey) string { + if k, ok := f[key]; ok { + return k + } + + return string(key) +} + +// JSONFormatter formats logs into parsable json +type JSONFormatter struct { + // TimestampFormat sets the format used for marshaling timestamps. + TimestampFormat string + + // DisableTimestamp allows disabling automatic timestamps in output + DisableTimestamp bool + + // DataKey allows users to put all the log entry parameters into a nested dictionary at a given key. + DataKey string + + // FieldMap allows users to customize the names of keys for default fields. + // As an example: + // formatter := &JSONFormatter{ + // FieldMap: FieldMap{ + // FieldKeyTime: "@timestamp", + // FieldKeyLevel: "@level", + // FieldKeyMsg: "@message", + // }, + // } + FieldMap FieldMap +} + +// Format renders a single log entry +func (f *JSONFormatter) Format(entry *Entry) ([]byte, error) { + data := make(Fields, len(entry.Data)+3) + for k, v := range entry.Data { + switch v := v.(type) { + case error: + // Otherwise errors are ignored by `encoding/json` + // https://github.com/sirupsen/logrus/issues/137 + data[k] = v.Error() + default: + data[k] = v + } + } + + if f.DataKey != "" { + newData := make(Fields, 4) + newData[f.DataKey] = data + data = newData + } + + prefixFieldClashes(data, f.FieldMap) + + timestampFormat := f.TimestampFormat + if timestampFormat == "" { + timestampFormat = defaultTimestampFormat + } + + if !f.DisableTimestamp { + data[f.FieldMap.resolve(FieldKeyTime)] = entry.Time.Format(timestampFormat) + } + data[f.FieldMap.resolve(FieldKeyMsg)] = entry.Message + data[f.FieldMap.resolve(FieldKeyLevel)] = entry.Level.String() + + var b *bytes.Buffer + if entry.Buffer != nil { + b = entry.Buffer + } else { + b = &bytes.Buffer{} + } + err := json.NewEncoder(b).Encode(data) + if err != nil { + return nil, fmt.Errorf("Failed to marshal fields to JSON, %v", err) + } + return b.Bytes(), nil +} diff --git a/vendor/github.com/sirupsen/logrus/logger.go b/vendor/github.com/sirupsen/logrus/logger.go new file mode 100644 index 000000000..b67bfcbd3 --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/logger.go @@ -0,0 +1,367 @@ +package logrus + +import ( + "io" + "os" + "sync" + "sync/atomic" + "time" +) + +type Logger struct { + // The logs are `io.Copy`'d to this in a mutex. It's common to set this to a + // file, or leave it default which is `os.Stderr`. You can also set this to + // something more adventurous, such as logging to Kafka. + Out io.Writer + // Hooks for the logger instance. These allow firing events based on logging + // levels and log entries. For example, to send errors to an error tracking + // service, log to StatsD or dump the core on fatal errors. + Hooks LevelHooks + // All log entries pass through the formatter before logged to Out. The + // included formatters are `TextFormatter` and `JSONFormatter` for which + // TextFormatter is the default. In development (when a TTY is attached) it + // logs with colors, but to a file it wouldn't. You can easily implement your + // own that implements the `Formatter` interface, see the `README` or included + // formatters for examples. + Formatter Formatter + // The logging level the logger should log at. This is typically (and defaults + // to) `logrus.Info`, which allows Info(), Warn(), Error() and Fatal() to be + // logged. + Level Level + // Used to sync writing to the log. Locking is enabled by Default + mu MutexWrap + // Reusable empty entry + entryPool sync.Pool +} + +type MutexWrap struct { + lock sync.Mutex + disabled bool +} + +func (mw *MutexWrap) Lock() { + if !mw.disabled { + mw.lock.Lock() + } +} + +func (mw *MutexWrap) Unlock() { + if !mw.disabled { + mw.lock.Unlock() + } +} + +func (mw *MutexWrap) Disable() { + mw.disabled = true +} + +// Creates a new logger. Configuration should be set by changing `Formatter`, +// `Out` and `Hooks` directly on the default logger instance. You can also just +// instantiate your own: +// +// var log = &Logger{ +// Out: os.Stderr, +// Formatter: new(JSONFormatter), +// Hooks: make(LevelHooks), +// Level: logrus.DebugLevel, +// } +// +// It's recommended to make this a global instance called `log`. +func New() *Logger { + return &Logger{ + Out: os.Stderr, + Formatter: new(TextFormatter), + Hooks: make(LevelHooks), + Level: InfoLevel, + } +} + +func (logger *Logger) newEntry() *Entry { + entry, ok := logger.entryPool.Get().(*Entry) + if ok { + return entry + } + return NewEntry(logger) +} + +func (logger *Logger) releaseEntry(entry *Entry) { + entry.Data = map[string]interface{}{} + logger.entryPool.Put(entry) +} + +// Adds a field to the log entry, note that it doesn't log until you call +// Debug, Print, Info, Warn, Error, Fatal or Panic. It only creates a log entry. +// If you want multiple fields, use `WithFields`. +func (logger *Logger) WithField(key string, value interface{}) *Entry { + entry := logger.newEntry() + defer logger.releaseEntry(entry) + return entry.WithField(key, value) +} + +// Adds a struct of fields to the log entry. All it does is call `WithField` for +// each `Field`. +func (logger *Logger) WithFields(fields Fields) *Entry { + entry := logger.newEntry() + defer logger.releaseEntry(entry) + return entry.WithFields(fields) +} + +// Add an error as single field to the log entry. All it does is call +// `WithError` for the given `error`. +func (logger *Logger) WithError(err error) *Entry { + entry := logger.newEntry() + defer logger.releaseEntry(entry) + return entry.WithError(err) +} + +// Overrides the time of the log entry. +func (logger *Logger) WithTime(t time.Time) *Entry { + entry := logger.newEntry() + defer logger.releaseEntry(entry) + return entry.WithTime(t) +} + +func (logger *Logger) Debugf(format string, args ...interface{}) { + if logger.IsLevelEnabled(DebugLevel) { + entry := logger.newEntry() + entry.Debugf(format, args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Infof(format string, args ...interface{}) { + if logger.IsLevelEnabled(InfoLevel) { + entry := logger.newEntry() + entry.Infof(format, args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Printf(format string, args ...interface{}) { + entry := logger.newEntry() + entry.Printf(format, args...) + logger.releaseEntry(entry) +} + +func (logger *Logger) Warnf(format string, args ...interface{}) { + if logger.IsLevelEnabled(WarnLevel) { + entry := logger.newEntry() + entry.Warnf(format, args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Warningf(format string, args ...interface{}) { + if logger.IsLevelEnabled(WarnLevel) { + entry := logger.newEntry() + entry.Warnf(format, args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Errorf(format string, args ...interface{}) { + if logger.IsLevelEnabled(ErrorLevel) { + entry := logger.newEntry() + entry.Errorf(format, args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Fatalf(format string, args ...interface{}) { + if logger.IsLevelEnabled(FatalLevel) { + entry := logger.newEntry() + entry.Fatalf(format, args...) + logger.releaseEntry(entry) + } + Exit(1) +} + +func (logger *Logger) Panicf(format string, args ...interface{}) { + if logger.IsLevelEnabled(PanicLevel) { + entry := logger.newEntry() + entry.Panicf(format, args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Debug(args ...interface{}) { + if logger.IsLevelEnabled(DebugLevel) { + entry := logger.newEntry() + entry.Debug(args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Info(args ...interface{}) { + if logger.IsLevelEnabled(InfoLevel) { + entry := logger.newEntry() + entry.Info(args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Print(args ...interface{}) { + entry := logger.newEntry() + entry.Info(args...) + logger.releaseEntry(entry) +} + +func (logger *Logger) Warn(args ...interface{}) { + if logger.IsLevelEnabled(WarnLevel) { + entry := logger.newEntry() + entry.Warn(args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Warning(args ...interface{}) { + if logger.IsLevelEnabled(WarnLevel) { + entry := logger.newEntry() + entry.Warn(args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Error(args ...interface{}) { + if logger.IsLevelEnabled(ErrorLevel) { + entry := logger.newEntry() + entry.Error(args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Fatal(args ...interface{}) { + if logger.IsLevelEnabled(FatalLevel) { + entry := logger.newEntry() + entry.Fatal(args...) + logger.releaseEntry(entry) + } + Exit(1) +} + +func (logger *Logger) Panic(args ...interface{}) { + if logger.IsLevelEnabled(PanicLevel) { + entry := logger.newEntry() + entry.Panic(args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Debugln(args ...interface{}) { + if logger.IsLevelEnabled(DebugLevel) { + entry := logger.newEntry() + entry.Debugln(args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Infoln(args ...interface{}) { + if logger.IsLevelEnabled(InfoLevel) { + entry := logger.newEntry() + entry.Infoln(args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Println(args ...interface{}) { + entry := logger.newEntry() + entry.Println(args...) + logger.releaseEntry(entry) +} + +func (logger *Logger) Warnln(args ...interface{}) { + if logger.IsLevelEnabled(WarnLevel) { + entry := logger.newEntry() + entry.Warnln(args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Warningln(args ...interface{}) { + if logger.IsLevelEnabled(WarnLevel) { + entry := logger.newEntry() + entry.Warnln(args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Errorln(args ...interface{}) { + if logger.IsLevelEnabled(ErrorLevel) { + entry := logger.newEntry() + entry.Errorln(args...) + logger.releaseEntry(entry) + } +} + +func (logger *Logger) Fatalln(args ...interface{}) { + if logger.IsLevelEnabled(FatalLevel) { + entry := logger.newEntry() + entry.Fatalln(args...) + logger.releaseEntry(entry) + } + Exit(1) +} + +func (logger *Logger) Panicln(args ...interface{}) { + if logger.IsLevelEnabled(PanicLevel) { + entry := logger.newEntry() + entry.Panicln(args...) + logger.releaseEntry(entry) + } +} + +//When file is opened with appending mode, it's safe to +//write concurrently to a file (within 4k message on Linux). +//In these cases user can choose to disable the lock. +func (logger *Logger) SetNoLock() { + logger.mu.Disable() +} + +func (logger *Logger) level() Level { + return Level(atomic.LoadUint32((*uint32)(&logger.Level))) +} + +// SetLevel sets the logger level. +func (logger *Logger) SetLevel(level Level) { + atomic.StoreUint32((*uint32)(&logger.Level), uint32(level)) +} + +// GetLevel returns the logger level. +func (logger *Logger) GetLevel() Level { + return logger.level() +} + +// AddHook adds a hook to the logger hooks. +func (logger *Logger) AddHook(hook Hook) { + logger.mu.Lock() + defer logger.mu.Unlock() + logger.Hooks.Add(hook) +} + +// IsLevelEnabled checks if the log level of the logger is greater than the level param +func (logger *Logger) IsLevelEnabled(level Level) bool { + return logger.level() >= level +} + +// SetFormatter sets the logger formatter. +func (logger *Logger) SetFormatter(formatter Formatter) { + logger.mu.Lock() + defer logger.mu.Unlock() + logger.Formatter = formatter +} + +// SetOutput sets the logger output. +func (logger *Logger) SetOutput(output io.Writer) { + logger.mu.Lock() + defer logger.mu.Unlock() + logger.Out = output +} + +// ReplaceHooks replaces the logger hooks and returns the old ones +func (logger *Logger) ReplaceHooks(hooks LevelHooks) LevelHooks { + logger.mu.Lock() + oldHooks := logger.Hooks + logger.Hooks = hooks + logger.mu.Unlock() + return oldHooks +} diff --git a/vendor/github.com/Sirupsen/logrus/logrus.go b/vendor/github.com/sirupsen/logrus/logrus.go similarity index 95% rename from vendor/github.com/Sirupsen/logrus/logrus.go rename to vendor/github.com/sirupsen/logrus/logrus.go index e59669111..fa0b9dea8 100644 --- a/vendor/github.com/Sirupsen/logrus/logrus.go +++ b/vendor/github.com/sirupsen/logrus/logrus.go @@ -10,7 +10,7 @@ import ( type Fields map[string]interface{} // Level type -type Level uint8 +type Level uint32 // Convert the Level to a string. E.g. PanicLevel becomes "panic". func (level Level) String() string { @@ -140,4 +140,11 @@ type FieldLogger interface { Errorln(args ...interface{}) Fatalln(args ...interface{}) Panicln(args ...interface{}) + + // IsDebugEnabled() bool + // IsInfoEnabled() bool + // IsWarnEnabled() bool + // IsErrorEnabled() bool + // IsFatalEnabled() bool + // IsPanicEnabled() bool } diff --git a/vendor/github.com/sirupsen/logrus/terminal_bsd.go b/vendor/github.com/sirupsen/logrus/terminal_bsd.go new file mode 100644 index 000000000..5b6212d24 --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/terminal_bsd.go @@ -0,0 +1,10 @@ +// +build darwin freebsd openbsd netbsd dragonfly +// +build !appengine,!js + +package logrus + +import "golang.org/x/sys/unix" + +const ioctlReadTermios = unix.TIOCGETA + +type Termios unix.Termios diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_appengine.go b/vendor/github.com/sirupsen/logrus/terminal_check_appengine.go new file mode 100644 index 000000000..2403de981 --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/terminal_check_appengine.go @@ -0,0 +1,11 @@ +// +build appengine + +package logrus + +import ( + "io" +) + +func checkIfTerminal(w io.Writer) bool { + return true +} diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_js.go b/vendor/github.com/sirupsen/logrus/terminal_check_js.go new file mode 100644 index 000000000..0c209750a --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/terminal_check_js.go @@ -0,0 +1,11 @@ +// +build js + +package logrus + +import ( + "io" +) + +func checkIfTerminal(w io.Writer) bool { + return false +} diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_notappengine.go b/vendor/github.com/sirupsen/logrus/terminal_check_notappengine.go new file mode 100644 index 000000000..87f0b8097 --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/terminal_check_notappengine.go @@ -0,0 +1,19 @@ +// +build !appengine,!js + +package logrus + +import ( + "io" + "os" + + "golang.org/x/crypto/ssh/terminal" +) + +func checkIfTerminal(w io.Writer) bool { + switch v := w.(type) { + case *os.File: + return terminal.IsTerminal(int(v.Fd())) + default: + return false + } +} diff --git a/vendor/github.com/Sirupsen/logrus/terminal_linux.go b/vendor/github.com/sirupsen/logrus/terminal_linux.go similarity index 62% rename from vendor/github.com/Sirupsen/logrus/terminal_linux.go rename to vendor/github.com/sirupsen/logrus/terminal_linux.go index a2c0b40db..634e39b30 100644 --- a/vendor/github.com/Sirupsen/logrus/terminal_linux.go +++ b/vendor/github.com/sirupsen/logrus/terminal_linux.go @@ -3,10 +3,12 @@ // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. +// +build !appengine,!js + package logrus -import "syscall" +import "golang.org/x/sys/unix" -const ioctlReadTermios = syscall.TCGETS +const ioctlReadTermios = unix.TCGETS -type Termios syscall.Termios +type Termios unix.Termios diff --git a/vendor/github.com/sirupsen/logrus/text_formatter.go b/vendor/github.com/sirupsen/logrus/text_formatter.go new file mode 100644 index 000000000..4b1f071f1 --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/text_formatter.go @@ -0,0 +1,217 @@ +package logrus + +import ( + "bytes" + "fmt" + "os" + "sort" + "strings" + "sync" + "time" +) + +const ( + nocolor = 0 + red = 31 + green = 32 + yellow = 33 + blue = 36 + gray = 37 +) + +var ( + baseTimestamp time.Time + emptyFieldMap FieldMap +) + +func init() { + baseTimestamp = time.Now() +} + +// TextFormatter formats logs into text +type TextFormatter struct { + // Set to true to bypass checking for a TTY before outputting colors. + ForceColors bool + + // Force disabling colors. + DisableColors bool + + // Override coloring based on CLICOLOR and CLICOLOR_FORCE. - https://bixense.com/clicolors/ + EnvironmentOverrideColors bool + + // Disable timestamp logging. useful when output is redirected to logging + // system that already adds timestamps. + DisableTimestamp bool + + // Enable logging the full timestamp when a TTY is attached instead of just + // the time passed since beginning of execution. + FullTimestamp bool + + // TimestampFormat to use for display when a full timestamp is printed + TimestampFormat string + + // The fields are sorted by default for a consistent output. For applications + // that log extremely frequently and don't use the JSON formatter this may not + // be desired. + DisableSorting bool + + // Disables the truncation of the level text to 4 characters. + DisableLevelTruncation bool + + // QuoteEmptyFields will wrap empty fields in quotes if true + QuoteEmptyFields bool + + // Whether the logger's out is to a terminal + isTerminal bool + + // FieldMap allows users to customize the names of keys for default fields. + // As an example: + // formatter := &TextFormatter{ + // FieldMap: FieldMap{ + // FieldKeyTime: "@timestamp", + // FieldKeyLevel: "@level", + // FieldKeyMsg: "@message"}} + FieldMap FieldMap + + sync.Once +} + +func (f *TextFormatter) init(entry *Entry) { + if entry.Logger != nil { + f.isTerminal = checkIfTerminal(entry.Logger.Out) + } +} + +func (f *TextFormatter) isColored() bool { + isColored := f.ForceColors || f.isTerminal + + if f.EnvironmentOverrideColors { + if force, ok := os.LookupEnv("CLICOLOR_FORCE"); ok && force != "0" { + isColored = true + } else if ok && force == "0" { + isColored = false + } else if os.Getenv("CLICOLOR") == "0" { + isColored = false + } + } + + return isColored && !f.DisableColors +} + +// Format renders a single log entry +func (f *TextFormatter) Format(entry *Entry) ([]byte, error) { + prefixFieldClashes(entry.Data, f.FieldMap) + + keys := make([]string, 0, len(entry.Data)) + for k := range entry.Data { + keys = append(keys, k) + } + + if !f.DisableSorting { + sort.Strings(keys) + } + + var b *bytes.Buffer + if entry.Buffer != nil { + b = entry.Buffer + } else { + b = &bytes.Buffer{} + } + + f.Do(func() { f.init(entry) }) + + timestampFormat := f.TimestampFormat + if timestampFormat == "" { + timestampFormat = defaultTimestampFormat + } + if f.isColored() { + f.printColored(b, entry, keys, timestampFormat) + } else { + if !f.DisableTimestamp { + f.appendKeyValue(b, f.FieldMap.resolve(FieldKeyTime), entry.Time.Format(timestampFormat)) + } + f.appendKeyValue(b, f.FieldMap.resolve(FieldKeyLevel), entry.Level.String()) + if entry.Message != "" { + f.appendKeyValue(b, f.FieldMap.resolve(FieldKeyMsg), entry.Message) + } + for _, key := range keys { + f.appendKeyValue(b, key, entry.Data[key]) + } + } + + b.WriteByte('\n') + return b.Bytes(), nil +} + +func (f *TextFormatter) printColored(b *bytes.Buffer, entry *Entry, keys []string, timestampFormat string) { + var levelColor int + switch entry.Level { + case DebugLevel: + levelColor = gray + case WarnLevel: + levelColor = yellow + case ErrorLevel, FatalLevel, PanicLevel: + levelColor = red + default: + levelColor = blue + } + + levelText := strings.ToUpper(entry.Level.String()) + if !f.DisableLevelTruncation { + levelText = levelText[0:4] + } + + // Remove a single newline if it already exists in the message to keep + // the behavior of logrus text_formatter the same as the stdlib log package + entry.Message = strings.TrimSuffix(entry.Message, "\n") + + if f.DisableTimestamp { + fmt.Fprintf(b, "\x1b[%dm%s\x1b[0m %-44s ", levelColor, levelText, entry.Message) + } else if !f.FullTimestamp { + fmt.Fprintf(b, "\x1b[%dm%s\x1b[0m[%04d] %-44s ", levelColor, levelText, int(entry.Time.Sub(baseTimestamp)/time.Second), entry.Message) + } else { + fmt.Fprintf(b, "\x1b[%dm%s\x1b[0m[%s] %-44s ", levelColor, levelText, entry.Time.Format(timestampFormat), entry.Message) + } + for _, k := range keys { + v := entry.Data[k] + fmt.Fprintf(b, " \x1b[%dm%s\x1b[0m=", levelColor, k) + f.appendValue(b, v) + } +} + +func (f *TextFormatter) needsQuoting(text string) bool { + if f.QuoteEmptyFields && len(text) == 0 { + return true + } + for _, ch := range text { + if !((ch >= 'a' && ch <= 'z') || + (ch >= 'A' && ch <= 'Z') || + (ch >= '0' && ch <= '9') || + ch == '-' || ch == '.' || ch == '_' || ch == '/' || ch == '@' || ch == '^' || ch == '+') { + return true + } + } + return false +} + +func (f *TextFormatter) appendKeyValue(b *bytes.Buffer, key string, value interface{}) { + if b.Len() > 0 { + b.WriteByte(' ') + } + b.WriteString(key) + b.WriteByte('=') + f.appendValue(b, value) +} + +func (f *TextFormatter) appendValue(b *bytes.Buffer, value interface{}) { + stringVal, ok := value.(string) + if !ok { + stringVal = fmt.Sprint(value) + } + + if !f.needsQuoting(stringVal) { + b.WriteString(stringVal) + } else { + b.WriteString(fmt.Sprintf("%q", stringVal)) + } +} diff --git a/vendor/github.com/sirupsen/logrus/writer.go b/vendor/github.com/sirupsen/logrus/writer.go new file mode 100644 index 000000000..7bdebedc6 --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/writer.go @@ -0,0 +1,62 @@ +package logrus + +import ( + "bufio" + "io" + "runtime" +) + +func (logger *Logger) Writer() *io.PipeWriter { + return logger.WriterLevel(InfoLevel) +} + +func (logger *Logger) WriterLevel(level Level) *io.PipeWriter { + return NewEntry(logger).WriterLevel(level) +} + +func (entry *Entry) Writer() *io.PipeWriter { + return entry.WriterLevel(InfoLevel) +} + +func (entry *Entry) WriterLevel(level Level) *io.PipeWriter { + reader, writer := io.Pipe() + + var printFunc func(args ...interface{}) + + switch level { + case DebugLevel: + printFunc = entry.Debug + case InfoLevel: + printFunc = entry.Info + case WarnLevel: + printFunc = entry.Warn + case ErrorLevel: + printFunc = entry.Error + case FatalLevel: + printFunc = entry.Fatal + case PanicLevel: + printFunc = entry.Panic + default: + printFunc = entry.Print + } + + go entry.writerScanner(reader, printFunc) + runtime.SetFinalizer(writer, writerFinalizer) + + return writer +} + +func (entry *Entry) writerScanner(reader *io.PipeReader, printFunc func(args ...interface{})) { + scanner := bufio.NewScanner(reader) + for scanner.Scan() { + printFunc(scanner.Text()) + } + if err := scanner.Err(); err != nil { + entry.Errorf("Error while reading from Writer: %s", err) + } + reader.Close() +} + +func writerFinalizer(writer *io.PipeWriter) { + writer.Close() +} diff --git a/vendor/github.com/ulikunitz/xz/example.go b/vendor/github.com/ulikunitz/xz/example.go deleted file mode 100644 index 855e60aee..000000000 --- a/vendor/github.com/ulikunitz/xz/example.go +++ /dev/null @@ -1,40 +0,0 @@ -// Copyright 2014-2017 Ulrich Kunitz. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build ignore - -package main - -import ( - "bytes" - "io" - "log" - "os" - - "github.com/ulikunitz/xz" -) - -func main() { - const text = "The quick brown fox jumps over the lazy dog.\n" - var buf bytes.Buffer - // compress text - w, err := xz.NewWriter(&buf) - if err != nil { - log.Fatalf("xz.NewWriter error %s", err) - } - if _, err := io.WriteString(w, text); err != nil { - log.Fatalf("WriteString error %s", err) - } - if err := w.Close(); err != nil { - log.Fatalf("w.Close error %s", err) - } - // decompress buffer and write output to stdout - r, err := xz.NewReader(&buf) - if err != nil { - log.Fatalf("NewReader error %s", err) - } - if _, err = io.Copy(os.Stdout, r); err != nil { - log.Fatalf("io.Copy error %s", err) - } -} diff --git a/vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/vishvananda/netlink/LICENSE b/vendor/github.com/vishvananda/netlink/LICENSE similarity index 100% rename from vendor/github.com/opencontainers/runc/Godeps/_workspace/src/github.com/vishvananda/netlink/LICENSE rename to vendor/github.com/vishvananda/netlink/LICENSE diff --git a/vendor/github.com/vishvananda/netlink/README.md b/vendor/github.com/vishvananda/netlink/README.md new file mode 100644 index 000000000..8cd50a93b --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/README.md @@ -0,0 +1,89 @@ +# netlink - netlink library for go # + +[![Build Status](https://travis-ci.org/vishvananda/netlink.png?branch=master)](https://travis-ci.org/vishvananda/netlink) [![GoDoc](https://godoc.org/github.com/vishvananda/netlink?status.svg)](https://godoc.org/github.com/vishvananda/netlink) + +The netlink package provides a simple netlink library for go. Netlink +is the interface a user-space program in linux uses to communicate with +the kernel. It can be used to add and remove interfaces, set ip addresses +and routes, and configure ipsec. Netlink communication requires elevated +privileges, so in most cases this code needs to be run as root. Since +low-level netlink messages are inscrutable at best, the library attempts +to provide an api that is loosely modeled on the CLI provied by iproute2. +Actions like `ip link add` will be accomplished via a similarly named +function like AddLink(). This library began its life as a fork of the +netlink functionality in +[docker/libcontainer](https://github.com/docker/libcontainer) but was +heavily rewritten to improve testability, performance, and to add new +functionality like ipsec xfrm handling. + +## Local Build and Test ## + +You can use go get command: + + go get github.com/vishvananda/netlink + +Testing dependencies: + + go get github.com/vishvananda/netns + +Testing (requires root): + + sudo -E go test github.com/vishvananda/netlink + +## Examples ## + +Add a new bridge and add eth1 into it: + +```go +package main + +import ( + "net" + "github.com/vishvananda/netlink" +) + +func main() { + la := netlink.NewLinkAttrs() + la.Name = "foo" + mybridge := &netlink.Bridge{la}} + _ := netlink.LinkAdd(mybridge) + eth1, _ := netlink.LinkByName("eth1") + netlink.LinkSetMaster(eth1, mybridge) +} + +``` +Note `NewLinkAttrs` constructor, it sets default values in structure. For now +it sets only `TxQLen` to `-1`, so kernel will set default by itself. If you're +using simple initialization(`LinkAttrs{Name: "foo"}`) `TxQLen` will be set to +`0` unless you specify it like `LinkAttrs{Name: "foo", TxQLen: 1000}`. + +Add a new ip address to loopback: + +```go +package main + +import ( + "net" + "github.com/vishvananda/netlink" +) + +func main() { + lo, _ := netlink.LinkByName("lo") + addr, _ := netlink.ParseAddr("169.254.169.254/32") + netlink.AddrAdd(lo, addr) +} + +``` + +## Future Work ## + +Many pieces of netlink are not yet fully supported in the high-level +interface. Aspects of virtually all of the high-level objects don't exist. +Many of the underlying primitives are there, so its a matter of putting +the right fields into the high-level objects and making sure that they +are serialized and deserialized correctly in the Add and List methods. + +There are also a few pieces of low level netlink functionality that still +need to be implemented. Routing rules are not in place and some of the +more advanced link types. Hopefully there is decent structure and testing +in place to make these fairly straightforward to add. diff --git a/vendor/github.com/vishvananda/netlink/addr.go b/vendor/github.com/vishvananda/netlink/addr.go new file mode 100644 index 000000000..9bbaf508e --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/addr.go @@ -0,0 +1,43 @@ +package netlink + +import ( + "fmt" + "net" + "strings" +) + +// Addr represents an IP address from netlink. Netlink ip addresses +// include a mask, so it stores the address as a net.IPNet. +type Addr struct { + *net.IPNet + Label string +} + +// String returns $ip/$netmask $label +func (a Addr) String() string { + return fmt.Sprintf("%s %s", a.IPNet, a.Label) +} + +// ParseAddr parses the string representation of an address in the +// form $ip/$netmask $label. The label portion is optional +func ParseAddr(s string) (*Addr, error) { + label := "" + parts := strings.Split(s, " ") + if len(parts) > 1 { + s = parts[0] + label = parts[1] + } + m, err := ParseIPNet(s) + if err != nil { + return nil, err + } + return &Addr{IPNet: m, Label: label}, nil +} + +// Equal returns true if both Addrs have the same net.IPNet value. +func (a Addr) Equal(x Addr) bool { + sizea, _ := a.Mask.Size() + sizeb, _ := x.Mask.Size() + // ignore label for comparison + return a.IP.Equal(x.IP) && sizea == sizeb +} diff --git a/vendor/github.com/vishvananda/netlink/addr_linux.go b/vendor/github.com/vishvananda/netlink/addr_linux.go new file mode 100644 index 000000000..19aac0fb9 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/addr_linux.go @@ -0,0 +1,128 @@ +package netlink + +import ( + "fmt" + "net" + "strings" + "syscall" + + "github.com/vishvananda/netlink/nl" +) + +// AddrAdd will add an IP address to a link device. +// Equivalent to: `ip addr add $addr dev $link` +func AddrAdd(link Link, addr *Addr) error { + + req := nl.NewNetlinkRequest(syscall.RTM_NEWADDR, syscall.NLM_F_CREATE|syscall.NLM_F_EXCL|syscall.NLM_F_ACK) + return addrHandle(link, addr, req) +} + +// AddrDel will delete an IP address from a link device. +// Equivalent to: `ip addr del $addr dev $link` +func AddrDel(link Link, addr *Addr) error { + req := nl.NewNetlinkRequest(syscall.RTM_DELADDR, syscall.NLM_F_ACK) + return addrHandle(link, addr, req) +} + +func addrHandle(link Link, addr *Addr, req *nl.NetlinkRequest) error { + base := link.Attrs() + if addr.Label != "" && !strings.HasPrefix(addr.Label, base.Name) { + return fmt.Errorf("label must begin with interface name") + } + ensureIndex(base) + + family := nl.GetIPFamily(addr.IP) + + msg := nl.NewIfAddrmsg(family) + msg.Index = uint32(base.Index) + prefixlen, _ := addr.Mask.Size() + msg.Prefixlen = uint8(prefixlen) + req.AddData(msg) + + var addrData []byte + if family == FAMILY_V4 { + addrData = addr.IP.To4() + } else { + addrData = addr.IP.To16() + } + + localData := nl.NewRtAttr(syscall.IFA_LOCAL, addrData) + req.AddData(localData) + + addressData := nl.NewRtAttr(syscall.IFA_ADDRESS, addrData) + req.AddData(addressData) + + if addr.Label != "" { + labelData := nl.NewRtAttr(syscall.IFA_LABEL, nl.ZeroTerminated(addr.Label)) + req.AddData(labelData) + } + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// AddrList gets a list of IP addresses in the system. +// Equivalent to: `ip addr show`. +// The list can be filtered by link and ip family. +func AddrList(link Link, family int) ([]Addr, error) { + req := nl.NewNetlinkRequest(syscall.RTM_GETADDR, syscall.NLM_F_DUMP) + msg := nl.NewIfInfomsg(family) + req.AddData(msg) + + msgs, err := req.Execute(syscall.NETLINK_ROUTE, syscall.RTM_NEWADDR) + if err != nil { + return nil, err + } + + index := 0 + if link != nil { + base := link.Attrs() + ensureIndex(base) + index = base.Index + } + + var res []Addr + for _, m := range msgs { + msg := nl.DeserializeIfAddrmsg(m) + + if link != nil && msg.Index != uint32(index) { + // Ignore messages from other interfaces + continue + } + + attrs, err := nl.ParseRouteAttr(m[msg.Len():]) + if err != nil { + return nil, err + } + + var local, dst *net.IPNet + var addr Addr + for _, attr := range attrs { + switch attr.Attr.Type { + case syscall.IFA_ADDRESS: + dst = &net.IPNet{ + IP: attr.Value, + Mask: net.CIDRMask(int(msg.Prefixlen), 8*len(attr.Value)), + } + case syscall.IFA_LOCAL: + local = &net.IPNet{ + IP: attr.Value, + Mask: net.CIDRMask(int(msg.Prefixlen), 8*len(attr.Value)), + } + case syscall.IFA_LABEL: + addr.Label = string(attr.Value[:len(attr.Value)-1]) + } + } + + // IFA_LOCAL should be there but if not, fall back to IFA_ADDRESS + if local != nil { + addr.IPNet = local + } else { + addr.IPNet = dst + } + + res = append(res, addr) + } + + return res, nil +} diff --git a/vendor/github.com/vishvananda/netlink/filter.go b/vendor/github.com/vishvananda/netlink/filter.go new file mode 100644 index 000000000..83ad7007b --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/filter.go @@ -0,0 +1,55 @@ +package netlink + +import ( + "fmt" +) + +type Filter interface { + Attrs() *FilterAttrs + Type() string +} + +// Filter represents a netlink filter. A filter is associated with a link, +// has a handle and a parent. The root filter of a device should have a +// parent == HANDLE_ROOT. +type FilterAttrs struct { + LinkIndex int + Handle uint32 + Parent uint32 + Priority uint16 // lower is higher priority + Protocol uint16 // syscall.ETH_P_* +} + +func (q FilterAttrs) String() string { + return fmt.Sprintf("{LinkIndex: %d, Handle: %s, Parent: %s, Priority: %d, Protocol: %d}", q.LinkIndex, HandleStr(q.Handle), HandleStr(q.Parent), q.Priority, q.Protocol) +} + +// U32 filters on many packet related properties +type U32 struct { + FilterAttrs + // Currently only supports redirecting to another interface + RedirIndex int +} + +func (filter *U32) Attrs() *FilterAttrs { + return &filter.FilterAttrs +} + +func (filter *U32) Type() string { + return "u32" +} + +// GenericFilter filters represent types that are not currently understood +// by this netlink library. +type GenericFilter struct { + FilterAttrs + FilterType string +} + +func (filter *GenericFilter) Attrs() *FilterAttrs { + return &filter.FilterAttrs +} + +func (filter *GenericFilter) Type() string { + return filter.FilterType +} diff --git a/vendor/github.com/vishvananda/netlink/filter_linux.go b/vendor/github.com/vishvananda/netlink/filter_linux.go new file mode 100644 index 000000000..1ec698702 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/filter_linux.go @@ -0,0 +1,191 @@ +package netlink + +import ( + "fmt" + "syscall" + + "github.com/vishvananda/netlink/nl" +) + +// FilterDel will delete a filter from the system. +// Equivalent to: `tc filter del $filter` +func FilterDel(filter Filter) error { + req := nl.NewNetlinkRequest(syscall.RTM_DELTFILTER, syscall.NLM_F_ACK) + base := filter.Attrs() + msg := &nl.TcMsg{ + Family: nl.FAMILY_ALL, + Ifindex: int32(base.LinkIndex), + Handle: base.Handle, + Parent: base.Parent, + Info: MakeHandle(base.Priority, nl.Swap16(base.Protocol)), + } + req.AddData(msg) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// FilterAdd will add a filter to the system. +// Equivalent to: `tc filter add $filter` +func FilterAdd(filter Filter) error { + req := nl.NewNetlinkRequest(syscall.RTM_NEWTFILTER, syscall.NLM_F_CREATE|syscall.NLM_F_EXCL|syscall.NLM_F_ACK) + base := filter.Attrs() + msg := &nl.TcMsg{ + Family: nl.FAMILY_ALL, + Ifindex: int32(base.LinkIndex), + Handle: base.Handle, + Parent: base.Parent, + Info: MakeHandle(base.Priority, nl.Swap16(base.Protocol)), + } + req.AddData(msg) + req.AddData(nl.NewRtAttr(nl.TCA_KIND, nl.ZeroTerminated(filter.Type()))) + + options := nl.NewRtAttr(nl.TCA_OPTIONS, nil) + if u32, ok := filter.(*U32); ok { + // match all + sel := nl.TcU32Sel{ + Nkeys: 1, + Flags: nl.TC_U32_TERMINAL, + } + sel.Keys = append(sel.Keys, nl.TcU32Key{}) + nl.NewRtAttrChild(options, nl.TCA_U32_SEL, sel.Serialize()) + actions := nl.NewRtAttrChild(options, nl.TCA_U32_ACT, nil) + table := nl.NewRtAttrChild(actions, nl.TCA_ACT_TAB, nil) + nl.NewRtAttrChild(table, nl.TCA_KIND, nl.ZeroTerminated("mirred")) + // redirect to other interface + mir := nl.TcMirred{ + Action: nl.TC_ACT_STOLEN, + Eaction: nl.TCA_EGRESS_REDIR, + Ifindex: uint32(u32.RedirIndex), + } + aopts := nl.NewRtAttrChild(table, nl.TCA_OPTIONS, nil) + nl.NewRtAttrChild(aopts, nl.TCA_MIRRED_PARMS, mir.Serialize()) + } + req.AddData(options) + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// FilterList gets a list of filters in the system. +// Equivalent to: `tc filter show`. +// Generally retunrs nothing if link and parent are not specified. +func FilterList(link Link, parent uint32) ([]Filter, error) { + req := nl.NewNetlinkRequest(syscall.RTM_GETTFILTER, syscall.NLM_F_DUMP) + msg := &nl.TcMsg{ + Family: nl.FAMILY_ALL, + Parent: parent, + } + if link != nil { + base := link.Attrs() + ensureIndex(base) + msg.Ifindex = int32(base.Index) + } + req.AddData(msg) + + msgs, err := req.Execute(syscall.NETLINK_ROUTE, syscall.RTM_NEWTFILTER) + if err != nil { + return nil, err + } + + var res []Filter + for _, m := range msgs { + msg := nl.DeserializeTcMsg(m) + + attrs, err := nl.ParseRouteAttr(m[msg.Len():]) + if err != nil { + return nil, err + } + + base := FilterAttrs{ + LinkIndex: int(msg.Ifindex), + Handle: msg.Handle, + Parent: msg.Parent, + } + base.Priority, base.Protocol = MajorMinor(msg.Info) + base.Protocol = nl.Swap16(base.Protocol) + + var filter Filter + filterType := "" + detailed := false + for _, attr := range attrs { + switch attr.Attr.Type { + case nl.TCA_KIND: + filterType = string(attr.Value[:len(attr.Value)-1]) + switch filterType { + case "u32": + filter = &U32{} + default: + filter = &GenericFilter{FilterType: filterType} + } + case nl.TCA_OPTIONS: + switch filterType { + case "u32": + data, err := nl.ParseRouteAttr(attr.Value) + if err != nil { + return nil, err + } + detailed, err = parseU32Data(filter, data) + if err != nil { + return nil, err + } + } + } + } + // only return the detailed version of the filter + if detailed { + *filter.Attrs() = base + res = append(res, filter) + } + } + + return res, nil +} + +func parseU32Data(filter Filter, data []syscall.NetlinkRouteAttr) (bool, error) { + native = nl.NativeEndian() + u32 := filter.(*U32) + detailed := false + for _, datum := range data { + switch datum.Attr.Type { + case nl.TCA_U32_SEL: + detailed = true + sel := nl.DeserializeTcU32Sel(datum.Value) + // only parse if we have a very basic redirect + if sel.Flags&nl.TC_U32_TERMINAL == 0 || sel.Nkeys != 1 { + return detailed, nil + } + case nl.TCA_U32_ACT: + table, err := nl.ParseRouteAttr(datum.Value) + if err != nil { + return detailed, err + } + if len(table) != 1 || table[0].Attr.Type != nl.TCA_ACT_TAB { + return detailed, fmt.Errorf("Action table not formed properly") + } + aattrs, err := nl.ParseRouteAttr(table[0].Value) + for _, aattr := range aattrs { + switch aattr.Attr.Type { + case nl.TCA_KIND: + actionType := string(aattr.Value[:len(aattr.Value)-1]) + // only parse if the action is mirred + if actionType != "mirred" { + return detailed, nil + } + case nl.TCA_OPTIONS: + adata, err := nl.ParseRouteAttr(aattr.Value) + if err != nil { + return detailed, err + } + for _, adatum := range adata { + switch adatum.Attr.Type { + case nl.TCA_MIRRED_PARMS: + mir := nl.DeserializeTcMirred(adatum.Value) + u32.RedirIndex = int(mir.Ifindex) + } + } + } + } + } + } + return detailed, nil +} diff --git a/vendor/github.com/vishvananda/netlink/link.go b/vendor/github.com/vishvananda/netlink/link.go new file mode 100644 index 000000000..18fd1759a --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/link.go @@ -0,0 +1,223 @@ +package netlink + +import "net" + +// Link represents a link device from netlink. Shared link attributes +// like name may be retrieved using the Attrs() method. Unique data +// can be retrieved by casting the object to the proper type. +type Link interface { + Attrs() *LinkAttrs + Type() string +} + +type ( + NsPid int + NsFd int +) + +// LinkAttrs represents data shared by most link types +type LinkAttrs struct { + Index int + MTU int + TxQLen int // Transmit Queue Length + Name string + HardwareAddr net.HardwareAddr + Flags net.Flags + ParentIndex int // index of the parent link device + MasterIndex int // must be the index of a bridge + Namespace interface{} // nil | NsPid | NsFd +} + +// NewLinkAttrs returns LinkAttrs structure filled with default values +func NewLinkAttrs() LinkAttrs { + return LinkAttrs{ + TxQLen: -1, + } +} + +// Device links cannot be created via netlink. These links +// are links created by udev like 'lo' and 'etho0' +type Device struct { + LinkAttrs +} + +func (device *Device) Attrs() *LinkAttrs { + return &device.LinkAttrs +} + +func (device *Device) Type() string { + return "device" +} + +// Dummy links are dummy ethernet devices +type Dummy struct { + LinkAttrs +} + +func (dummy *Dummy) Attrs() *LinkAttrs { + return &dummy.LinkAttrs +} + +func (dummy *Dummy) Type() string { + return "dummy" +} + +// Ifb links are advanced dummy devices for packet filtering +type Ifb struct { + LinkAttrs +} + +func (ifb *Ifb) Attrs() *LinkAttrs { + return &ifb.LinkAttrs +} + +func (ifb *Ifb) Type() string { + return "ifb" +} + +// Bridge links are simple linux bridges +type Bridge struct { + LinkAttrs +} + +func (bridge *Bridge) Attrs() *LinkAttrs { + return &bridge.LinkAttrs +} + +func (bridge *Bridge) Type() string { + return "bridge" +} + +// Vlan links have ParentIndex set in their Attrs() +type Vlan struct { + LinkAttrs + VlanId int +} + +func (vlan *Vlan) Attrs() *LinkAttrs { + return &vlan.LinkAttrs +} + +func (vlan *Vlan) Type() string { + return "vlan" +} + +type MacvlanMode uint16 + +const ( + MACVLAN_MODE_DEFAULT MacvlanMode = iota + MACVLAN_MODE_PRIVATE + MACVLAN_MODE_VEPA + MACVLAN_MODE_BRIDGE + MACVLAN_MODE_PASSTHRU + MACVLAN_MODE_SOURCE +) + +// Macvlan links have ParentIndex set in their Attrs() +type Macvlan struct { + LinkAttrs + Mode MacvlanMode +} + +func (macvlan *Macvlan) Attrs() *LinkAttrs { + return &macvlan.LinkAttrs +} + +func (macvlan *Macvlan) Type() string { + return "macvlan" +} + +// Macvtap - macvtap is a virtual interfaces based on macvlan +type Macvtap struct { + Macvlan +} + +func (macvtap Macvtap) Type() string { + return "macvtap" +} + +// Veth devices must specify PeerName on create +type Veth struct { + LinkAttrs + PeerName string // veth on create only +} + +func (veth *Veth) Attrs() *LinkAttrs { + return &veth.LinkAttrs +} + +func (veth *Veth) Type() string { + return "veth" +} + +// GenericLink links represent types that are not currently understood +// by this netlink library. +type GenericLink struct { + LinkAttrs + LinkType string +} + +func (generic *GenericLink) Attrs() *LinkAttrs { + return &generic.LinkAttrs +} + +func (generic *GenericLink) Type() string { + return generic.LinkType +} + +type Vxlan struct { + LinkAttrs + VxlanId int + VtepDevIndex int + SrcAddr net.IP + Group net.IP + TTL int + TOS int + Learning bool + Proxy bool + RSC bool + L2miss bool + L3miss bool + NoAge bool + GBP bool + Age int + Limit int + Port int + PortLow int + PortHigh int +} + +func (vxlan *Vxlan) Attrs() *LinkAttrs { + return &vxlan.LinkAttrs +} + +func (vxlan *Vxlan) Type() string { + return "vxlan" +} + +type IPVlanMode uint16 + +const ( + IPVLAN_MODE_L2 IPVlanMode = iota + IPVLAN_MODE_L3 + IPVLAN_MODE_MAX +) + +type IPVlan struct { + LinkAttrs + Mode IPVlanMode +} + +func (ipvlan *IPVlan) Attrs() *LinkAttrs { + return &ipvlan.LinkAttrs +} + +func (ipvlan *IPVlan) Type() string { + return "ipvlan" +} + +// iproute2 supported devices; +// vlan | veth | vcan | dummy | ifb | macvlan | macvtap | +// bridge | bond | ipoib | ip6tnl | ipip | sit | vxlan | +// gre | gretap | ip6gre | ip6gretap | vti | nlmon | +// bond_slave | ipvlan diff --git a/vendor/github.com/vishvananda/netlink/link_linux.go b/vendor/github.com/vishvananda/netlink/link_linux.go new file mode 100644 index 000000000..685115044 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/link_linux.go @@ -0,0 +1,750 @@ +package netlink + +import ( + "bytes" + "encoding/binary" + "fmt" + "net" + "syscall" + + "github.com/vishvananda/netlink/nl" +) + +var native = nl.NativeEndian() +var lookupByDump = false + +var macvlanModes = [...]uint32{ + 0, + nl.MACVLAN_MODE_PRIVATE, + nl.MACVLAN_MODE_VEPA, + nl.MACVLAN_MODE_BRIDGE, + nl.MACVLAN_MODE_PASSTHRU, + nl.MACVLAN_MODE_SOURCE, +} + +func ensureIndex(link *LinkAttrs) { + if link != nil && link.Index == 0 { + newlink, _ := LinkByName(link.Name) + if newlink != nil { + link.Index = newlink.Attrs().Index + } + } +} + +// LinkSetUp enables the link device. +// Equivalent to: `ip link set $link up` +func LinkSetUp(link Link) error { + base := link.Attrs() + ensureIndex(base) + req := nl.NewNetlinkRequest(syscall.RTM_NEWLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + msg.Change = syscall.IFF_UP + msg.Flags = syscall.IFF_UP + msg.Index = int32(base.Index) + req.AddData(msg) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// LinkSetDown disables link device. +// Equivalent to: `ip link set $link down` +func LinkSetDown(link Link) error { + base := link.Attrs() + ensureIndex(base) + req := nl.NewNetlinkRequest(syscall.RTM_NEWLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + msg.Change = syscall.IFF_UP + msg.Flags = 0 & ^syscall.IFF_UP + msg.Index = int32(base.Index) + req.AddData(msg) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// LinkSetMTU sets the mtu of the link device. +// Equivalent to: `ip link set $link mtu $mtu` +func LinkSetMTU(link Link, mtu int) error { + base := link.Attrs() + ensureIndex(base) + req := nl.NewNetlinkRequest(syscall.RTM_SETLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + msg.Index = int32(base.Index) + req.AddData(msg) + + b := make([]byte, 4) + native.PutUint32(b, uint32(mtu)) + + data := nl.NewRtAttr(syscall.IFLA_MTU, b) + req.AddData(data) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// LinkSetName sets the name of the link device. +// Equivalent to: `ip link set $link name $name` +func LinkSetName(link Link, name string) error { + base := link.Attrs() + ensureIndex(base) + req := nl.NewNetlinkRequest(syscall.RTM_SETLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + msg.Index = int32(base.Index) + req.AddData(msg) + + data := nl.NewRtAttr(syscall.IFLA_IFNAME, []byte(name)) + req.AddData(data) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// LinkSetHardwareAddr sets the hardware address of the link device. +// Equivalent to: `ip link set $link address $hwaddr` +func LinkSetHardwareAddr(link Link, hwaddr net.HardwareAddr) error { + base := link.Attrs() + ensureIndex(base) + req := nl.NewNetlinkRequest(syscall.RTM_SETLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + msg.Index = int32(base.Index) + req.AddData(msg) + + data := nl.NewRtAttr(syscall.IFLA_ADDRESS, []byte(hwaddr)) + req.AddData(data) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// LinkSetMaster sets the master of the link device. +// Equivalent to: `ip link set $link master $master` +func LinkSetMaster(link Link, master *Bridge) error { + index := 0 + if master != nil { + masterBase := master.Attrs() + ensureIndex(masterBase) + index = masterBase.Index + } + return LinkSetMasterByIndex(link, index) +} + +// LinkSetMasterByIndex sets the master of the link device. +// Equivalent to: `ip link set $link master $master` +func LinkSetMasterByIndex(link Link, masterIndex int) error { + base := link.Attrs() + ensureIndex(base) + req := nl.NewNetlinkRequest(syscall.RTM_SETLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + msg.Index = int32(base.Index) + req.AddData(msg) + + b := make([]byte, 4) + native.PutUint32(b, uint32(masterIndex)) + + data := nl.NewRtAttr(syscall.IFLA_MASTER, b) + req.AddData(data) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// LinkSetNsPid puts the device into a new network namespace. The +// pid must be a pid of a running process. +// Equivalent to: `ip link set $link netns $pid` +func LinkSetNsPid(link Link, nspid int) error { + base := link.Attrs() + ensureIndex(base) + req := nl.NewNetlinkRequest(syscall.RTM_SETLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + msg.Index = int32(base.Index) + req.AddData(msg) + + b := make([]byte, 4) + native.PutUint32(b, uint32(nspid)) + + data := nl.NewRtAttr(syscall.IFLA_NET_NS_PID, b) + req.AddData(data) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// LinkSetNsFd puts the device into a new network namespace. The +// fd must be an open file descriptor to a network namespace. +// Similar to: `ip link set $link netns $ns` +func LinkSetNsFd(link Link, fd int) error { + base := link.Attrs() + ensureIndex(base) + req := nl.NewNetlinkRequest(syscall.RTM_SETLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + msg.Index = int32(base.Index) + req.AddData(msg) + + b := make([]byte, 4) + native.PutUint32(b, uint32(fd)) + + data := nl.NewRtAttr(nl.IFLA_NET_NS_FD, b) + req.AddData(data) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +func boolAttr(val bool) []byte { + var v uint8 + if val { + v = 1 + } + return nl.Uint8Attr(v) +} + +type vxlanPortRange struct { + Lo, Hi uint16 +} + +func addVxlanAttrs(vxlan *Vxlan, linkInfo *nl.RtAttr) { + data := nl.NewRtAttrChild(linkInfo, nl.IFLA_INFO_DATA, nil) + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_ID, nl.Uint32Attr(uint32(vxlan.VxlanId))) + if vxlan.VtepDevIndex != 0 { + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_LINK, nl.Uint32Attr(uint32(vxlan.VtepDevIndex))) + } + if vxlan.SrcAddr != nil { + ip := vxlan.SrcAddr.To4() + if ip != nil { + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_LOCAL, []byte(ip)) + } else { + ip = vxlan.SrcAddr.To16() + if ip != nil { + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_LOCAL6, []byte(ip)) + } + } + } + if vxlan.Group != nil { + group := vxlan.Group.To4() + if group != nil { + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_GROUP, []byte(group)) + } else { + group = vxlan.Group.To16() + if group != nil { + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_GROUP6, []byte(group)) + } + } + } + + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_TTL, nl.Uint8Attr(uint8(vxlan.TTL))) + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_TOS, nl.Uint8Attr(uint8(vxlan.TOS))) + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_LEARNING, boolAttr(vxlan.Learning)) + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_PROXY, boolAttr(vxlan.Proxy)) + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_RSC, boolAttr(vxlan.RSC)) + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_L2MISS, boolAttr(vxlan.L2miss)) + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_L3MISS, boolAttr(vxlan.L3miss)) + + if vxlan.GBP { + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_GBP, boolAttr(vxlan.GBP)) + } + + if vxlan.NoAge { + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_AGEING, nl.Uint32Attr(0)) + } else if vxlan.Age > 0 { + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_AGEING, nl.Uint32Attr(uint32(vxlan.Age))) + } + if vxlan.Limit > 0 { + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_LIMIT, nl.Uint32Attr(uint32(vxlan.Limit))) + } + if vxlan.Port > 0 { + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_PORT, nl.Uint16Attr(uint16(vxlan.Port))) + } + if vxlan.PortLow > 0 || vxlan.PortHigh > 0 { + pr := vxlanPortRange{uint16(vxlan.PortLow), uint16(vxlan.PortHigh)} + + buf := new(bytes.Buffer) + binary.Write(buf, binary.BigEndian, &pr) + + nl.NewRtAttrChild(data, nl.IFLA_VXLAN_PORT_RANGE, buf.Bytes()) + } +} + +// LinkAdd adds a new link device. The type and features of the device +// are taken fromt the parameters in the link object. +// Equivalent to: `ip link add $link` +func LinkAdd(link Link) error { + // TODO: set mtu and hardware address + // TODO: support extra data for macvlan + base := link.Attrs() + + if base.Name == "" { + return fmt.Errorf("LinkAttrs.Name cannot be empty!") + } + + req := nl.NewNetlinkRequest(syscall.RTM_NEWLINK, syscall.NLM_F_CREATE|syscall.NLM_F_EXCL|syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + req.AddData(msg) + + if base.ParentIndex != 0 { + b := make([]byte, 4) + native.PutUint32(b, uint32(base.ParentIndex)) + data := nl.NewRtAttr(syscall.IFLA_LINK, b) + req.AddData(data) + } else if link.Type() == "ipvlan" { + return fmt.Errorf("Can't create ipvlan link without ParentIndex") + } + + nameData := nl.NewRtAttr(syscall.IFLA_IFNAME, nl.ZeroTerminated(base.Name)) + req.AddData(nameData) + + if base.MTU > 0 { + mtu := nl.NewRtAttr(syscall.IFLA_MTU, nl.Uint32Attr(uint32(base.MTU))) + req.AddData(mtu) + } + + if base.TxQLen >= 0 { + qlen := nl.NewRtAttr(syscall.IFLA_TXQLEN, nl.Uint32Attr(uint32(base.TxQLen))) + req.AddData(qlen) + } + + if base.Namespace != nil { + var attr *nl.RtAttr + switch base.Namespace.(type) { + case NsPid: + val := nl.Uint32Attr(uint32(base.Namespace.(NsPid))) + attr = nl.NewRtAttr(syscall.IFLA_NET_NS_PID, val) + case NsFd: + val := nl.Uint32Attr(uint32(base.Namespace.(NsFd))) + attr = nl.NewRtAttr(nl.IFLA_NET_NS_FD, val) + } + + req.AddData(attr) + } + + linkInfo := nl.NewRtAttr(syscall.IFLA_LINKINFO, nil) + nl.NewRtAttrChild(linkInfo, nl.IFLA_INFO_KIND, nl.NonZeroTerminated(link.Type())) + + if vlan, ok := link.(*Vlan); ok { + b := make([]byte, 2) + native.PutUint16(b, uint16(vlan.VlanId)) + data := nl.NewRtAttrChild(linkInfo, nl.IFLA_INFO_DATA, nil) + nl.NewRtAttrChild(data, nl.IFLA_VLAN_ID, b) + } else if veth, ok := link.(*Veth); ok { + data := nl.NewRtAttrChild(linkInfo, nl.IFLA_INFO_DATA, nil) + peer := nl.NewRtAttrChild(data, nl.VETH_INFO_PEER, nil) + nl.NewIfInfomsgChild(peer, syscall.AF_UNSPEC) + nl.NewRtAttrChild(peer, syscall.IFLA_IFNAME, nl.ZeroTerminated(veth.PeerName)) + if base.TxQLen >= 0 { + nl.NewRtAttrChild(peer, syscall.IFLA_TXQLEN, nl.Uint32Attr(uint32(base.TxQLen))) + } + if base.MTU > 0 { + nl.NewRtAttrChild(peer, syscall.IFLA_MTU, nl.Uint32Attr(uint32(base.MTU))) + } + + } else if vxlan, ok := link.(*Vxlan); ok { + addVxlanAttrs(vxlan, linkInfo) + } else if ipv, ok := link.(*IPVlan); ok { + data := nl.NewRtAttrChild(linkInfo, nl.IFLA_INFO_DATA, nil) + nl.NewRtAttrChild(data, nl.IFLA_IPVLAN_MODE, nl.Uint16Attr(uint16(ipv.Mode))) + } else if macv, ok := link.(*Macvlan); ok { + if macv.Mode != MACVLAN_MODE_DEFAULT { + data := nl.NewRtAttrChild(linkInfo, nl.IFLA_INFO_DATA, nil) + nl.NewRtAttrChild(data, nl.IFLA_MACVLAN_MODE, nl.Uint32Attr(macvlanModes[macv.Mode])) + } + } + + req.AddData(linkInfo) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + if err != nil { + return err + } + + ensureIndex(base) + + // can't set master during create, so set it afterwards + if base.MasterIndex != 0 { + // TODO: verify MasterIndex is actually a bridge? + return LinkSetMasterByIndex(link, base.MasterIndex) + } + return nil +} + +// LinkDel deletes link device. Either Index or Name must be set in +// the link object for it to be deleted. The other values are ignored. +// Equivalent to: `ip link del $link` +func LinkDel(link Link) error { + base := link.Attrs() + + ensureIndex(base) + + req := nl.NewNetlinkRequest(syscall.RTM_DELLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + msg.Index = int32(base.Index) + req.AddData(msg) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +func linkByNameDump(name string) (Link, error) { + links, err := LinkList() + if err != nil { + return nil, err + } + + for _, link := range links { + if link.Attrs().Name == name { + return link, nil + } + } + return nil, fmt.Errorf("Link %s not found", name) +} + +// LinkByName finds a link by name and returns a pointer to the object. +func LinkByName(name string) (Link, error) { + if lookupByDump { + return linkByNameDump(name) + } + + req := nl.NewNetlinkRequest(syscall.RTM_GETLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + req.AddData(msg) + + nameData := nl.NewRtAttr(syscall.IFLA_IFNAME, nl.ZeroTerminated(name)) + req.AddData(nameData) + + link, err := execGetLink(req) + if err == syscall.EINVAL { + // older kernels don't support looking up via IFLA_IFNAME + // so fall back to dumping all links + lookupByDump = true + return linkByNameDump(name) + } + + return link, err +} + +// LinkByIndex finds a link by index and returns a pointer to the object. +func LinkByIndex(index int) (Link, error) { + req := nl.NewNetlinkRequest(syscall.RTM_GETLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + msg.Index = int32(index) + req.AddData(msg) + + return execGetLink(req) +} + +func execGetLink(req *nl.NetlinkRequest) (Link, error) { + msgs, err := req.Execute(syscall.NETLINK_ROUTE, 0) + if err != nil { + if errno, ok := err.(syscall.Errno); ok { + if errno == syscall.ENODEV { + return nil, fmt.Errorf("Link not found") + } + } + return nil, err + } + + switch { + case len(msgs) == 0: + return nil, fmt.Errorf("Link not found") + + case len(msgs) == 1: + return linkDeserialize(msgs[0]) + + default: + return nil, fmt.Errorf("More than one link found") + } +} + +// linkDeserialize deserializes a raw message received from netlink into +// a link object. +func linkDeserialize(m []byte) (Link, error) { + msg := nl.DeserializeIfInfomsg(m) + + attrs, err := nl.ParseRouteAttr(m[msg.Len():]) + if err != nil { + return nil, err + } + + base := LinkAttrs{Index: int(msg.Index), Flags: linkFlags(msg.Flags)} + var link Link + linkType := "" + for _, attr := range attrs { + switch attr.Attr.Type { + case syscall.IFLA_LINKINFO: + infos, err := nl.ParseRouteAttr(attr.Value) + if err != nil { + return nil, err + } + for _, info := range infos { + switch info.Attr.Type { + case nl.IFLA_INFO_KIND: + linkType = string(info.Value[:len(info.Value)-1]) + switch linkType { + case "dummy": + link = &Dummy{} + case "ifb": + link = &Ifb{} + case "bridge": + link = &Bridge{} + case "vlan": + link = &Vlan{} + case "veth": + link = &Veth{} + case "vxlan": + link = &Vxlan{} + case "ipvlan": + link = &IPVlan{} + case "macvlan": + link = &Macvlan{} + case "macvtap": + link = &Macvtap{} + default: + link = &GenericLink{LinkType: linkType} + } + case nl.IFLA_INFO_DATA: + data, err := nl.ParseRouteAttr(info.Value) + if err != nil { + return nil, err + } + switch linkType { + case "vlan": + parseVlanData(link, data) + case "vxlan": + parseVxlanData(link, data) + case "ipvlan": + parseIPVlanData(link, data) + case "macvlan": + parseMacvlanData(link, data) + case "macvtap": + parseMacvtapData(link, data) + } + } + } + case syscall.IFLA_ADDRESS: + var nonzero bool + for _, b := range attr.Value { + if b != 0 { + nonzero = true + } + } + if nonzero { + base.HardwareAddr = attr.Value[:] + } + case syscall.IFLA_IFNAME: + base.Name = string(attr.Value[:len(attr.Value)-1]) + case syscall.IFLA_MTU: + base.MTU = int(native.Uint32(attr.Value[0:4])) + case syscall.IFLA_LINK: + base.ParentIndex = int(native.Uint32(attr.Value[0:4])) + case syscall.IFLA_MASTER: + base.MasterIndex = int(native.Uint32(attr.Value[0:4])) + case syscall.IFLA_TXQLEN: + base.TxQLen = int(native.Uint32(attr.Value[0:4])) + } + } + // Links that don't have IFLA_INFO_KIND are hardware devices + if link == nil { + link = &Device{} + } + *link.Attrs() = base + + return link, nil +} + +// LinkList gets a list of link devices. +// Equivalent to: `ip link show` +func LinkList() ([]Link, error) { + // NOTE(vish): This duplicates functionality in net/iface_linux.go, but we need + // to get the message ourselves to parse link type. + req := nl.NewNetlinkRequest(syscall.RTM_GETLINK, syscall.NLM_F_DUMP) + + msg := nl.NewIfInfomsg(syscall.AF_UNSPEC) + req.AddData(msg) + + msgs, err := req.Execute(syscall.NETLINK_ROUTE, syscall.RTM_NEWLINK) + if err != nil { + return nil, err + } + + var res []Link + for _, m := range msgs { + link, err := linkDeserialize(m) + if err != nil { + return nil, err + } + res = append(res, link) + } + + return res, nil +} + +func LinkSetHairpin(link Link, mode bool) error { + return setProtinfoAttr(link, mode, nl.IFLA_BRPORT_MODE) +} + +func LinkSetGuard(link Link, mode bool) error { + return setProtinfoAttr(link, mode, nl.IFLA_BRPORT_GUARD) +} + +func LinkSetFastLeave(link Link, mode bool) error { + return setProtinfoAttr(link, mode, nl.IFLA_BRPORT_FAST_LEAVE) +} + +func LinkSetLearning(link Link, mode bool) error { + return setProtinfoAttr(link, mode, nl.IFLA_BRPORT_LEARNING) +} + +func LinkSetRootBlock(link Link, mode bool) error { + return setProtinfoAttr(link, mode, nl.IFLA_BRPORT_PROTECT) +} + +func LinkSetFlood(link Link, mode bool) error { + return setProtinfoAttr(link, mode, nl.IFLA_BRPORT_UNICAST_FLOOD) +} + +func setProtinfoAttr(link Link, mode bool, attr int) error { + base := link.Attrs() + ensureIndex(base) + req := nl.NewNetlinkRequest(syscall.RTM_SETLINK, syscall.NLM_F_ACK) + + msg := nl.NewIfInfomsg(syscall.AF_BRIDGE) + msg.Index = int32(base.Index) + req.AddData(msg) + + br := nl.NewRtAttr(syscall.IFLA_PROTINFO|syscall.NLA_F_NESTED, nil) + nl.NewRtAttrChild(br, attr, boolToByte(mode)) + req.AddData(br) + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + if err != nil { + return err + } + return nil +} + +func parseVlanData(link Link, data []syscall.NetlinkRouteAttr) { + vlan := link.(*Vlan) + for _, datum := range data { + switch datum.Attr.Type { + case nl.IFLA_VLAN_ID: + vlan.VlanId = int(native.Uint16(datum.Value[0:2])) + } + } +} + +func parseVxlanData(link Link, data []syscall.NetlinkRouteAttr) { + vxlan := link.(*Vxlan) + for _, datum := range data { + switch datum.Attr.Type { + case nl.IFLA_VXLAN_ID: + vxlan.VxlanId = int(native.Uint32(datum.Value[0:4])) + case nl.IFLA_VXLAN_LINK: + vxlan.VtepDevIndex = int(native.Uint32(datum.Value[0:4])) + case nl.IFLA_VXLAN_LOCAL: + vxlan.SrcAddr = net.IP(datum.Value[0:4]) + case nl.IFLA_VXLAN_LOCAL6: + vxlan.SrcAddr = net.IP(datum.Value[0:16]) + case nl.IFLA_VXLAN_GROUP: + vxlan.Group = net.IP(datum.Value[0:4]) + case nl.IFLA_VXLAN_GROUP6: + vxlan.Group = net.IP(datum.Value[0:16]) + case nl.IFLA_VXLAN_TTL: + vxlan.TTL = int(datum.Value[0]) + case nl.IFLA_VXLAN_TOS: + vxlan.TOS = int(datum.Value[0]) + case nl.IFLA_VXLAN_LEARNING: + vxlan.Learning = int8(datum.Value[0]) != 0 + case nl.IFLA_VXLAN_PROXY: + vxlan.Proxy = int8(datum.Value[0]) != 0 + case nl.IFLA_VXLAN_RSC: + vxlan.RSC = int8(datum.Value[0]) != 0 + case nl.IFLA_VXLAN_L2MISS: + vxlan.L2miss = int8(datum.Value[0]) != 0 + case nl.IFLA_VXLAN_L3MISS: + vxlan.L3miss = int8(datum.Value[0]) != 0 + case nl.IFLA_VXLAN_GBP: + vxlan.GBP = int8(datum.Value[0]) != 0 + case nl.IFLA_VXLAN_AGEING: + vxlan.Age = int(native.Uint32(datum.Value[0:4])) + vxlan.NoAge = vxlan.Age == 0 + case nl.IFLA_VXLAN_LIMIT: + vxlan.Limit = int(native.Uint32(datum.Value[0:4])) + case nl.IFLA_VXLAN_PORT: + vxlan.Port = int(native.Uint16(datum.Value[0:2])) + case nl.IFLA_VXLAN_PORT_RANGE: + buf := bytes.NewBuffer(datum.Value[0:4]) + var pr vxlanPortRange + if binary.Read(buf, binary.BigEndian, &pr) != nil { + vxlan.PortLow = int(pr.Lo) + vxlan.PortHigh = int(pr.Hi) + } + } + } +} + +func parseIPVlanData(link Link, data []syscall.NetlinkRouteAttr) { + ipv := link.(*IPVlan) + for _, datum := range data { + if datum.Attr.Type == nl.IFLA_IPVLAN_MODE { + ipv.Mode = IPVlanMode(native.Uint32(datum.Value[0:4])) + return + } + } +} + +func parseMacvtapData(link Link, data []syscall.NetlinkRouteAttr) { + macv := link.(*Macvtap) + parseMacvlanData(&macv.Macvlan, data) +} + +func parseMacvlanData(link Link, data []syscall.NetlinkRouteAttr) { + macv := link.(*Macvlan) + for _, datum := range data { + if datum.Attr.Type == nl.IFLA_MACVLAN_MODE { + switch native.Uint32(datum.Value[0:4]) { + case nl.MACVLAN_MODE_PRIVATE: + macv.Mode = MACVLAN_MODE_PRIVATE + case nl.MACVLAN_MODE_VEPA: + macv.Mode = MACVLAN_MODE_VEPA + case nl.MACVLAN_MODE_BRIDGE: + macv.Mode = MACVLAN_MODE_BRIDGE + case nl.MACVLAN_MODE_PASSTHRU: + macv.Mode = MACVLAN_MODE_PASSTHRU + case nl.MACVLAN_MODE_SOURCE: + macv.Mode = MACVLAN_MODE_SOURCE + } + return + } + } +} + +// copied from pkg/net_linux.go +func linkFlags(rawFlags uint32) net.Flags { + var f net.Flags + if rawFlags&syscall.IFF_UP != 0 { + f |= net.FlagUp + } + if rawFlags&syscall.IFF_BROADCAST != 0 { + f |= net.FlagBroadcast + } + if rawFlags&syscall.IFF_LOOPBACK != 0 { + f |= net.FlagLoopback + } + if rawFlags&syscall.IFF_POINTOPOINT != 0 { + f |= net.FlagPointToPoint + } + if rawFlags&syscall.IFF_MULTICAST != 0 { + f |= net.FlagMulticast + } + return f +} diff --git a/vendor/github.com/vishvananda/netlink/neigh.go b/vendor/github.com/vishvananda/netlink/neigh.go new file mode 100644 index 000000000..0e5eb90c9 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/neigh.go @@ -0,0 +1,22 @@ +package netlink + +import ( + "fmt" + "net" +) + +// Neigh represents a link layer neighbor from netlink. +type Neigh struct { + LinkIndex int + Family int + State int + Type int + Flags int + IP net.IP + HardwareAddr net.HardwareAddr +} + +// String returns $ip/$hwaddr $label +func (neigh *Neigh) String() string { + return fmt.Sprintf("%s %s", neigh.IP, neigh.HardwareAddr) +} diff --git a/vendor/github.com/vishvananda/netlink/neigh_linux.go b/vendor/github.com/vishvananda/netlink/neigh_linux.go new file mode 100644 index 000000000..620a0ee70 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/neigh_linux.go @@ -0,0 +1,189 @@ +package netlink + +import ( + "net" + "syscall" + "unsafe" + + "github.com/vishvananda/netlink/nl" +) + +const ( + NDA_UNSPEC = iota + NDA_DST + NDA_LLADDR + NDA_CACHEINFO + NDA_PROBES + NDA_VLAN + NDA_PORT + NDA_VNI + NDA_IFINDEX + NDA_MAX = NDA_IFINDEX +) + +// Neighbor Cache Entry States. +const ( + NUD_NONE = 0x00 + NUD_INCOMPLETE = 0x01 + NUD_REACHABLE = 0x02 + NUD_STALE = 0x04 + NUD_DELAY = 0x08 + NUD_PROBE = 0x10 + NUD_FAILED = 0x20 + NUD_NOARP = 0x40 + NUD_PERMANENT = 0x80 +) + +// Neighbor Flags +const ( + NTF_USE = 0x01 + NTF_SELF = 0x02 + NTF_MASTER = 0x04 + NTF_PROXY = 0x08 + NTF_ROUTER = 0x80 +) + +type Ndmsg struct { + Family uint8 + Index uint32 + State uint16 + Flags uint8 + Type uint8 +} + +func deserializeNdmsg(b []byte) *Ndmsg { + var dummy Ndmsg + return (*Ndmsg)(unsafe.Pointer(&b[0:unsafe.Sizeof(dummy)][0])) +} + +func (msg *Ndmsg) Serialize() []byte { + return (*(*[unsafe.Sizeof(*msg)]byte)(unsafe.Pointer(msg)))[:] +} + +func (msg *Ndmsg) Len() int { + return int(unsafe.Sizeof(*msg)) +} + +// NeighAdd will add an IP to MAC mapping to the ARP table +// Equivalent to: `ip neigh add ....` +func NeighAdd(neigh *Neigh) error { + return neighAdd(neigh, syscall.NLM_F_CREATE|syscall.NLM_F_EXCL) +} + +// NeighAdd will add or replace an IP to MAC mapping to the ARP table +// Equivalent to: `ip neigh replace....` +func NeighSet(neigh *Neigh) error { + return neighAdd(neigh, syscall.NLM_F_CREATE) +} + +// NeighAppend will append an entry to FDB +// Equivalent to: `bridge fdb append...` +func NeighAppend(neigh *Neigh) error { + return neighAdd(neigh, syscall.NLM_F_CREATE|syscall.NLM_F_APPEND) +} + +func neighAdd(neigh *Neigh, mode int) error { + req := nl.NewNetlinkRequest(syscall.RTM_NEWNEIGH, mode|syscall.NLM_F_ACK) + return neighHandle(neigh, req) +} + +// NeighDel will delete an IP address from a link device. +// Equivalent to: `ip addr del $addr dev $link` +func NeighDel(neigh *Neigh) error { + req := nl.NewNetlinkRequest(syscall.RTM_DELNEIGH, syscall.NLM_F_ACK) + return neighHandle(neigh, req) +} + +func neighHandle(neigh *Neigh, req *nl.NetlinkRequest) error { + var family int + if neigh.Family > 0 { + family = neigh.Family + } else { + family = nl.GetIPFamily(neigh.IP) + } + + msg := Ndmsg{ + Family: uint8(family), + Index: uint32(neigh.LinkIndex), + State: uint16(neigh.State), + Type: uint8(neigh.Type), + Flags: uint8(neigh.Flags), + } + req.AddData(&msg) + + ipData := neigh.IP.To4() + if ipData == nil { + ipData = neigh.IP.To16() + } + + dstData := nl.NewRtAttr(NDA_DST, ipData) + req.AddData(dstData) + + hwData := nl.NewRtAttr(NDA_LLADDR, []byte(neigh.HardwareAddr)) + req.AddData(hwData) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// NeighList gets a list of IP-MAC mappings in the system (ARP table). +// Equivalent to: `ip neighbor show`. +// The list can be filtered by link and ip family. +func NeighList(linkIndex, family int) ([]Neigh, error) { + req := nl.NewNetlinkRequest(syscall.RTM_GETNEIGH, syscall.NLM_F_DUMP) + msg := Ndmsg{ + Family: uint8(family), + } + req.AddData(&msg) + + msgs, err := req.Execute(syscall.NETLINK_ROUTE, syscall.RTM_NEWNEIGH) + if err != nil { + return nil, err + } + + var res []Neigh + for _, m := range msgs { + ndm := deserializeNdmsg(m) + if linkIndex != 0 && int(ndm.Index) != linkIndex { + // Ignore messages from other interfaces + continue + } + + neigh, err := NeighDeserialize(m) + if err != nil { + continue + } + + res = append(res, *neigh) + } + + return res, nil +} + +func NeighDeserialize(m []byte) (*Neigh, error) { + msg := deserializeNdmsg(m) + + neigh := Neigh{ + LinkIndex: int(msg.Index), + Family: int(msg.Family), + State: int(msg.State), + Type: int(msg.Type), + Flags: int(msg.Flags), + } + + attrs, err := nl.ParseRouteAttr(m[msg.Len():]) + if err != nil { + return nil, err + } + + for _, attr := range attrs { + switch attr.Attr.Type { + case NDA_DST: + neigh.IP = net.IP(attr.Value) + case NDA_LLADDR: + neigh.HardwareAddr = net.HardwareAddr(attr.Value) + } + } + + return &neigh, nil +} diff --git a/vendor/github.com/vishvananda/netlink/netlink.go b/vendor/github.com/vishvananda/netlink/netlink.go new file mode 100644 index 000000000..41ebdb11f --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/netlink.go @@ -0,0 +1,39 @@ +// Package netlink provides a simple library for netlink. Netlink is +// the interface a user-space program in linux uses to communicate with +// the kernel. It can be used to add and remove interfaces, set up ip +// addresses and routes, and confiugre ipsec. Netlink communication +// requires elevated privileges, so in most cases this code needs to +// be run as root. The low level primitives for netlink are contained +// in the nl subpackage. This package attempts to provide a high-level +// interface that is loosly modeled on the iproute2 cli. +package netlink + +import ( + "net" + + "github.com/vishvananda/netlink/nl" +) + +const ( + // Family type definitions + FAMILY_ALL = nl.FAMILY_ALL + FAMILY_V4 = nl.FAMILY_V4 + FAMILY_V6 = nl.FAMILY_V6 +) + +// ParseIPNet parses a string in ip/net format and returns a net.IPNet. +// This is valuable because addresses in netlink are often IPNets and +// ParseCIDR returns an IPNet with the IP part set to the base IP of the +// range. +func ParseIPNet(s string) (*net.IPNet, error) { + ip, ipNet, err := net.ParseCIDR(s) + if err != nil { + return nil, err + } + return &net.IPNet{IP: ip, Mask: ipNet.Mask}, nil +} + +// NewIPNet generates an IPNet from an ip address using a netmask of 32. +func NewIPNet(ip net.IP) *net.IPNet { + return &net.IPNet{IP: ip, Mask: net.CIDRMask(32, 32)} +} diff --git a/vendor/github.com/vishvananda/netlink/netlink_unspecified.go b/vendor/github.com/vishvananda/netlink/netlink_unspecified.go new file mode 100644 index 000000000..10c49c1bf --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/netlink_unspecified.go @@ -0,0 +1,143 @@ +// +build !linux + +package netlink + +import ( + "errors" +) + +var ( + ErrNotImplemented = errors.New("not implemented") +) + +func LinkSetUp(link *Link) error { + return ErrNotImplemented +} + +func LinkSetDown(link *Link) error { + return ErrNotImplemented +} + +func LinkSetMTU(link *Link, mtu int) error { + return ErrNotImplemented +} + +func LinkSetMaster(link *Link, master *Link) error { + return ErrNotImplemented +} + +func LinkSetNsPid(link *Link, nspid int) error { + return ErrNotImplemented +} + +func LinkSetNsFd(link *Link, fd int) error { + return ErrNotImplemented +} + +func LinkAdd(link *Link) error { + return ErrNotImplemented +} + +func LinkDel(link *Link) error { + return ErrNotImplemented +} + +func SetHairpin(link Link, mode bool) error { + return ErrNotImplemented +} + +func SetGuard(link Link, mode bool) error { + return ErrNotImplemented +} + +func SetFastLeave(link Link, mode bool) error { + return ErrNotImplemented +} + +func SetLearning(link Link, mode bool) error { + return ErrNotImplemented +} + +func SetRootBlock(link Link, mode bool) error { + return ErrNotImplemented +} + +func SetFlood(link Link, mode bool) error { + return ErrNotImplemented +} + +func LinkList() ([]Link, error) { + return nil, ErrNotImplemented +} + +func AddrAdd(link *Link, addr *Addr) error { + return ErrNotImplemented +} + +func AddrDel(link *Link, addr *Addr) error { + return ErrNotImplemented +} + +func AddrList(link *Link, family int) ([]Addr, error) { + return nil, ErrNotImplemented +} + +func RouteAdd(route *Route) error { + return ErrNotImplemented +} + +func RouteDel(route *Route) error { + return ErrNotImplemented +} + +func RouteList(link *Link, family int) ([]Route, error) { + return nil, ErrNotImplemented +} + +func XfrmPolicyAdd(policy *XfrmPolicy) error { + return ErrNotImplemented +} + +func XfrmPolicyDel(policy *XfrmPolicy) error { + return ErrNotImplemented +} + +func XfrmPolicyList(family int) ([]XfrmPolicy, error) { + return nil, ErrNotImplemented +} + +func XfrmStateAdd(policy *XfrmState) error { + return ErrNotImplemented +} + +func XfrmStateDel(policy *XfrmState) error { + return ErrNotImplemented +} + +func XfrmStateList(family int) ([]XfrmState, error) { + return nil, ErrNotImplemented +} + +func NeighAdd(neigh *Neigh) error { + return ErrNotImplemented +} + +func NeighSet(neigh *Neigh) error { + return ErrNotImplemented +} + +func NeighAppend(neigh *Neigh) error { + return ErrNotImplemented +} + +func NeighDel(neigh *Neigh) error { + return ErrNotImplemented +} + +func NeighList(linkIndex, family int) ([]Neigh, error) { + return nil, ErrNotImplemented +} + +func NeighDeserialize(m []byte) (*Ndmsg, *Neigh, error) { + return nil, nil, ErrNotImplemented +} diff --git a/vendor/github.com/vishvananda/netlink/nl/addr_linux.go b/vendor/github.com/vishvananda/netlink/nl/addr_linux.go new file mode 100644 index 000000000..17088fa0c --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/nl/addr_linux.go @@ -0,0 +1,47 @@ +package nl + +import ( + "syscall" + "unsafe" +) + +type IfAddrmsg struct { + syscall.IfAddrmsg +} + +func NewIfAddrmsg(family int) *IfAddrmsg { + return &IfAddrmsg{ + IfAddrmsg: syscall.IfAddrmsg{ + Family: uint8(family), + }, + } +} + +// struct ifaddrmsg { +// __u8 ifa_family; +// __u8 ifa_prefixlen; /* The prefix length */ +// __u8 ifa_flags; /* Flags */ +// __u8 ifa_scope; /* Address scope */ +// __u32 ifa_index; /* Link index */ +// }; + +// type IfAddrmsg struct { +// Family uint8 +// Prefixlen uint8 +// Flags uint8 +// Scope uint8 +// Index uint32 +// } +// SizeofIfAddrmsg = 0x8 + +func DeserializeIfAddrmsg(b []byte) *IfAddrmsg { + return (*IfAddrmsg)(unsafe.Pointer(&b[0:syscall.SizeofIfAddrmsg][0])) +} + +func (msg *IfAddrmsg) Serialize() []byte { + return (*(*[syscall.SizeofIfAddrmsg]byte)(unsafe.Pointer(msg)))[:] +} + +func (msg *IfAddrmsg) Len() int { + return syscall.SizeofIfAddrmsg +} diff --git a/vendor/github.com/vishvananda/netlink/nl/link_linux.go b/vendor/github.com/vishvananda/netlink/nl/link_linux.go new file mode 100644 index 000000000..1f9ab088f --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/nl/link_linux.go @@ -0,0 +1,104 @@ +package nl + +const ( + DEFAULT_CHANGE = 0xFFFFFFFF +) + +const ( + IFLA_INFO_UNSPEC = iota + IFLA_INFO_KIND + IFLA_INFO_DATA + IFLA_INFO_XSTATS + IFLA_INFO_MAX = IFLA_INFO_XSTATS +) + +const ( + IFLA_VLAN_UNSPEC = iota + IFLA_VLAN_ID + IFLA_VLAN_FLAGS + IFLA_VLAN_EGRESS_QOS + IFLA_VLAN_INGRESS_QOS + IFLA_VLAN_PROTOCOL + IFLA_VLAN_MAX = IFLA_VLAN_PROTOCOL +) + +const ( + VETH_INFO_UNSPEC = iota + VETH_INFO_PEER + VETH_INFO_MAX = VETH_INFO_PEER +) + +const ( + IFLA_VXLAN_UNSPEC = iota + IFLA_VXLAN_ID + IFLA_VXLAN_GROUP + IFLA_VXLAN_LINK + IFLA_VXLAN_LOCAL + IFLA_VXLAN_TTL + IFLA_VXLAN_TOS + IFLA_VXLAN_LEARNING + IFLA_VXLAN_AGEING + IFLA_VXLAN_LIMIT + IFLA_VXLAN_PORT_RANGE + IFLA_VXLAN_PROXY + IFLA_VXLAN_RSC + IFLA_VXLAN_L2MISS + IFLA_VXLAN_L3MISS + IFLA_VXLAN_PORT + IFLA_VXLAN_GROUP6 + IFLA_VXLAN_LOCAL6 + IFLA_VXLAN_UDP_CSUM + IFLA_VXLAN_UDP_ZERO_CSUM6_TX + IFLA_VXLAN_UDP_ZERO_CSUM6_RX + IFLA_VXLAN_REMCSUM_TX + IFLA_VXLAN_REMCSUM_RX + IFLA_VXLAN_GBP + IFLA_VXLAN_REMCSUM_NOPARTIAL + IFLA_VXLAN_FLOWBASED + IFLA_VXLAN_MAX = IFLA_VXLAN_FLOWBASED +) + +const ( + BRIDGE_MODE_UNSPEC = iota + BRIDGE_MODE_HAIRPIN +) + +const ( + IFLA_BRPORT_UNSPEC = iota + IFLA_BRPORT_STATE + IFLA_BRPORT_PRIORITY + IFLA_BRPORT_COST + IFLA_BRPORT_MODE + IFLA_BRPORT_GUARD + IFLA_BRPORT_PROTECT + IFLA_BRPORT_FAST_LEAVE + IFLA_BRPORT_LEARNING + IFLA_BRPORT_UNICAST_FLOOD + IFLA_BRPORT_MAX = IFLA_BRPORT_UNICAST_FLOOD +) + +const ( + IFLA_IPVLAN_UNSPEC = iota + IFLA_IPVLAN_MODE + IFLA_IPVLAN_MAX = IFLA_IPVLAN_MODE +) + +const ( + // not defined in syscall + IFLA_NET_NS_FD = 28 +) + +const ( + IFLA_MACVLAN_UNSPEC = iota + IFLA_MACVLAN_MODE + IFLA_MACVLAN_FLAGS + IFLA_MACVLAN_MAX = IFLA_MACVLAN_FLAGS +) + +const ( + MACVLAN_MODE_PRIVATE = 1 + MACVLAN_MODE_VEPA = 2 + MACVLAN_MODE_BRIDGE = 4 + MACVLAN_MODE_PASSTHRU = 8 + MACVLAN_MODE_SOURCE = 16 +) diff --git a/vendor/github.com/vishvananda/netlink/nl/nl_linux.go b/vendor/github.com/vishvananda/netlink/nl/nl_linux.go new file mode 100644 index 000000000..8dbd92b81 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/nl/nl_linux.go @@ -0,0 +1,418 @@ +// Package nl has low level primitives for making Netlink calls. +package nl + +import ( + "bytes" + "encoding/binary" + "fmt" + "net" + "sync/atomic" + "syscall" + "unsafe" +) + +const ( + // Family type definitions + FAMILY_ALL = syscall.AF_UNSPEC + FAMILY_V4 = syscall.AF_INET + FAMILY_V6 = syscall.AF_INET6 +) + +var nextSeqNr uint32 + +// GetIPFamily returns the family type of a net.IP. +func GetIPFamily(ip net.IP) int { + if len(ip) <= net.IPv4len { + return FAMILY_V4 + } + if ip.To4() != nil { + return FAMILY_V4 + } + return FAMILY_V6 +} + +var nativeEndian binary.ByteOrder + +// Get native endianness for the system +func NativeEndian() binary.ByteOrder { + if nativeEndian == nil { + var x uint32 = 0x01020304 + if *(*byte)(unsafe.Pointer(&x)) == 0x01 { + nativeEndian = binary.BigEndian + } else { + nativeEndian = binary.LittleEndian + } + } + return nativeEndian +} + +// Byte swap a 16 bit value if we aren't big endian +func Swap16(i uint16) uint16 { + if NativeEndian() == binary.BigEndian { + return i + } + return (i&0xff00)>>8 | (i&0xff)<<8 +} + +// Byte swap a 32 bit value if aren't big endian +func Swap32(i uint32) uint32 { + if NativeEndian() == binary.BigEndian { + return i + } + return (i&0xff000000)>>24 | (i&0xff0000)>>8 | (i&0xff00)<<8 | (i&0xff)<<24 +} + +type NetlinkRequestData interface { + Len() int + Serialize() []byte +} + +// IfInfomsg is related to links, but it is used for list requests as well +type IfInfomsg struct { + syscall.IfInfomsg +} + +// Create an IfInfomsg with family specified +func NewIfInfomsg(family int) *IfInfomsg { + return &IfInfomsg{ + IfInfomsg: syscall.IfInfomsg{ + Family: uint8(family), + }, + } +} + +func DeserializeIfInfomsg(b []byte) *IfInfomsg { + return (*IfInfomsg)(unsafe.Pointer(&b[0:syscall.SizeofIfInfomsg][0])) +} + +func (msg *IfInfomsg) Serialize() []byte { + return (*(*[syscall.SizeofIfInfomsg]byte)(unsafe.Pointer(msg)))[:] +} + +func (msg *IfInfomsg) Len() int { + return syscall.SizeofIfInfomsg +} + +func rtaAlignOf(attrlen int) int { + return (attrlen + syscall.RTA_ALIGNTO - 1) & ^(syscall.RTA_ALIGNTO - 1) +} + +func NewIfInfomsgChild(parent *RtAttr, family int) *IfInfomsg { + msg := NewIfInfomsg(family) + parent.children = append(parent.children, msg) + return msg +} + +// Extend RtAttr to handle data and children +type RtAttr struct { + syscall.RtAttr + Data []byte + children []NetlinkRequestData +} + +// Create a new Extended RtAttr object +func NewRtAttr(attrType int, data []byte) *RtAttr { + return &RtAttr{ + RtAttr: syscall.RtAttr{ + Type: uint16(attrType), + }, + children: []NetlinkRequestData{}, + Data: data, + } +} + +// Create a new RtAttr obj anc add it as a child of an existing object +func NewRtAttrChild(parent *RtAttr, attrType int, data []byte) *RtAttr { + attr := NewRtAttr(attrType, data) + parent.children = append(parent.children, attr) + return attr +} + +func (a *RtAttr) Len() int { + if len(a.children) == 0 { + return (syscall.SizeofRtAttr + len(a.Data)) + } + + l := 0 + for _, child := range a.children { + l += rtaAlignOf(child.Len()) + } + l += syscall.SizeofRtAttr + return rtaAlignOf(l + len(a.Data)) +} + +// Serialize the RtAttr into a byte array +// This can't just unsafe.cast because it must iterate through children. +func (a *RtAttr) Serialize() []byte { + native := NativeEndian() + + length := a.Len() + buf := make([]byte, rtaAlignOf(length)) + + if a.Data != nil { + copy(buf[4:], a.Data) + } else { + next := 4 + for _, child := range a.children { + childBuf := child.Serialize() + copy(buf[next:], childBuf) + next += rtaAlignOf(len(childBuf)) + } + } + + if l := uint16(length); l != 0 { + native.PutUint16(buf[0:2], l) + } + native.PutUint16(buf[2:4], a.Type) + return buf +} + +type NetlinkRequest struct { + syscall.NlMsghdr + Data []NetlinkRequestData +} + +// Serialize the Netlink Request into a byte array +func (req *NetlinkRequest) Serialize() []byte { + length := syscall.SizeofNlMsghdr + dataBytes := make([][]byte, len(req.Data)) + for i, data := range req.Data { + dataBytes[i] = data.Serialize() + length = length + len(dataBytes[i]) + } + req.Len = uint32(length) + b := make([]byte, length) + hdr := (*(*[syscall.SizeofNlMsghdr]byte)(unsafe.Pointer(req)))[:] + next := syscall.SizeofNlMsghdr + copy(b[0:next], hdr) + for _, data := range dataBytes { + for _, dataByte := range data { + b[next] = dataByte + next = next + 1 + } + } + return b +} + +func (req *NetlinkRequest) AddData(data NetlinkRequestData) { + if data != nil { + req.Data = append(req.Data, data) + } +} + +// Execute the request against a the given sockType. +// Returns a list of netlink messages in seriaized format, optionally filtered +// by resType. +func (req *NetlinkRequest) Execute(sockType int, resType uint16) ([][]byte, error) { + s, err := getNetlinkSocket(sockType) + if err != nil { + return nil, err + } + defer s.Close() + + if err := s.Send(req); err != nil { + return nil, err + } + + pid, err := s.GetPid() + if err != nil { + return nil, err + } + + var res [][]byte + +done: + for { + msgs, err := s.Receive() + if err != nil { + return nil, err + } + for _, m := range msgs { + if m.Header.Seq != req.Seq { + return nil, fmt.Errorf("Wrong Seq nr %d, expected 1", m.Header.Seq) + } + if m.Header.Pid != pid { + return nil, fmt.Errorf("Wrong pid %d, expected %d", m.Header.Pid, pid) + } + if m.Header.Type == syscall.NLMSG_DONE { + break done + } + if m.Header.Type == syscall.NLMSG_ERROR { + native := NativeEndian() + error := int32(native.Uint32(m.Data[0:4])) + if error == 0 { + break done + } + return nil, syscall.Errno(-error) + } + if resType != 0 && m.Header.Type != resType { + continue + } + res = append(res, m.Data) + if m.Header.Flags&syscall.NLM_F_MULTI == 0 { + break done + } + } + } + return res, nil +} + +// Create a new netlink request from proto and flags +// Note the Len value will be inaccurate once data is added until +// the message is serialized +func NewNetlinkRequest(proto, flags int) *NetlinkRequest { + return &NetlinkRequest{ + NlMsghdr: syscall.NlMsghdr{ + Len: uint32(syscall.SizeofNlMsghdr), + Type: uint16(proto), + Flags: syscall.NLM_F_REQUEST | uint16(flags), + Seq: atomic.AddUint32(&nextSeqNr, 1), + }, + } +} + +type NetlinkSocket struct { + fd int + lsa syscall.SockaddrNetlink +} + +func getNetlinkSocket(protocol int) (*NetlinkSocket, error) { + fd, err := syscall.Socket(syscall.AF_NETLINK, syscall.SOCK_RAW, protocol) + if err != nil { + return nil, err + } + s := &NetlinkSocket{ + fd: fd, + } + s.lsa.Family = syscall.AF_NETLINK + if err := syscall.Bind(fd, &s.lsa); err != nil { + syscall.Close(fd) + return nil, err + } + + return s, nil +} + +// Create a netlink socket with a given protocol (e.g. NETLINK_ROUTE) +// and subscribe it to multicast groups passed in variable argument list. +// Returns the netlink socket on which Receive() method can be called +// to retrieve the messages from the kernel. +func Subscribe(protocol int, groups ...uint) (*NetlinkSocket, error) { + fd, err := syscall.Socket(syscall.AF_NETLINK, syscall.SOCK_RAW, protocol) + if err != nil { + return nil, err + } + s := &NetlinkSocket{ + fd: fd, + } + s.lsa.Family = syscall.AF_NETLINK + + for _, g := range groups { + s.lsa.Groups |= (1 << (g - 1)) + } + + if err := syscall.Bind(fd, &s.lsa); err != nil { + syscall.Close(fd) + return nil, err + } + + return s, nil +} + +func (s *NetlinkSocket) Close() { + syscall.Close(s.fd) +} + +func (s *NetlinkSocket) Send(request *NetlinkRequest) error { + if err := syscall.Sendto(s.fd, request.Serialize(), 0, &s.lsa); err != nil { + return err + } + return nil +} + +func (s *NetlinkSocket) Receive() ([]syscall.NetlinkMessage, error) { + rb := make([]byte, syscall.Getpagesize()) + nr, _, err := syscall.Recvfrom(s.fd, rb, 0) + if err != nil { + return nil, err + } + if nr < syscall.NLMSG_HDRLEN { + return nil, fmt.Errorf("Got short response from netlink") + } + rb = rb[:nr] + return syscall.ParseNetlinkMessage(rb) +} + +func (s *NetlinkSocket) GetPid() (uint32, error) { + lsa, err := syscall.Getsockname(s.fd) + if err != nil { + return 0, err + } + switch v := lsa.(type) { + case *syscall.SockaddrNetlink: + return v.Pid, nil + } + return 0, fmt.Errorf("Wrong socket type") +} + +func ZeroTerminated(s string) []byte { + bytes := make([]byte, len(s)+1) + for i := 0; i < len(s); i++ { + bytes[i] = s[i] + } + bytes[len(s)] = 0 + return bytes +} + +func NonZeroTerminated(s string) []byte { + bytes := make([]byte, len(s)) + for i := 0; i < len(s); i++ { + bytes[i] = s[i] + } + return bytes +} + +func BytesToString(b []byte) string { + n := bytes.Index(b, []byte{0}) + return string(b[:n]) +} + +func Uint8Attr(v uint8) []byte { + return []byte{byte(v)} +} + +func Uint16Attr(v uint16) []byte { + native := NativeEndian() + bytes := make([]byte, 2) + native.PutUint16(bytes, v) + return bytes +} + +func Uint32Attr(v uint32) []byte { + native := NativeEndian() + bytes := make([]byte, 4) + native.PutUint32(bytes, v) + return bytes +} + +func ParseRouteAttr(b []byte) ([]syscall.NetlinkRouteAttr, error) { + var attrs []syscall.NetlinkRouteAttr + for len(b) >= syscall.SizeofRtAttr { + a, vbuf, alen, err := netlinkRouteAttrAndValue(b) + if err != nil { + return nil, err + } + ra := syscall.NetlinkRouteAttr{Attr: *a, Value: vbuf[:int(a.Len)-syscall.SizeofRtAttr]} + attrs = append(attrs, ra) + b = b[alen:] + } + return attrs, nil +} + +func netlinkRouteAttrAndValue(b []byte) (*syscall.RtAttr, []byte, int, error) { + a := (*syscall.RtAttr)(unsafe.Pointer(&b[0])) + if int(a.Len) < syscall.SizeofRtAttr || int(a.Len) > len(b) { + return nil, nil, 0, syscall.EINVAL + } + return a, b[syscall.SizeofRtAttr:], rtaAlignOf(int(a.Len)), nil +} diff --git a/vendor/github.com/vishvananda/netlink/nl/route_linux.go b/vendor/github.com/vishvananda/netlink/nl/route_linux.go new file mode 100644 index 000000000..447e83e5a --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/nl/route_linux.go @@ -0,0 +1,42 @@ +package nl + +import ( + "syscall" + "unsafe" +) + +type RtMsg struct { + syscall.RtMsg +} + +func NewRtMsg() *RtMsg { + return &RtMsg{ + RtMsg: syscall.RtMsg{ + Table: syscall.RT_TABLE_MAIN, + Scope: syscall.RT_SCOPE_UNIVERSE, + Protocol: syscall.RTPROT_BOOT, + Type: syscall.RTN_UNICAST, + }, + } +} + +func NewRtDelMsg() *RtMsg { + return &RtMsg{ + RtMsg: syscall.RtMsg{ + Table: syscall.RT_TABLE_MAIN, + Scope: syscall.RT_SCOPE_NOWHERE, + }, + } +} + +func (msg *RtMsg) Len() int { + return syscall.SizeofRtMsg +} + +func DeserializeRtMsg(b []byte) *RtMsg { + return (*RtMsg)(unsafe.Pointer(&b[0:syscall.SizeofRtMsg][0])) +} + +func (msg *RtMsg) Serialize() []byte { + return (*(*[syscall.SizeofRtMsg]byte)(unsafe.Pointer(msg)))[:] +} diff --git a/vendor/github.com/vishvananda/netlink/nl/tc_linux.go b/vendor/github.com/vishvananda/netlink/nl/tc_linux.go new file mode 100644 index 000000000..c9bfe8dfd --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/nl/tc_linux.go @@ -0,0 +1,359 @@ +package nl + +import ( + "unsafe" +) + +// Message types +const ( + TCA_UNSPEC = iota + TCA_KIND + TCA_OPTIONS + TCA_STATS + TCA_XSTATS + TCA_RATE + TCA_FCNT + TCA_STATS2 + TCA_STAB + TCA_MAX = TCA_STAB +) + +const ( + TCA_ACT_TAB = 1 + TCAA_MAX = 1 +) + +const ( + TCA_PRIO_UNSPEC = iota + TCA_PRIO_MQ + TCA_PRIO_MAX = TCA_PRIO_MQ +) + +const ( + SizeofTcMsg = 0x14 + SizeofTcActionMsg = 0x04 + SizeofTcPrioMap = 0x14 + SizeofTcRateSpec = 0x0c + SizeofTcTbfQopt = 2*SizeofTcRateSpec + 0x0c + SizeofTcU32Key = 0x10 + SizeofTcU32Sel = 0x10 // without keys + SizeofTcMirred = 0x1c +) + +// struct tcmsg { +// unsigned char tcm_family; +// unsigned char tcm__pad1; +// unsigned short tcm__pad2; +// int tcm_ifindex; +// __u32 tcm_handle; +// __u32 tcm_parent; +// __u32 tcm_info; +// }; + +type TcMsg struct { + Family uint8 + Pad [3]byte + Ifindex int32 + Handle uint32 + Parent uint32 + Info uint32 +} + +func (msg *TcMsg) Len() int { + return SizeofTcMsg +} + +func DeserializeTcMsg(b []byte) *TcMsg { + return (*TcMsg)(unsafe.Pointer(&b[0:SizeofTcMsg][0])) +} + +func (x *TcMsg) Serialize() []byte { + return (*(*[SizeofTcMsg]byte)(unsafe.Pointer(x)))[:] +} + +// struct tcamsg { +// unsigned char tca_family; +// unsigned char tca__pad1; +// unsigned short tca__pad2; +// }; + +type TcActionMsg struct { + Family uint8 + Pad [3]byte +} + +func (msg *TcActionMsg) Len() int { + return SizeofTcActionMsg +} + +func DeserializeTcActionMsg(b []byte) *TcActionMsg { + return (*TcActionMsg)(unsafe.Pointer(&b[0:SizeofTcActionMsg][0])) +} + +func (x *TcActionMsg) Serialize() []byte { + return (*(*[SizeofTcActionMsg]byte)(unsafe.Pointer(x)))[:] +} + +const ( + TC_PRIO_MAX = 15 +) + +// struct tc_prio_qopt { +// int bands; /* Number of bands */ +// __u8 priomap[TC_PRIO_MAX+1]; /* Map: logical priority -> PRIO band */ +// }; + +type TcPrioMap struct { + Bands int32 + Priomap [TC_PRIO_MAX + 1]uint8 +} + +func (msg *TcPrioMap) Len() int { + return SizeofTcPrioMap +} + +func DeserializeTcPrioMap(b []byte) *TcPrioMap { + return (*TcPrioMap)(unsafe.Pointer(&b[0:SizeofTcPrioMap][0])) +} + +func (x *TcPrioMap) Serialize() []byte { + return (*(*[SizeofTcPrioMap]byte)(unsafe.Pointer(x)))[:] +} + +const ( + TCA_TBF_UNSPEC = iota + TCA_TBF_PARMS + TCA_TBF_RTAB + TCA_TBF_PTAB + TCA_TBF_RATE64 + TCA_TBF_PRATE64 + TCA_TBF_BURST + TCA_TBF_PBURST + TCA_TBF_MAX = TCA_TBF_PBURST +) + +// struct tc_ratespec { +// unsigned char cell_log; +// __u8 linklayer; /* lower 4 bits */ +// unsigned short overhead; +// short cell_align; +// unsigned short mpu; +// __u32 rate; +// }; + +type TcRateSpec struct { + CellLog uint8 + Linklayer uint8 + Overhead uint16 + CellAlign int16 + Mpu uint16 + Rate uint32 +} + +func (msg *TcRateSpec) Len() int { + return SizeofTcRateSpec +} + +func DeserializeTcRateSpec(b []byte) *TcRateSpec { + return (*TcRateSpec)(unsafe.Pointer(&b[0:SizeofTcRateSpec][0])) +} + +func (x *TcRateSpec) Serialize() []byte { + return (*(*[SizeofTcRateSpec]byte)(unsafe.Pointer(x)))[:] +} + +// struct tc_tbf_qopt { +// struct tc_ratespec rate; +// struct tc_ratespec peakrate; +// __u32 limit; +// __u32 buffer; +// __u32 mtu; +// }; + +type TcTbfQopt struct { + Rate TcRateSpec + Peakrate TcRateSpec + Limit uint32 + Buffer uint32 + Mtu uint32 +} + +func (msg *TcTbfQopt) Len() int { + return SizeofTcTbfQopt +} + +func DeserializeTcTbfQopt(b []byte) *TcTbfQopt { + return (*TcTbfQopt)(unsafe.Pointer(&b[0:SizeofTcTbfQopt][0])) +} + +func (x *TcTbfQopt) Serialize() []byte { + return (*(*[SizeofTcTbfQopt]byte)(unsafe.Pointer(x)))[:] +} + +const ( + TCA_U32_UNSPEC = iota + TCA_U32_CLASSID + TCA_U32_HASH + TCA_U32_LINK + TCA_U32_DIVISOR + TCA_U32_SEL + TCA_U32_POLICE + TCA_U32_ACT + TCA_U32_INDEV + TCA_U32_PCNT + TCA_U32_MARK + TCA_U32_MAX = TCA_U32_MARK +) + +// struct tc_u32_key { +// __be32 mask; +// __be32 val; +// int off; +// int offmask; +// }; + +type TcU32Key struct { + Mask uint32 // big endian + Val uint32 // big endian + Off int32 + OffMask int32 +} + +func (msg *TcU32Key) Len() int { + return SizeofTcU32Key +} + +func DeserializeTcU32Key(b []byte) *TcU32Key { + return (*TcU32Key)(unsafe.Pointer(&b[0:SizeofTcU32Key][0])) +} + +func (x *TcU32Key) Serialize() []byte { + return (*(*[SizeofTcU32Key]byte)(unsafe.Pointer(x)))[:] +} + +// struct tc_u32_sel { +// unsigned char flags; +// unsigned char offshift; +// unsigned char nkeys; +// +// __be16 offmask; +// __u16 off; +// short offoff; +// +// short hoff; +// __be32 hmask; +// struct tc_u32_key keys[0]; +// }; + +const ( + TC_U32_TERMINAL = 1 << iota + TC_U32_OFFSET = 1 << iota + TC_U32_VAROFFSET = 1 << iota + TC_U32_EAT = 1 << iota +) + +type TcU32Sel struct { + Flags uint8 + Offshift uint8 + Nkeys uint8 + Pad uint8 + Offmask uint16 // big endian + Off uint16 + Offoff int16 + Hoff int16 + Hmask uint32 // big endian + Keys []TcU32Key +} + +func (msg *TcU32Sel) Len() int { + return SizeofTcU32Sel + int(msg.Nkeys)*SizeofTcU32Key +} + +func DeserializeTcU32Sel(b []byte) *TcU32Sel { + x := &TcU32Sel{} + copy((*(*[SizeofTcU32Sel]byte)(unsafe.Pointer(x)))[:], b) + next := SizeofTcU32Sel + var i uint8 + for i = 0; i < x.Nkeys; i++ { + x.Keys = append(x.Keys, *DeserializeTcU32Key(b[next:])) + next += SizeofTcU32Key + } + return x +} + +func (x *TcU32Sel) Serialize() []byte { + // This can't just unsafe.cast because it must iterate through keys. + buf := make([]byte, x.Len()) + copy(buf, (*(*[SizeofTcU32Sel]byte)(unsafe.Pointer(x)))[:]) + next := SizeofTcU32Sel + for _, key := range x.Keys { + keyBuf := key.Serialize() + copy(buf[next:], keyBuf) + next += SizeofTcU32Key + } + return buf +} + +const ( + TCA_ACT_MIRRED = 8 +) + +const ( + TCA_MIRRED_UNSPEC = iota + TCA_MIRRED_TM + TCA_MIRRED_PARMS + TCA_MIRRED_MAX = TCA_MIRRED_PARMS +) + +const ( + TCA_EGRESS_REDIR = 1 /* packet redirect to EGRESS*/ + TCA_EGRESS_MIRROR = 2 /* mirror packet to EGRESS */ + TCA_INGRESS_REDIR = 3 /* packet redirect to INGRESS*/ + TCA_INGRESS_MIRROR = 4 /* mirror packet to INGRESS */ +) + +const ( + TC_ACT_UNSPEC = int32(-1) + TC_ACT_OK = 0 + TC_ACT_RECLASSIFY = 1 + TC_ACT_SHOT = 2 + TC_ACT_PIPE = 3 + TC_ACT_STOLEN = 4 + TC_ACT_QUEUED = 5 + TC_ACT_REPEAT = 6 + TC_ACT_JUMP = 0x10000000 +) + +// #define tc_gen \ +// __u32 index; \ +// __u32 capab; \ +// int action; \ +// int refcnt; \ +// int bindcnt +// struct tc_mirred { +// tc_gen; +// int eaction; /* one of IN/EGRESS_MIRROR/REDIR */ +// __u32 ifindex; /* ifindex of egress port */ +// }; + +type TcMirred struct { + Index uint32 + Capab uint32 + Action int32 + Refcnt int32 + Bindcnt int32 + Eaction int32 + Ifindex uint32 +} + +func (msg *TcMirred) Len() int { + return SizeofTcMirred +} + +func DeserializeTcMirred(b []byte) *TcMirred { + return (*TcMirred)(unsafe.Pointer(&b[0:SizeofTcMirred][0])) +} + +func (x *TcMirred) Serialize() []byte { + return (*(*[SizeofTcMirred]byte)(unsafe.Pointer(x)))[:] +} diff --git a/vendor/github.com/vishvananda/netlink/nl/xfrm_linux.go b/vendor/github.com/vishvananda/netlink/nl/xfrm_linux.go new file mode 100644 index 000000000..d24637d27 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/nl/xfrm_linux.go @@ -0,0 +1,258 @@ +package nl + +import ( + "bytes" + "net" + "unsafe" +) + +// Infinity for packet and byte counts +const ( + XFRM_INF = ^uint64(0) +) + +// Message Types +const ( + XFRM_MSG_BASE = 0x10 + XFRM_MSG_NEWSA = 0x10 + XFRM_MSG_DELSA = 0x11 + XFRM_MSG_GETSA = 0x12 + XFRM_MSG_NEWPOLICY = 0x13 + XFRM_MSG_DELPOLICY = 0x14 + XFRM_MSG_GETPOLICY = 0x15 + XFRM_MSG_ALLOCSPI = 0x16 + XFRM_MSG_ACQUIRE = 0x17 + XFRM_MSG_EXPIRE = 0x18 + XFRM_MSG_UPDPOLICY = 0x19 + XFRM_MSG_UPDSA = 0x1a + XFRM_MSG_POLEXPIRE = 0x1b + XFRM_MSG_FLUSHSA = 0x1c + XFRM_MSG_FLUSHPOLICY = 0x1d + XFRM_MSG_NEWAE = 0x1e + XFRM_MSG_GETAE = 0x1f + XFRM_MSG_REPORT = 0x20 + XFRM_MSG_MIGRATE = 0x21 + XFRM_MSG_NEWSADINFO = 0x22 + XFRM_MSG_GETSADINFO = 0x23 + XFRM_MSG_NEWSPDINFO = 0x24 + XFRM_MSG_GETSPDINFO = 0x25 + XFRM_MSG_MAPPING = 0x26 + XFRM_MSG_MAX = 0x26 + XFRM_NR_MSGTYPES = 0x17 +) + +// Attribute types +const ( + /* Netlink message attributes. */ + XFRMA_UNSPEC = 0x00 + XFRMA_ALG_AUTH = 0x01 /* struct xfrm_algo */ + XFRMA_ALG_CRYPT = 0x02 /* struct xfrm_algo */ + XFRMA_ALG_COMP = 0x03 /* struct xfrm_algo */ + XFRMA_ENCAP = 0x04 /* struct xfrm_algo + struct xfrm_encap_tmpl */ + XFRMA_TMPL = 0x05 /* 1 or more struct xfrm_user_tmpl */ + XFRMA_SA = 0x06 /* struct xfrm_usersa_info */ + XFRMA_POLICY = 0x07 /* struct xfrm_userpolicy_info */ + XFRMA_SEC_CTX = 0x08 /* struct xfrm_sec_ctx */ + XFRMA_LTIME_VAL = 0x09 + XFRMA_REPLAY_VAL = 0x0a + XFRMA_REPLAY_THRESH = 0x0b + XFRMA_ETIMER_THRESH = 0x0c + XFRMA_SRCADDR = 0x0d /* xfrm_address_t */ + XFRMA_COADDR = 0x0e /* xfrm_address_t */ + XFRMA_LASTUSED = 0x0f /* unsigned long */ + XFRMA_POLICY_TYPE = 0x10 /* struct xfrm_userpolicy_type */ + XFRMA_MIGRATE = 0x11 + XFRMA_ALG_AEAD = 0x12 /* struct xfrm_algo_aead */ + XFRMA_KMADDRESS = 0x13 /* struct xfrm_user_kmaddress */ + XFRMA_ALG_AUTH_TRUNC = 0x14 /* struct xfrm_algo_auth */ + XFRMA_MARK = 0x15 /* struct xfrm_mark */ + XFRMA_TFCPAD = 0x16 /* __u32 */ + XFRMA_REPLAY_ESN_VAL = 0x17 /* struct xfrm_replay_esn */ + XFRMA_SA_EXTRA_FLAGS = 0x18 /* __u32 */ + XFRMA_MAX = 0x18 +) + +const ( + SizeofXfrmAddress = 0x10 + SizeofXfrmSelector = 0x38 + SizeofXfrmLifetimeCfg = 0x40 + SizeofXfrmLifetimeCur = 0x20 + SizeofXfrmId = 0x18 +) + +// typedef union { +// __be32 a4; +// __be32 a6[4]; +// } xfrm_address_t; + +type XfrmAddress [SizeofXfrmAddress]byte + +func (x *XfrmAddress) ToIP() net.IP { + var empty = [12]byte{} + ip := make(net.IP, net.IPv6len) + if bytes.Equal(x[4:16], empty[:]) { + ip[10] = 0xff + ip[11] = 0xff + copy(ip[12:16], x[0:4]) + } else { + copy(ip[:], x[:]) + } + return ip +} + +func (x *XfrmAddress) ToIPNet(prefixlen uint8) *net.IPNet { + ip := x.ToIP() + if GetIPFamily(ip) == FAMILY_V4 { + return &net.IPNet{IP: ip, Mask: net.CIDRMask(int(prefixlen), 32)} + } + return &net.IPNet{IP: ip, Mask: net.CIDRMask(int(prefixlen), 128)} +} + +func (x *XfrmAddress) FromIP(ip net.IP) { + var empty = [16]byte{} + if len(ip) < net.IPv4len { + copy(x[4:16], empty[:]) + } else if GetIPFamily(ip) == FAMILY_V4 { + copy(x[0:4], ip.To4()[0:4]) + copy(x[4:16], empty[:12]) + } else { + copy(x[0:16], ip.To16()[0:16]) + } +} + +func DeserializeXfrmAddress(b []byte) *XfrmAddress { + return (*XfrmAddress)(unsafe.Pointer(&b[0:SizeofXfrmAddress][0])) +} + +func (x *XfrmAddress) Serialize() []byte { + return (*(*[SizeofXfrmAddress]byte)(unsafe.Pointer(x)))[:] +} + +// struct xfrm_selector { +// xfrm_address_t daddr; +// xfrm_address_t saddr; +// __be16 dport; +// __be16 dport_mask; +// __be16 sport; +// __be16 sport_mask; +// __u16 family; +// __u8 prefixlen_d; +// __u8 prefixlen_s; +// __u8 proto; +// int ifindex; +// __kernel_uid32_t user; +// }; + +type XfrmSelector struct { + Daddr XfrmAddress + Saddr XfrmAddress + Dport uint16 // big endian + DportMask uint16 // big endian + Sport uint16 // big endian + SportMask uint16 // big endian + Family uint16 + PrefixlenD uint8 + PrefixlenS uint8 + Proto uint8 + Pad [3]byte + Ifindex int32 + User uint32 +} + +func (msg *XfrmSelector) Len() int { + return SizeofXfrmSelector +} + +func DeserializeXfrmSelector(b []byte) *XfrmSelector { + return (*XfrmSelector)(unsafe.Pointer(&b[0:SizeofXfrmSelector][0])) +} + +func (msg *XfrmSelector) Serialize() []byte { + return (*(*[SizeofXfrmSelector]byte)(unsafe.Pointer(msg)))[:] +} + +// struct xfrm_lifetime_cfg { +// __u64 soft_byte_limit; +// __u64 hard_byte_limit; +// __u64 soft_packet_limit; +// __u64 hard_packet_limit; +// __u64 soft_add_expires_seconds; +// __u64 hard_add_expires_seconds; +// __u64 soft_use_expires_seconds; +// __u64 hard_use_expires_seconds; +// }; +// + +type XfrmLifetimeCfg struct { + SoftByteLimit uint64 + HardByteLimit uint64 + SoftPacketLimit uint64 + HardPacketLimit uint64 + SoftAddExpiresSeconds uint64 + HardAddExpiresSeconds uint64 + SoftUseExpiresSeconds uint64 + HardUseExpiresSeconds uint64 +} + +func (msg *XfrmLifetimeCfg) Len() int { + return SizeofXfrmLifetimeCfg +} + +func DeserializeXfrmLifetimeCfg(b []byte) *XfrmLifetimeCfg { + return (*XfrmLifetimeCfg)(unsafe.Pointer(&b[0:SizeofXfrmLifetimeCfg][0])) +} + +func (msg *XfrmLifetimeCfg) Serialize() []byte { + return (*(*[SizeofXfrmLifetimeCfg]byte)(unsafe.Pointer(msg)))[:] +} + +// struct xfrm_lifetime_cur { +// __u64 bytes; +// __u64 packets; +// __u64 add_time; +// __u64 use_time; +// }; + +type XfrmLifetimeCur struct { + Bytes uint64 + Packets uint64 + AddTime uint64 + UseTime uint64 +} + +func (msg *XfrmLifetimeCur) Len() int { + return SizeofXfrmLifetimeCur +} + +func DeserializeXfrmLifetimeCur(b []byte) *XfrmLifetimeCur { + return (*XfrmLifetimeCur)(unsafe.Pointer(&b[0:SizeofXfrmLifetimeCur][0])) +} + +func (msg *XfrmLifetimeCur) Serialize() []byte { + return (*(*[SizeofXfrmLifetimeCur]byte)(unsafe.Pointer(msg)))[:] +} + +// struct xfrm_id { +// xfrm_address_t daddr; +// __be32 spi; +// __u8 proto; +// }; + +type XfrmId struct { + Daddr XfrmAddress + Spi uint32 // big endian + Proto uint8 + Pad [3]byte +} + +func (msg *XfrmId) Len() int { + return SizeofXfrmId +} + +func DeserializeXfrmId(b []byte) *XfrmId { + return (*XfrmId)(unsafe.Pointer(&b[0:SizeofXfrmId][0])) +} + +func (msg *XfrmId) Serialize() []byte { + return (*(*[SizeofXfrmId]byte)(unsafe.Pointer(msg)))[:] +} diff --git a/vendor/github.com/vishvananda/netlink/nl/xfrm_policy_linux.go b/vendor/github.com/vishvananda/netlink/nl/xfrm_policy_linux.go new file mode 100644 index 000000000..66f7e03d2 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/nl/xfrm_policy_linux.go @@ -0,0 +1,119 @@ +package nl + +import ( + "unsafe" +) + +const ( + SizeofXfrmUserpolicyId = 0x40 + SizeofXfrmUserpolicyInfo = 0xa8 + SizeofXfrmUserTmpl = 0x40 +) + +// struct xfrm_userpolicy_id { +// struct xfrm_selector sel; +// __u32 index; +// __u8 dir; +// }; +// + +type XfrmUserpolicyId struct { + Sel XfrmSelector + Index uint32 + Dir uint8 + Pad [3]byte +} + +func (msg *XfrmUserpolicyId) Len() int { + return SizeofXfrmUserpolicyId +} + +func DeserializeXfrmUserpolicyId(b []byte) *XfrmUserpolicyId { + return (*XfrmUserpolicyId)(unsafe.Pointer(&b[0:SizeofXfrmUserpolicyId][0])) +} + +func (msg *XfrmUserpolicyId) Serialize() []byte { + return (*(*[SizeofXfrmUserpolicyId]byte)(unsafe.Pointer(msg)))[:] +} + +// struct xfrm_userpolicy_info { +// struct xfrm_selector sel; +// struct xfrm_lifetime_cfg lft; +// struct xfrm_lifetime_cur curlft; +// __u32 priority; +// __u32 index; +// __u8 dir; +// __u8 action; +// #define XFRM_POLICY_ALLOW 0 +// #define XFRM_POLICY_BLOCK 1 +// __u8 flags; +// #define XFRM_POLICY_LOCALOK 1 /* Allow user to override global policy */ +// /* Automatically expand selector to include matching ICMP payloads. */ +// #define XFRM_POLICY_ICMP 2 +// __u8 share; +// }; + +type XfrmUserpolicyInfo struct { + Sel XfrmSelector + Lft XfrmLifetimeCfg + Curlft XfrmLifetimeCur + Priority uint32 + Index uint32 + Dir uint8 + Action uint8 + Flags uint8 + Share uint8 + Pad [4]byte +} + +func (msg *XfrmUserpolicyInfo) Len() int { + return SizeofXfrmUserpolicyInfo +} + +func DeserializeXfrmUserpolicyInfo(b []byte) *XfrmUserpolicyInfo { + return (*XfrmUserpolicyInfo)(unsafe.Pointer(&b[0:SizeofXfrmUserpolicyInfo][0])) +} + +func (msg *XfrmUserpolicyInfo) Serialize() []byte { + return (*(*[SizeofXfrmUserpolicyInfo]byte)(unsafe.Pointer(msg)))[:] +} + +// struct xfrm_user_tmpl { +// struct xfrm_id id; +// __u16 family; +// xfrm_address_t saddr; +// __u32 reqid; +// __u8 mode; +// __u8 share; +// __u8 optional; +// __u32 aalgos; +// __u32 ealgos; +// __u32 calgos; +// } + +type XfrmUserTmpl struct { + XfrmId XfrmId + Family uint16 + Pad1 [2]byte + Saddr XfrmAddress + Reqid uint32 + Mode uint8 + Share uint8 + Optional uint8 + Pad2 byte + Aalgos uint32 + Ealgos uint32 + Calgos uint32 +} + +func (msg *XfrmUserTmpl) Len() int { + return SizeofXfrmUserTmpl +} + +func DeserializeXfrmUserTmpl(b []byte) *XfrmUserTmpl { + return (*XfrmUserTmpl)(unsafe.Pointer(&b[0:SizeofXfrmUserTmpl][0])) +} + +func (msg *XfrmUserTmpl) Serialize() []byte { + return (*(*[SizeofXfrmUserTmpl]byte)(unsafe.Pointer(msg)))[:] +} diff --git a/vendor/github.com/vishvananda/netlink/nl/xfrm_state_linux.go b/vendor/github.com/vishvananda/netlink/nl/xfrm_state_linux.go new file mode 100644 index 000000000..4876ce458 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/nl/xfrm_state_linux.go @@ -0,0 +1,221 @@ +package nl + +import ( + "unsafe" +) + +const ( + SizeofXfrmUsersaId = 0x18 + SizeofXfrmStats = 0x0c + SizeofXfrmUsersaInfo = 0xe0 + SizeofXfrmAlgo = 0x44 + SizeofXfrmAlgoAuth = 0x48 + SizeofXfrmEncapTmpl = 0x18 +) + +// struct xfrm_usersa_id { +// xfrm_address_t daddr; +// __be32 spi; +// __u16 family; +// __u8 proto; +// }; + +type XfrmUsersaId struct { + Daddr XfrmAddress + Spi uint32 // big endian + Family uint16 + Proto uint8 + Pad byte +} + +func (msg *XfrmUsersaId) Len() int { + return SizeofXfrmUsersaId +} + +func DeserializeXfrmUsersaId(b []byte) *XfrmUsersaId { + return (*XfrmUsersaId)(unsafe.Pointer(&b[0:SizeofXfrmUsersaId][0])) +} + +func (msg *XfrmUsersaId) Serialize() []byte { + return (*(*[SizeofXfrmUsersaId]byte)(unsafe.Pointer(msg)))[:] +} + +// struct xfrm_stats { +// __u32 replay_window; +// __u32 replay; +// __u32 integrity_failed; +// }; + +type XfrmStats struct { + ReplayWindow uint32 + Replay uint32 + IntegrityFailed uint32 +} + +func (msg *XfrmStats) Len() int { + return SizeofXfrmStats +} + +func DeserializeXfrmStats(b []byte) *XfrmStats { + return (*XfrmStats)(unsafe.Pointer(&b[0:SizeofXfrmStats][0])) +} + +func (msg *XfrmStats) Serialize() []byte { + return (*(*[SizeofXfrmStats]byte)(unsafe.Pointer(msg)))[:] +} + +// struct xfrm_usersa_info { +// struct xfrm_selector sel; +// struct xfrm_id id; +// xfrm_address_t saddr; +// struct xfrm_lifetime_cfg lft; +// struct xfrm_lifetime_cur curlft; +// struct xfrm_stats stats; +// __u32 seq; +// __u32 reqid; +// __u16 family; +// __u8 mode; /* XFRM_MODE_xxx */ +// __u8 replay_window; +// __u8 flags; +// #define XFRM_STATE_NOECN 1 +// #define XFRM_STATE_DECAP_DSCP 2 +// #define XFRM_STATE_NOPMTUDISC 4 +// #define XFRM_STATE_WILDRECV 8 +// #define XFRM_STATE_ICMP 16 +// #define XFRM_STATE_AF_UNSPEC 32 +// #define XFRM_STATE_ALIGN4 64 +// #define XFRM_STATE_ESN 128 +// }; +// +// #define XFRM_SA_XFLAG_DONT_ENCAP_DSCP 1 +// + +type XfrmUsersaInfo struct { + Sel XfrmSelector + Id XfrmId + Saddr XfrmAddress + Lft XfrmLifetimeCfg + Curlft XfrmLifetimeCur + Stats XfrmStats + Seq uint32 + Reqid uint32 + Family uint16 + Mode uint8 + ReplayWindow uint8 + Flags uint8 + Pad [7]byte +} + +func (msg *XfrmUsersaInfo) Len() int { + return SizeofXfrmUsersaInfo +} + +func DeserializeXfrmUsersaInfo(b []byte) *XfrmUsersaInfo { + return (*XfrmUsersaInfo)(unsafe.Pointer(&b[0:SizeofXfrmUsersaInfo][0])) +} + +func (msg *XfrmUsersaInfo) Serialize() []byte { + return (*(*[SizeofXfrmUsersaInfo]byte)(unsafe.Pointer(msg)))[:] +} + +// struct xfrm_algo { +// char alg_name[64]; +// unsigned int alg_key_len; /* in bits */ +// char alg_key[0]; +// }; + +type XfrmAlgo struct { + AlgName [64]byte + AlgKeyLen uint32 + AlgKey []byte +} + +func (msg *XfrmAlgo) Len() int { + return SizeofXfrmAlgo + int(msg.AlgKeyLen/8) +} + +func DeserializeXfrmAlgo(b []byte) *XfrmAlgo { + ret := XfrmAlgo{} + copy(ret.AlgName[:], b[0:64]) + ret.AlgKeyLen = *(*uint32)(unsafe.Pointer(&b[64])) + ret.AlgKey = b[68:ret.Len()] + return &ret +} + +func (msg *XfrmAlgo) Serialize() []byte { + b := make([]byte, msg.Len()) + copy(b[0:64], msg.AlgName[:]) + copy(b[64:68], (*(*[4]byte)(unsafe.Pointer(&msg.AlgKeyLen)))[:]) + copy(b[68:msg.Len()], msg.AlgKey[:]) + return b +} + +// struct xfrm_algo_auth { +// char alg_name[64]; +// unsigned int alg_key_len; /* in bits */ +// unsigned int alg_trunc_len; /* in bits */ +// char alg_key[0]; +// }; + +type XfrmAlgoAuth struct { + AlgName [64]byte + AlgKeyLen uint32 + AlgTruncLen uint32 + AlgKey []byte +} + +func (msg *XfrmAlgoAuth) Len() int { + return SizeofXfrmAlgoAuth + int(msg.AlgKeyLen/8) +} + +func DeserializeXfrmAlgoAuth(b []byte) *XfrmAlgoAuth { + ret := XfrmAlgoAuth{} + copy(ret.AlgName[:], b[0:64]) + ret.AlgKeyLen = *(*uint32)(unsafe.Pointer(&b[64])) + ret.AlgTruncLen = *(*uint32)(unsafe.Pointer(&b[68])) + ret.AlgKey = b[72:ret.Len()] + return &ret +} + +func (msg *XfrmAlgoAuth) Serialize() []byte { + b := make([]byte, msg.Len()) + copy(b[0:64], msg.AlgName[:]) + copy(b[64:68], (*(*[4]byte)(unsafe.Pointer(&msg.AlgKeyLen)))[:]) + copy(b[68:72], (*(*[4]byte)(unsafe.Pointer(&msg.AlgTruncLen)))[:]) + copy(b[72:msg.Len()], msg.AlgKey[:]) + return b +} + +// struct xfrm_algo_aead { +// char alg_name[64]; +// unsigned int alg_key_len; /* in bits */ +// unsigned int alg_icv_len; /* in bits */ +// char alg_key[0]; +// } + +// struct xfrm_encap_tmpl { +// __u16 encap_type; +// __be16 encap_sport; +// __be16 encap_dport; +// xfrm_address_t encap_oa; +// }; + +type XfrmEncapTmpl struct { + EncapType uint16 + EncapSport uint16 // big endian + EncapDport uint16 // big endian + Pad [2]byte + EncapOa XfrmAddress +} + +func (msg *XfrmEncapTmpl) Len() int { + return SizeofXfrmEncapTmpl +} + +func DeserializeXfrmEncapTmpl(b []byte) *XfrmEncapTmpl { + return (*XfrmEncapTmpl)(unsafe.Pointer(&b[0:SizeofXfrmEncapTmpl][0])) +} + +func (msg *XfrmEncapTmpl) Serialize() []byte { + return (*(*[SizeofXfrmEncapTmpl]byte)(unsafe.Pointer(msg)))[:] +} diff --git a/vendor/github.com/vishvananda/netlink/protinfo.go b/vendor/github.com/vishvananda/netlink/protinfo.go new file mode 100644 index 000000000..f39ab8f4e --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/protinfo.go @@ -0,0 +1,53 @@ +package netlink + +import ( + "strings" +) + +// Protinfo represents bridge flags from netlink. +type Protinfo struct { + Hairpin bool + Guard bool + FastLeave bool + RootBlock bool + Learning bool + Flood bool +} + +// String returns a list of enabled flags +func (prot *Protinfo) String() string { + var boolStrings []string + if prot.Hairpin { + boolStrings = append(boolStrings, "Hairpin") + } + if prot.Guard { + boolStrings = append(boolStrings, "Guard") + } + if prot.FastLeave { + boolStrings = append(boolStrings, "FastLeave") + } + if prot.RootBlock { + boolStrings = append(boolStrings, "RootBlock") + } + if prot.Learning { + boolStrings = append(boolStrings, "Learning") + } + if prot.Flood { + boolStrings = append(boolStrings, "Flood") + } + return strings.Join(boolStrings, " ") +} + +func boolToByte(x bool) []byte { + if x { + return []byte{1} + } + return []byte{0} +} + +func byteToBool(x byte) bool { + if uint8(x) != 0 { + return true + } + return false +} diff --git a/vendor/github.com/vishvananda/netlink/protinfo_linux.go b/vendor/github.com/vishvananda/netlink/protinfo_linux.go new file mode 100644 index 000000000..7181eba10 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/protinfo_linux.go @@ -0,0 +1,60 @@ +package netlink + +import ( + "fmt" + "syscall" + + "github.com/vishvananda/netlink/nl" +) + +func LinkGetProtinfo(link Link) (Protinfo, error) { + base := link.Attrs() + ensureIndex(base) + var pi Protinfo + req := nl.NewNetlinkRequest(syscall.RTM_GETLINK, syscall.NLM_F_DUMP) + msg := nl.NewIfInfomsg(syscall.AF_BRIDGE) + req.AddData(msg) + msgs, err := req.Execute(syscall.NETLINK_ROUTE, 0) + if err != nil { + return pi, err + } + + for _, m := range msgs { + ans := nl.DeserializeIfInfomsg(m) + if int(ans.Index) != base.Index { + continue + } + attrs, err := nl.ParseRouteAttr(m[ans.Len():]) + if err != nil { + return pi, err + } + for _, attr := range attrs { + if attr.Attr.Type != syscall.IFLA_PROTINFO|syscall.NLA_F_NESTED { + continue + } + infos, err := nl.ParseRouteAttr(attr.Value) + if err != nil { + return pi, err + } + var pi Protinfo + for _, info := range infos { + switch info.Attr.Type { + case nl.IFLA_BRPORT_MODE: + pi.Hairpin = byteToBool(info.Value[0]) + case nl.IFLA_BRPORT_GUARD: + pi.Guard = byteToBool(info.Value[0]) + case nl.IFLA_BRPORT_FAST_LEAVE: + pi.FastLeave = byteToBool(info.Value[0]) + case nl.IFLA_BRPORT_PROTECT: + pi.RootBlock = byteToBool(info.Value[0]) + case nl.IFLA_BRPORT_LEARNING: + pi.Learning = byteToBool(info.Value[0]) + case nl.IFLA_BRPORT_UNICAST_FLOOD: + pi.Flood = byteToBool(info.Value[0]) + } + } + return pi, nil + } + } + return pi, fmt.Errorf("Device with index %d not found", base.Index) +} diff --git a/vendor/github.com/vishvananda/netlink/qdisc.go b/vendor/github.com/vishvananda/netlink/qdisc.go new file mode 100644 index 000000000..8e3d020fd --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/qdisc.go @@ -0,0 +1,138 @@ +package netlink + +import ( + "fmt" +) + +const ( + HANDLE_NONE = 0 + HANDLE_INGRESS = 0xFFFFFFF1 + HANDLE_ROOT = 0xFFFFFFFF + PRIORITY_MAP_LEN = 16 +) + +type Qdisc interface { + Attrs() *QdiscAttrs + Type() string +} + +// Qdisc represents a netlink qdisc. A qdisc is associated with a link, +// has a handle, a parent and a refcnt. The root qdisc of a device should +// have parent == HANDLE_ROOT. +type QdiscAttrs struct { + LinkIndex int + Handle uint32 + Parent uint32 + Refcnt uint32 // read only +} + +func (q QdiscAttrs) String() string { + return fmt.Sprintf("{LinkIndex: %d, Handle: %s, Parent: %s, Refcnt: %s}", q.LinkIndex, HandleStr(q.Handle), HandleStr(q.Parent), q.Refcnt) +} + +func MakeHandle(major, minor uint16) uint32 { + return (uint32(major) << 16) | uint32(minor) +} + +func MajorMinor(handle uint32) (uint16, uint16) { + return uint16((handle & 0xFFFF0000) >> 16), uint16(handle & 0x0000FFFFF) +} + +func HandleStr(handle uint32) string { + switch handle { + case HANDLE_NONE: + return "none" + case HANDLE_INGRESS: + return "ingress" + case HANDLE_ROOT: + return "root" + default: + major, minor := MajorMinor(handle) + return fmt.Sprintf("%x:%x", major, minor) + } +} + +// PfifoFast is the default qdisc created by the kernel if one has not +// been defined for the interface +type PfifoFast struct { + QdiscAttrs + Bands uint8 + PriorityMap [PRIORITY_MAP_LEN]uint8 +} + +func (qdisc *PfifoFast) Attrs() *QdiscAttrs { + return &qdisc.QdiscAttrs +} + +func (qdisc *PfifoFast) Type() string { + return "pfifo_fast" +} + +// Prio is a basic qdisc that works just like PfifoFast +type Prio struct { + QdiscAttrs + Bands uint8 + PriorityMap [PRIORITY_MAP_LEN]uint8 +} + +func NewPrio(attrs QdiscAttrs) *Prio { + return &Prio{ + QdiscAttrs: attrs, + Bands: 3, + PriorityMap: [PRIORITY_MAP_LEN]uint8{1, 2, 2, 2, 1, 2, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1}, + } +} + +func (qdisc *Prio) Attrs() *QdiscAttrs { + return &qdisc.QdiscAttrs +} + +func (qdisc *Prio) Type() string { + return "prio" +} + +// Tbf is a classful qdisc that rate limits based on tokens +type Tbf struct { + QdiscAttrs + // TODO: handle 64bit rate properly + Rate uint64 + Limit uint32 + Buffer uint32 + // TODO: handle other settings +} + +func (qdisc *Tbf) Attrs() *QdiscAttrs { + return &qdisc.QdiscAttrs +} + +func (qdisc *Tbf) Type() string { + return "tbf" +} + +// Ingress is a qdisc for adding ingress filters +type Ingress struct { + QdiscAttrs +} + +func (qdisc *Ingress) Attrs() *QdiscAttrs { + return &qdisc.QdiscAttrs +} + +func (qdisc *Ingress) Type() string { + return "ingress" +} + +// GenericQdisc qdiscs represent types that are not currently understood +// by this netlink library. +type GenericQdisc struct { + QdiscAttrs + QdiscType string +} + +func (qdisc *GenericQdisc) Attrs() *QdiscAttrs { + return &qdisc.QdiscAttrs +} + +func (qdisc *GenericQdisc) Type() string { + return qdisc.QdiscType +} diff --git a/vendor/github.com/vishvananda/netlink/qdisc_linux.go b/vendor/github.com/vishvananda/netlink/qdisc_linux.go new file mode 100644 index 000000000..2531c9dd1 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/qdisc_linux.go @@ -0,0 +1,263 @@ +package netlink + +import ( + "fmt" + "io/ioutil" + "strconv" + "strings" + "syscall" + + "github.com/vishvananda/netlink/nl" +) + +// QdiscDel will delete a qdisc from the system. +// Equivalent to: `tc qdisc del $qdisc` +func QdiscDel(qdisc Qdisc) error { + req := nl.NewNetlinkRequest(syscall.RTM_DELQDISC, syscall.NLM_F_ACK) + base := qdisc.Attrs() + msg := &nl.TcMsg{ + Family: nl.FAMILY_ALL, + Ifindex: int32(base.LinkIndex), + Handle: base.Handle, + Parent: base.Parent, + } + req.AddData(msg) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// QdiscAdd will add a qdisc to the system. +// Equivalent to: `tc qdisc add $qdisc` +func QdiscAdd(qdisc Qdisc) error { + req := nl.NewNetlinkRequest(syscall.RTM_NEWQDISC, syscall.NLM_F_CREATE|syscall.NLM_F_EXCL|syscall.NLM_F_ACK) + base := qdisc.Attrs() + msg := &nl.TcMsg{ + Family: nl.FAMILY_ALL, + Ifindex: int32(base.LinkIndex), + Handle: base.Handle, + Parent: base.Parent, + } + req.AddData(msg) + req.AddData(nl.NewRtAttr(nl.TCA_KIND, nl.ZeroTerminated(qdisc.Type()))) + + options := nl.NewRtAttr(nl.TCA_OPTIONS, nil) + if prio, ok := qdisc.(*Prio); ok { + tcmap := nl.TcPrioMap{ + Bands: int32(prio.Bands), + Priomap: prio.PriorityMap, + } + options = nl.NewRtAttr(nl.TCA_OPTIONS, tcmap.Serialize()) + } else if tbf, ok := qdisc.(*Tbf); ok { + opt := nl.TcTbfQopt{} + // TODO: handle rate > uint32 + opt.Rate.Rate = uint32(tbf.Rate) + opt.Limit = tbf.Limit + opt.Buffer = tbf.Buffer + nl.NewRtAttrChild(options, nl.TCA_TBF_PARMS, opt.Serialize()) + } else if _, ok := qdisc.(*Ingress); ok { + // ingress filters must use the proper handle + if msg.Parent != HANDLE_INGRESS { + return fmt.Errorf("Ingress filters must set Parent to HANDLE_INGRESS") + } + } + req.AddData(options) + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// QdiscList gets a list of qdiscs in the system. +// Equivalent to: `tc qdisc show`. +// The list can be filtered by link. +func QdiscList(link Link) ([]Qdisc, error) { + req := nl.NewNetlinkRequest(syscall.RTM_GETQDISC, syscall.NLM_F_DUMP) + index := int32(0) + if link != nil { + base := link.Attrs() + ensureIndex(base) + index = int32(base.Index) + } + msg := &nl.TcMsg{ + Family: nl.FAMILY_ALL, + Ifindex: index, + } + req.AddData(msg) + + msgs, err := req.Execute(syscall.NETLINK_ROUTE, syscall.RTM_NEWQDISC) + if err != nil { + return nil, err + } + + var res []Qdisc + for _, m := range msgs { + msg := nl.DeserializeTcMsg(m) + + attrs, err := nl.ParseRouteAttr(m[msg.Len():]) + if err != nil { + return nil, err + } + + // skip qdiscs from other interfaces + if link != nil && msg.Ifindex != index { + continue + } + + base := QdiscAttrs{ + LinkIndex: int(msg.Ifindex), + Handle: msg.Handle, + Parent: msg.Parent, + Refcnt: msg.Info, + } + var qdisc Qdisc + qdiscType := "" + for _, attr := range attrs { + switch attr.Attr.Type { + case nl.TCA_KIND: + qdiscType = string(attr.Value[:len(attr.Value)-1]) + switch qdiscType { + case "pfifo_fast": + qdisc = &PfifoFast{} + case "prio": + qdisc = &Prio{} + case "tbf": + qdisc = &Tbf{} + case "ingress": + qdisc = &Ingress{} + default: + qdisc = &GenericQdisc{QdiscType: qdiscType} + } + case nl.TCA_OPTIONS: + switch qdiscType { + case "pfifo_fast": + // pfifo returns TcPrioMap directly without wrapping it in rtattr + if err := parsePfifoFastData(qdisc, attr.Value); err != nil { + return nil, err + } + case "prio": + // prio returns TcPrioMap directly without wrapping it in rtattr + if err := parsePrioData(qdisc, attr.Value); err != nil { + return nil, err + } + case "tbf": + data, err := nl.ParseRouteAttr(attr.Value) + if err != nil { + return nil, err + } + if err := parseTbfData(qdisc, data); err != nil { + return nil, err + } + // no options for ingress + } + } + } + *qdisc.Attrs() = base + res = append(res, qdisc) + } + + return res, nil +} + +func parsePfifoFastData(qdisc Qdisc, value []byte) error { + pfifo := qdisc.(*PfifoFast) + tcmap := nl.DeserializeTcPrioMap(value) + pfifo.PriorityMap = tcmap.Priomap + pfifo.Bands = uint8(tcmap.Bands) + return nil +} + +func parsePrioData(qdisc Qdisc, value []byte) error { + prio := qdisc.(*Prio) + tcmap := nl.DeserializeTcPrioMap(value) + prio.PriorityMap = tcmap.Priomap + prio.Bands = uint8(tcmap.Bands) + return nil +} + +func parseTbfData(qdisc Qdisc, data []syscall.NetlinkRouteAttr) error { + native = nl.NativeEndian() + tbf := qdisc.(*Tbf) + for _, datum := range data { + switch datum.Attr.Type { + case nl.TCA_TBF_PARMS: + opt := nl.DeserializeTcTbfQopt(datum.Value) + tbf.Rate = uint64(opt.Rate.Rate) + tbf.Limit = opt.Limit + tbf.Buffer = opt.Buffer + case nl.TCA_TBF_RATE64: + tbf.Rate = native.Uint64(datum.Value[0:4]) + } + } + return nil +} + +const ( + TIME_UNITS_PER_SEC = 1000000 +) + +var ( + tickInUsec float64 = 0.0 + clockFactor float64 = 0.0 +) + +func initClock() { + data, err := ioutil.ReadFile("/proc/net/psched") + if err != nil { + return + } + parts := strings.Split(strings.TrimSpace(string(data)), " ") + if len(parts) < 3 { + return + } + var vals [3]uint64 + for i := range vals { + val, err := strconv.ParseUint(parts[i], 16, 32) + if err != nil { + return + } + vals[i] = val + } + // compatibility + if vals[2] == 1000000000 { + vals[0] = vals[1] + } + clockFactor = float64(vals[2]) / TIME_UNITS_PER_SEC + tickInUsec = float64(vals[0]) / float64(vals[1]) * clockFactor +} + +func TickInUsec() float64 { + if tickInUsec == 0.0 { + initClock() + } + return tickInUsec +} + +func ClockFactor() float64 { + if clockFactor == 0.0 { + initClock() + } + return clockFactor +} + +func time2Tick(time uint32) uint32 { + return uint32(float64(time) * TickInUsec()) +} + +func tick2Time(tick uint32) uint32 { + return uint32(float64(tick) / TickInUsec()) +} + +func time2Ktime(time uint32) uint32 { + return uint32(float64(time) * ClockFactor()) +} + +func ktime2Time(ktime uint32) uint32 { + return uint32(float64(ktime) / ClockFactor()) +} + +func burst(rate uint64, buffer uint32) uint32 { + return uint32(float64(rate) * float64(tick2Time(buffer)) / TIME_UNITS_PER_SEC) +} + +func latency(rate uint64, limit, buffer uint32) float64 { + return TIME_UNITS_PER_SEC*(float64(limit)/float64(rate)) - float64(tick2Time(buffer)) +} diff --git a/vendor/github.com/vishvananda/netlink/route.go b/vendor/github.com/vishvananda/netlink/route.go new file mode 100644 index 000000000..6218546f8 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/route.go @@ -0,0 +1,35 @@ +package netlink + +import ( + "fmt" + "net" + "syscall" +) + +// Scope is an enum representing a route scope. +type Scope uint8 + +const ( + SCOPE_UNIVERSE Scope = syscall.RT_SCOPE_UNIVERSE + SCOPE_SITE Scope = syscall.RT_SCOPE_SITE + SCOPE_LINK Scope = syscall.RT_SCOPE_LINK + SCOPE_HOST Scope = syscall.RT_SCOPE_HOST + SCOPE_NOWHERE Scope = syscall.RT_SCOPE_NOWHERE +) + +// Route represents a netlink route. A route is associated with a link, +// has a destination network, an optional source ip, and optional +// gateway. Advanced route parameters and non-main routing tables are +// currently not supported. +type Route struct { + LinkIndex int + Scope Scope + Dst *net.IPNet + Src net.IP + Gw net.IP +} + +func (r Route) String() string { + return fmt.Sprintf("{Ifindex: %d Dst: %s Src: %s Gw: %s}", r.LinkIndex, r.Dst, + r.Src, r.Gw) +} diff --git a/vendor/github.com/vishvananda/netlink/route_linux.go b/vendor/github.com/vishvananda/netlink/route_linux.go new file mode 100644 index 000000000..9e76d4414 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/route_linux.go @@ -0,0 +1,225 @@ +package netlink + +import ( + "fmt" + "net" + "syscall" + + "github.com/vishvananda/netlink/nl" +) + +// RtAttr is shared so it is in netlink_linux.go + +// RouteAdd will add a route to the system. +// Equivalent to: `ip route add $route` +func RouteAdd(route *Route) error { + req := nl.NewNetlinkRequest(syscall.RTM_NEWROUTE, syscall.NLM_F_CREATE|syscall.NLM_F_EXCL|syscall.NLM_F_ACK) + return routeHandle(route, req, nl.NewRtMsg()) +} + +// RouteAdd will delete a route from the system. +// Equivalent to: `ip route del $route` +func RouteDel(route *Route) error { + req := nl.NewNetlinkRequest(syscall.RTM_DELROUTE, syscall.NLM_F_ACK) + return routeHandle(route, req, nl.NewRtDelMsg()) +} + +func routeHandle(route *Route, req *nl.NetlinkRequest, msg *nl.RtMsg) error { + if (route.Dst == nil || route.Dst.IP == nil) && route.Src == nil && route.Gw == nil { + return fmt.Errorf("one of Dst.IP, Src, or Gw must not be nil") + } + + msg.Scope = uint8(route.Scope) + family := -1 + var rtAttrs []*nl.RtAttr + + if route.Dst != nil && route.Dst.IP != nil { + dstLen, _ := route.Dst.Mask.Size() + msg.Dst_len = uint8(dstLen) + dstFamily := nl.GetIPFamily(route.Dst.IP) + family = dstFamily + var dstData []byte + if dstFamily == FAMILY_V4 { + dstData = route.Dst.IP.To4() + } else { + dstData = route.Dst.IP.To16() + } + rtAttrs = append(rtAttrs, nl.NewRtAttr(syscall.RTA_DST, dstData)) + } + + if route.Src != nil { + srcFamily := nl.GetIPFamily(route.Src) + if family != -1 && family != srcFamily { + return fmt.Errorf("source and destination ip are not the same IP family") + } + family = srcFamily + var srcData []byte + if srcFamily == FAMILY_V4 { + srcData = route.Src.To4() + } else { + srcData = route.Src.To16() + } + // The commonly used src ip for routes is actually PREFSRC + rtAttrs = append(rtAttrs, nl.NewRtAttr(syscall.RTA_PREFSRC, srcData)) + } + + if route.Gw != nil { + gwFamily := nl.GetIPFamily(route.Gw) + if family != -1 && family != gwFamily { + return fmt.Errorf("gateway, source, and destination ip are not the same IP family") + } + family = gwFamily + var gwData []byte + if gwFamily == FAMILY_V4 { + gwData = route.Gw.To4() + } else { + gwData = route.Gw.To16() + } + rtAttrs = append(rtAttrs, nl.NewRtAttr(syscall.RTA_GATEWAY, gwData)) + } + + msg.Family = uint8(family) + + req.AddData(msg) + for _, attr := range rtAttrs { + req.AddData(attr) + } + + var ( + b = make([]byte, 4) + native = nl.NativeEndian() + ) + native.PutUint32(b, uint32(route.LinkIndex)) + + req.AddData(nl.NewRtAttr(syscall.RTA_OIF, b)) + + _, err := req.Execute(syscall.NETLINK_ROUTE, 0) + return err +} + +// RouteList gets a list of routes in the system. +// Equivalent to: `ip route show`. +// The list can be filtered by link and ip family. +func RouteList(link Link, family int) ([]Route, error) { + req := nl.NewNetlinkRequest(syscall.RTM_GETROUTE, syscall.NLM_F_DUMP) + msg := nl.NewIfInfomsg(family) + req.AddData(msg) + + msgs, err := req.Execute(syscall.NETLINK_ROUTE, syscall.RTM_NEWROUTE) + if err != nil { + return nil, err + } + + index := 0 + if link != nil { + base := link.Attrs() + ensureIndex(base) + index = base.Index + } + + native := nl.NativeEndian() + var res []Route +MsgLoop: + for _, m := range msgs { + msg := nl.DeserializeRtMsg(m) + + if msg.Flags&syscall.RTM_F_CLONED != 0 { + // Ignore cloned routes + continue + } + + if msg.Table != syscall.RT_TABLE_MAIN { + // Ignore non-main tables + continue + } + + attrs, err := nl.ParseRouteAttr(m[msg.Len():]) + if err != nil { + return nil, err + } + + route := Route{Scope: Scope(msg.Scope)} + for _, attr := range attrs { + switch attr.Attr.Type { + case syscall.RTA_GATEWAY: + route.Gw = net.IP(attr.Value) + case syscall.RTA_PREFSRC: + route.Src = net.IP(attr.Value) + case syscall.RTA_DST: + route.Dst = &net.IPNet{ + IP: attr.Value, + Mask: net.CIDRMask(int(msg.Dst_len), 8*len(attr.Value)), + } + case syscall.RTA_OIF: + routeIndex := int(native.Uint32(attr.Value[0:4])) + if link != nil && routeIndex != index { + // Ignore routes from other interfaces + continue MsgLoop + } + route.LinkIndex = routeIndex + } + } + res = append(res, route) + } + + return res, nil +} + +// RouteGet gets a route to a specific destination from the host system. +// Equivalent to: 'ip route get'. +func RouteGet(destination net.IP) ([]Route, error) { + req := nl.NewNetlinkRequest(syscall.RTM_GETROUTE, syscall.NLM_F_REQUEST) + family := nl.GetIPFamily(destination) + var destinationData []byte + var bitlen uint8 + if family == FAMILY_V4 { + destinationData = destination.To4() + bitlen = 32 + } else { + destinationData = destination.To16() + bitlen = 128 + } + msg := &nl.RtMsg{} + msg.Family = uint8(family) + msg.Dst_len = bitlen + req.AddData(msg) + + rtaDst := nl.NewRtAttr(syscall.RTA_DST, destinationData) + req.AddData(rtaDst) + + msgs, err := req.Execute(syscall.NETLINK_ROUTE, syscall.RTM_NEWROUTE) + if err != nil { + return nil, err + } + + native := nl.NativeEndian() + var res []Route + for _, m := range msgs { + msg := nl.DeserializeRtMsg(m) + attrs, err := nl.ParseRouteAttr(m[msg.Len():]) + if err != nil { + return nil, err + } + + route := Route{} + for _, attr := range attrs { + switch attr.Attr.Type { + case syscall.RTA_GATEWAY: + route.Gw = net.IP(attr.Value) + case syscall.RTA_PREFSRC: + route.Src = net.IP(attr.Value) + case syscall.RTA_DST: + route.Dst = &net.IPNet{ + IP: attr.Value, + Mask: net.CIDRMask(int(msg.Dst_len), 8*len(attr.Value)), + } + case syscall.RTA_OIF: + routeIndex := int(native.Uint32(attr.Value[0:4])) + route.LinkIndex = routeIndex + } + } + res = append(res, route) + } + return res, nil + +} diff --git a/vendor/github.com/vishvananda/netlink/xfrm.go b/vendor/github.com/vishvananda/netlink/xfrm.go new file mode 100644 index 000000000..621ffb6c6 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/xfrm.go @@ -0,0 +1,64 @@ +package netlink + +import ( + "fmt" + "syscall" +) + +// Proto is an enum representing an ipsec protocol. +type Proto uint8 + +const ( + XFRM_PROTO_ROUTE2 Proto = syscall.IPPROTO_ROUTING + XFRM_PROTO_ESP Proto = syscall.IPPROTO_ESP + XFRM_PROTO_AH Proto = syscall.IPPROTO_AH + XFRM_PROTO_HAO Proto = syscall.IPPROTO_DSTOPTS + XFRM_PROTO_COMP Proto = syscall.IPPROTO_COMP + XFRM_PROTO_IPSEC_ANY Proto = syscall.IPPROTO_RAW +) + +func (p Proto) String() string { + switch p { + case XFRM_PROTO_ROUTE2: + return "route2" + case XFRM_PROTO_ESP: + return "esp" + case XFRM_PROTO_AH: + return "ah" + case XFRM_PROTO_HAO: + return "hao" + case XFRM_PROTO_COMP: + return "comp" + case XFRM_PROTO_IPSEC_ANY: + return "ipsec-any" + } + return fmt.Sprintf("%d", p) +} + +// Mode is an enum representing an ipsec transport. +type Mode uint8 + +const ( + XFRM_MODE_TRANSPORT Mode = iota + XFRM_MODE_TUNNEL + XFRM_MODE_ROUTEOPTIMIZATION + XFRM_MODE_IN_TRIGGER + XFRM_MODE_BEET + XFRM_MODE_MAX +) + +func (m Mode) String() string { + switch m { + case XFRM_MODE_TRANSPORT: + return "transport" + case XFRM_MODE_TUNNEL: + return "tunnel" + case XFRM_MODE_ROUTEOPTIMIZATION: + return "ro" + case XFRM_MODE_IN_TRIGGER: + return "in_trigger" + case XFRM_MODE_BEET: + return "beet" + } + return fmt.Sprintf("%d", m) +} diff --git a/vendor/github.com/vishvananda/netlink/xfrm_policy.go b/vendor/github.com/vishvananda/netlink/xfrm_policy.go new file mode 100644 index 000000000..d85c65d2d --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/xfrm_policy.go @@ -0,0 +1,59 @@ +package netlink + +import ( + "fmt" + "net" +) + +// Dir is an enum representing an ipsec template direction. +type Dir uint8 + +const ( + XFRM_DIR_IN Dir = iota + XFRM_DIR_OUT + XFRM_DIR_FWD + XFRM_SOCKET_IN + XFRM_SOCKET_OUT + XFRM_SOCKET_FWD +) + +func (d Dir) String() string { + switch d { + case XFRM_DIR_IN: + return "dir in" + case XFRM_DIR_OUT: + return "dir out" + case XFRM_DIR_FWD: + return "dir fwd" + case XFRM_SOCKET_IN: + return "socket in" + case XFRM_SOCKET_OUT: + return "socket out" + case XFRM_SOCKET_FWD: + return "socket fwd" + } + return fmt.Sprintf("socket %d", d-XFRM_SOCKET_IN) +} + +// XfrmPolicyTmpl encapsulates a rule for the base addresses of an ipsec +// policy. These rules are matched with XfrmState to determine encryption +// and authentication algorithms. +type XfrmPolicyTmpl struct { + Dst net.IP + Src net.IP + Proto Proto + Mode Mode + Reqid int +} + +// XfrmPolicy represents an ipsec policy. It represents the overlay network +// and has a list of XfrmPolicyTmpls representing the base addresses of +// the policy. +type XfrmPolicy struct { + Dst *net.IPNet + Src *net.IPNet + Dir Dir + Priority int + Index int + Tmpls []XfrmPolicyTmpl +} diff --git a/vendor/github.com/vishvananda/netlink/xfrm_policy_linux.go b/vendor/github.com/vishvananda/netlink/xfrm_policy_linux.go new file mode 100644 index 000000000..2daf6dc8b --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/xfrm_policy_linux.go @@ -0,0 +1,127 @@ +package netlink + +import ( + "syscall" + + "github.com/vishvananda/netlink/nl" +) + +func selFromPolicy(sel *nl.XfrmSelector, policy *XfrmPolicy) { + sel.Family = uint16(nl.GetIPFamily(policy.Dst.IP)) + sel.Daddr.FromIP(policy.Dst.IP) + sel.Saddr.FromIP(policy.Src.IP) + prefixlenD, _ := policy.Dst.Mask.Size() + sel.PrefixlenD = uint8(prefixlenD) + prefixlenS, _ := policy.Src.Mask.Size() + sel.PrefixlenS = uint8(prefixlenS) +} + +// XfrmPolicyAdd will add an xfrm policy to the system. +// Equivalent to: `ip xfrm policy add $policy` +func XfrmPolicyAdd(policy *XfrmPolicy) error { + req := nl.NewNetlinkRequest(nl.XFRM_MSG_NEWPOLICY, syscall.NLM_F_CREATE|syscall.NLM_F_EXCL|syscall.NLM_F_ACK) + + msg := &nl.XfrmUserpolicyInfo{} + selFromPolicy(&msg.Sel, policy) + msg.Priority = uint32(policy.Priority) + msg.Index = uint32(policy.Index) + msg.Dir = uint8(policy.Dir) + msg.Lft.SoftByteLimit = nl.XFRM_INF + msg.Lft.HardByteLimit = nl.XFRM_INF + msg.Lft.SoftPacketLimit = nl.XFRM_INF + msg.Lft.HardPacketLimit = nl.XFRM_INF + req.AddData(msg) + + tmplData := make([]byte, nl.SizeofXfrmUserTmpl*len(policy.Tmpls)) + for i, tmpl := range policy.Tmpls { + start := i * nl.SizeofXfrmUserTmpl + userTmpl := nl.DeserializeXfrmUserTmpl(tmplData[start : start+nl.SizeofXfrmUserTmpl]) + userTmpl.XfrmId.Daddr.FromIP(tmpl.Dst) + userTmpl.Saddr.FromIP(tmpl.Src) + userTmpl.XfrmId.Proto = uint8(tmpl.Proto) + userTmpl.Mode = uint8(tmpl.Mode) + userTmpl.Reqid = uint32(tmpl.Reqid) + userTmpl.Aalgos = ^uint32(0) + userTmpl.Ealgos = ^uint32(0) + userTmpl.Calgos = ^uint32(0) + } + if len(tmplData) > 0 { + tmpls := nl.NewRtAttr(nl.XFRMA_TMPL, tmplData) + req.AddData(tmpls) + } + + _, err := req.Execute(syscall.NETLINK_XFRM, 0) + return err +} + +// XfrmPolicyDel will delete an xfrm policy from the system. Note that +// the Tmpls are ignored when matching the policy to delete. +// Equivalent to: `ip xfrm policy del $policy` +func XfrmPolicyDel(policy *XfrmPolicy) error { + req := nl.NewNetlinkRequest(nl.XFRM_MSG_DELPOLICY, syscall.NLM_F_ACK) + + msg := &nl.XfrmUserpolicyId{} + selFromPolicy(&msg.Sel, policy) + msg.Index = uint32(policy.Index) + msg.Dir = uint8(policy.Dir) + req.AddData(msg) + + _, err := req.Execute(syscall.NETLINK_XFRM, 0) + return err +} + +// XfrmPolicyList gets a list of xfrm policies in the system. +// Equivalent to: `ip xfrm policy show`. +// The list can be filtered by ip family. +func XfrmPolicyList(family int) ([]XfrmPolicy, error) { + req := nl.NewNetlinkRequest(nl.XFRM_MSG_GETPOLICY, syscall.NLM_F_DUMP) + + msg := nl.NewIfInfomsg(family) + req.AddData(msg) + + msgs, err := req.Execute(syscall.NETLINK_XFRM, nl.XFRM_MSG_NEWPOLICY) + if err != nil { + return nil, err + } + + var res []XfrmPolicy + for _, m := range msgs { + msg := nl.DeserializeXfrmUserpolicyInfo(m) + + if family != FAMILY_ALL && family != int(msg.Sel.Family) { + continue + } + + var policy XfrmPolicy + + policy.Dst = msg.Sel.Daddr.ToIPNet(msg.Sel.PrefixlenD) + policy.Src = msg.Sel.Saddr.ToIPNet(msg.Sel.PrefixlenS) + policy.Priority = int(msg.Priority) + policy.Index = int(msg.Index) + policy.Dir = Dir(msg.Dir) + + attrs, err := nl.ParseRouteAttr(m[msg.Len():]) + if err != nil { + return nil, err + } + + for _, attr := range attrs { + switch attr.Attr.Type { + case nl.XFRMA_TMPL: + max := len(attr.Value) + for i := 0; i < max; i += nl.SizeofXfrmUserTmpl { + var resTmpl XfrmPolicyTmpl + tmpl := nl.DeserializeXfrmUserTmpl(attr.Value[i : i+nl.SizeofXfrmUserTmpl]) + resTmpl.Dst = tmpl.XfrmId.Daddr.ToIP() + resTmpl.Src = tmpl.Saddr.ToIP() + resTmpl.Proto = Proto(tmpl.XfrmId.Proto) + resTmpl.Mode = Mode(tmpl.Mode) + resTmpl.Reqid = int(tmpl.Reqid) + policy.Tmpls = append(policy.Tmpls, resTmpl) + } + } + } + res = append(res, policy) + } + return res, nil +} diff --git a/vendor/github.com/vishvananda/netlink/xfrm_state.go b/vendor/github.com/vishvananda/netlink/xfrm_state.go new file mode 100644 index 000000000..5b8f2df70 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/xfrm_state.go @@ -0,0 +1,53 @@ +package netlink + +import ( + "net" +) + +// XfrmStateAlgo represents the algorithm to use for the ipsec encryption. +type XfrmStateAlgo struct { + Name string + Key []byte + TruncateLen int // Auth only +} + +// EncapType is an enum representing an ipsec template direction. +type EncapType uint8 + +const ( + XFRM_ENCAP_ESPINUDP_NONIKE EncapType = iota + 1 + XFRM_ENCAP_ESPINUDP +) + +func (e EncapType) String() string { + switch e { + case XFRM_ENCAP_ESPINUDP_NONIKE: + return "espinudp-nonike" + case XFRM_ENCAP_ESPINUDP: + return "espinudp" + } + return "unknown" +} + +// XfrmEncap represents the encapsulation to use for the ipsec encryption. +type XfrmStateEncap struct { + Type EncapType + SrcPort int + DstPort int + OriginalAddress net.IP +} + +// XfrmState represents the state of an ipsec policy. It optionally +// contains an XfrmStateAlgo for encryption and one for authentication. +type XfrmState struct { + Dst net.IP + Src net.IP + Proto Proto + Mode Mode + Spi int + Reqid int + ReplayWindow int + Auth *XfrmStateAlgo + Crypt *XfrmStateAlgo + Encap *XfrmStateEncap +} diff --git a/vendor/github.com/vishvananda/netlink/xfrm_state_linux.go b/vendor/github.com/vishvananda/netlink/xfrm_state_linux.go new file mode 100644 index 000000000..5f44ec852 --- /dev/null +++ b/vendor/github.com/vishvananda/netlink/xfrm_state_linux.go @@ -0,0 +1,181 @@ +package netlink + +import ( + "fmt" + "syscall" + + "github.com/vishvananda/netlink/nl" +) + +func writeStateAlgo(a *XfrmStateAlgo) []byte { + algo := nl.XfrmAlgo{ + AlgKeyLen: uint32(len(a.Key) * 8), + AlgKey: a.Key, + } + end := len(a.Name) + if end > 64 { + end = 64 + } + copy(algo.AlgName[:end], a.Name) + return algo.Serialize() +} + +func writeStateAlgoAuth(a *XfrmStateAlgo) []byte { + algo := nl.XfrmAlgoAuth{ + AlgKeyLen: uint32(len(a.Key) * 8), + AlgTruncLen: uint32(a.TruncateLen), + AlgKey: a.Key, + } + end := len(a.Name) + if end > 64 { + end = 64 + } + copy(algo.AlgName[:end], a.Name) + return algo.Serialize() +} + +// XfrmStateAdd will add an xfrm state to the system. +// Equivalent to: `ip xfrm state add $state` +func XfrmStateAdd(state *XfrmState) error { + // A state with spi 0 can't be deleted so don't allow it to be set + if state.Spi == 0 { + return fmt.Errorf("Spi must be set when adding xfrm state.") + } + req := nl.NewNetlinkRequest(nl.XFRM_MSG_NEWSA, syscall.NLM_F_CREATE|syscall.NLM_F_EXCL|syscall.NLM_F_ACK) + + msg := &nl.XfrmUsersaInfo{} + msg.Family = uint16(nl.GetIPFamily(state.Dst)) + msg.Id.Daddr.FromIP(state.Dst) + msg.Saddr.FromIP(state.Src) + msg.Id.Proto = uint8(state.Proto) + msg.Mode = uint8(state.Mode) + msg.Id.Spi = nl.Swap32(uint32(state.Spi)) + msg.Reqid = uint32(state.Reqid) + msg.ReplayWindow = uint8(state.ReplayWindow) + msg.Lft.SoftByteLimit = nl.XFRM_INF + msg.Lft.HardByteLimit = nl.XFRM_INF + msg.Lft.SoftPacketLimit = nl.XFRM_INF + msg.Lft.HardPacketLimit = nl.XFRM_INF + req.AddData(msg) + + if state.Auth != nil { + out := nl.NewRtAttr(nl.XFRMA_ALG_AUTH_TRUNC, writeStateAlgoAuth(state.Auth)) + req.AddData(out) + } + if state.Crypt != nil { + out := nl.NewRtAttr(nl.XFRMA_ALG_CRYPT, writeStateAlgo(state.Crypt)) + req.AddData(out) + } + if state.Encap != nil { + encapData := make([]byte, nl.SizeofXfrmEncapTmpl) + encap := nl.DeserializeXfrmEncapTmpl(encapData) + encap.EncapType = uint16(state.Encap.Type) + encap.EncapSport = nl.Swap16(uint16(state.Encap.SrcPort)) + encap.EncapDport = nl.Swap16(uint16(state.Encap.DstPort)) + encap.EncapOa.FromIP(state.Encap.OriginalAddress) + out := nl.NewRtAttr(nl.XFRMA_ENCAP, encapData) + req.AddData(out) + } + + _, err := req.Execute(syscall.NETLINK_XFRM, 0) + return err +} + +// XfrmStateDel will delete an xfrm state from the system. Note that +// the Algos are ignored when matching the state to delete. +// Equivalent to: `ip xfrm state del $state` +func XfrmStateDel(state *XfrmState) error { + req := nl.NewNetlinkRequest(nl.XFRM_MSG_DELSA, syscall.NLM_F_ACK) + + msg := &nl.XfrmUsersaId{} + msg.Daddr.FromIP(state.Dst) + msg.Family = uint16(nl.GetIPFamily(state.Dst)) + msg.Proto = uint8(state.Proto) + msg.Spi = nl.Swap32(uint32(state.Spi)) + req.AddData(msg) + + saddr := nl.XfrmAddress{} + saddr.FromIP(state.Src) + srcdata := nl.NewRtAttr(nl.XFRMA_SRCADDR, saddr.Serialize()) + + req.AddData(srcdata) + + _, err := req.Execute(syscall.NETLINK_XFRM, 0) + return err +} + +// XfrmStateList gets a list of xfrm states in the system. +// Equivalent to: `ip xfrm state show`. +// The list can be filtered by ip family. +func XfrmStateList(family int) ([]XfrmState, error) { + req := nl.NewNetlinkRequest(nl.XFRM_MSG_GETSA, syscall.NLM_F_DUMP) + + msg := nl.NewIfInfomsg(family) + req.AddData(msg) + + msgs, err := req.Execute(syscall.NETLINK_XFRM, nl.XFRM_MSG_NEWSA) + if err != nil { + return nil, err + } + + var res []XfrmState + for _, m := range msgs { + msg := nl.DeserializeXfrmUsersaInfo(m) + + if family != FAMILY_ALL && family != int(msg.Family) { + continue + } + + var state XfrmState + + state.Dst = msg.Id.Daddr.ToIP() + state.Src = msg.Saddr.ToIP() + state.Proto = Proto(msg.Id.Proto) + state.Mode = Mode(msg.Mode) + state.Spi = int(nl.Swap32(msg.Id.Spi)) + state.Reqid = int(msg.Reqid) + state.ReplayWindow = int(msg.ReplayWindow) + + attrs, err := nl.ParseRouteAttr(m[msg.Len():]) + if err != nil { + return nil, err + } + + for _, attr := range attrs { + switch attr.Attr.Type { + case nl.XFRMA_ALG_AUTH, nl.XFRMA_ALG_CRYPT: + var resAlgo *XfrmStateAlgo + if attr.Attr.Type == nl.XFRMA_ALG_AUTH { + if state.Auth == nil { + state.Auth = new(XfrmStateAlgo) + } + resAlgo = state.Auth + } else { + state.Crypt = new(XfrmStateAlgo) + resAlgo = state.Crypt + } + algo := nl.DeserializeXfrmAlgo(attr.Value[:]) + (*resAlgo).Name = nl.BytesToString(algo.AlgName[:]) + (*resAlgo).Key = algo.AlgKey + case nl.XFRMA_ALG_AUTH_TRUNC: + if state.Auth == nil { + state.Auth = new(XfrmStateAlgo) + } + algo := nl.DeserializeXfrmAlgoAuth(attr.Value[:]) + state.Auth.Name = nl.BytesToString(algo.AlgName[:]) + state.Auth.Key = algo.AlgKey + state.Auth.TruncateLen = int(algo.AlgTruncLen) + case nl.XFRMA_ENCAP: + encap := nl.DeserializeXfrmEncapTmpl(attr.Value[:]) + state.Encap = new(XfrmStateEncap) + state.Encap.Type = EncapType(encap.EncapType) + state.Encap.SrcPort = int(nl.Swap16(encap.EncapSport)) + state.Encap.DstPort = int(nl.Swap16(encap.EncapDport)) + state.Encap.OriginalAddress = encap.EncapOa.ToIP() + } + + } + res = append(res, state) + } + return res, nil +} diff --git a/vendor/github.com/zclconf/go-cty/cty/convert/mismatch_msg.go b/vendor/github.com/zclconf/go-cty/cty/convert/mismatch_msg.go index d91cabf24..88a4a251b 100644 --- a/vendor/github.com/zclconf/go-cty/cty/convert/mismatch_msg.go +++ b/vendor/github.com/zclconf/go-cty/cty/convert/mismatch_msg.go @@ -49,7 +49,7 @@ func MismatchMessage(got, want cty.Type) string { default: // If we have nothing better to say, we'll just state what was required. - return want.FriendlyNameForConstraint() + " required; got %v" + got.GoString() + return want.FriendlyNameForConstraint() + " required" } } diff --git a/vendor/golang.org/x/net/proxy/direct.go b/vendor/golang.org/x/net/proxy/direct.go new file mode 100644 index 000000000..4c5ad88b1 --- /dev/null +++ b/vendor/golang.org/x/net/proxy/direct.go @@ -0,0 +1,18 @@ +// Copyright 2011 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proxy + +import ( + "net" +) + +type direct struct{} + +// Direct is a direct proxy: one that makes network connections directly. +var Direct = direct{} + +func (direct) Dial(network, addr string) (net.Conn, error) { + return net.Dial(network, addr) +} diff --git a/vendor/golang.org/x/net/proxy/per_host.go b/vendor/golang.org/x/net/proxy/per_host.go new file mode 100644 index 000000000..0689bb6a7 --- /dev/null +++ b/vendor/golang.org/x/net/proxy/per_host.go @@ -0,0 +1,140 @@ +// Copyright 2011 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proxy + +import ( + "net" + "strings" +) + +// A PerHost directs connections to a default Dialer unless the host name +// requested matches one of a number of exceptions. +type PerHost struct { + def, bypass Dialer + + bypassNetworks []*net.IPNet + bypassIPs []net.IP + bypassZones []string + bypassHosts []string +} + +// NewPerHost returns a PerHost Dialer that directs connections to either +// defaultDialer or bypass, depending on whether the connection matches one of +// the configured rules. +func NewPerHost(defaultDialer, bypass Dialer) *PerHost { + return &PerHost{ + def: defaultDialer, + bypass: bypass, + } +} + +// Dial connects to the address addr on the given network through either +// defaultDialer or bypass. +func (p *PerHost) Dial(network, addr string) (c net.Conn, err error) { + host, _, err := net.SplitHostPort(addr) + if err != nil { + return nil, err + } + + return p.dialerForRequest(host).Dial(network, addr) +} + +func (p *PerHost) dialerForRequest(host string) Dialer { + if ip := net.ParseIP(host); ip != nil { + for _, net := range p.bypassNetworks { + if net.Contains(ip) { + return p.bypass + } + } + for _, bypassIP := range p.bypassIPs { + if bypassIP.Equal(ip) { + return p.bypass + } + } + return p.def + } + + for _, zone := range p.bypassZones { + if strings.HasSuffix(host, zone) { + return p.bypass + } + if host == zone[1:] { + // For a zone ".example.com", we match "example.com" + // too. + return p.bypass + } + } + for _, bypassHost := range p.bypassHosts { + if bypassHost == host { + return p.bypass + } + } + return p.def +} + +// AddFromString parses a string that contains comma-separated values +// specifying hosts that should use the bypass proxy. Each value is either an +// IP address, a CIDR range, a zone (*.example.com) or a host name +// (localhost). A best effort is made to parse the string and errors are +// ignored. +func (p *PerHost) AddFromString(s string) { + hosts := strings.Split(s, ",") + for _, host := range hosts { + host = strings.TrimSpace(host) + if len(host) == 0 { + continue + } + if strings.Contains(host, "/") { + // We assume that it's a CIDR address like 127.0.0.0/8 + if _, net, err := net.ParseCIDR(host); err == nil { + p.AddNetwork(net) + } + continue + } + if ip := net.ParseIP(host); ip != nil { + p.AddIP(ip) + continue + } + if strings.HasPrefix(host, "*.") { + p.AddZone(host[1:]) + continue + } + p.AddHost(host) + } +} + +// AddIP specifies an IP address that will use the bypass proxy. Note that +// this will only take effect if a literal IP address is dialed. A connection +// to a named host will never match an IP. +func (p *PerHost) AddIP(ip net.IP) { + p.bypassIPs = append(p.bypassIPs, ip) +} + +// AddNetwork specifies an IP range that will use the bypass proxy. Note that +// this will only take effect if a literal IP address is dialed. A connection +// to a named host will never match. +func (p *PerHost) AddNetwork(net *net.IPNet) { + p.bypassNetworks = append(p.bypassNetworks, net) +} + +// AddZone specifies a DNS suffix that will use the bypass proxy. A zone of +// "example.com" matches "example.com" and all of its subdomains. +func (p *PerHost) AddZone(zone string) { + if strings.HasSuffix(zone, ".") { + zone = zone[:len(zone)-1] + } + if !strings.HasPrefix(zone, ".") { + zone = "." + zone + } + p.bypassZones = append(p.bypassZones, zone) +} + +// AddHost specifies a host name that will use the bypass proxy. +func (p *PerHost) AddHost(host string) { + if strings.HasSuffix(host, ".") { + host = host[:len(host)-1] + } + p.bypassHosts = append(p.bypassHosts, host) +} diff --git a/vendor/golang.org/x/net/proxy/proxy.go b/vendor/golang.org/x/net/proxy/proxy.go new file mode 100644 index 000000000..553ead7cf --- /dev/null +++ b/vendor/golang.org/x/net/proxy/proxy.go @@ -0,0 +1,134 @@ +// Copyright 2011 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package proxy provides support for a variety of protocols to proxy network +// data. +package proxy // import "golang.org/x/net/proxy" + +import ( + "errors" + "net" + "net/url" + "os" + "sync" +) + +// A Dialer is a means to establish a connection. +type Dialer interface { + // Dial connects to the given address via the proxy. + Dial(network, addr string) (c net.Conn, err error) +} + +// Auth contains authentication parameters that specific Dialers may require. +type Auth struct { + User, Password string +} + +// FromEnvironment returns the dialer specified by the proxy related variables in +// the environment. +func FromEnvironment() Dialer { + allProxy := allProxyEnv.Get() + if len(allProxy) == 0 { + return Direct + } + + proxyURL, err := url.Parse(allProxy) + if err != nil { + return Direct + } + proxy, err := FromURL(proxyURL, Direct) + if err != nil { + return Direct + } + + noProxy := noProxyEnv.Get() + if len(noProxy) == 0 { + return proxy + } + + perHost := NewPerHost(proxy, Direct) + perHost.AddFromString(noProxy) + return perHost +} + +// proxySchemes is a map from URL schemes to a function that creates a Dialer +// from a URL with such a scheme. +var proxySchemes map[string]func(*url.URL, Dialer) (Dialer, error) + +// RegisterDialerType takes a URL scheme and a function to generate Dialers from +// a URL with that scheme and a forwarding Dialer. Registered schemes are used +// by FromURL. +func RegisterDialerType(scheme string, f func(*url.URL, Dialer) (Dialer, error)) { + if proxySchemes == nil { + proxySchemes = make(map[string]func(*url.URL, Dialer) (Dialer, error)) + } + proxySchemes[scheme] = f +} + +// FromURL returns a Dialer given a URL specification and an underlying +// Dialer for it to make network requests. +func FromURL(u *url.URL, forward Dialer) (Dialer, error) { + var auth *Auth + if u.User != nil { + auth = new(Auth) + auth.User = u.User.Username() + if p, ok := u.User.Password(); ok { + auth.Password = p + } + } + + switch u.Scheme { + case "socks5": + return SOCKS5("tcp", u.Host, auth, forward) + } + + // If the scheme doesn't match any of the built-in schemes, see if it + // was registered by another package. + if proxySchemes != nil { + if f, ok := proxySchemes[u.Scheme]; ok { + return f(u, forward) + } + } + + return nil, errors.New("proxy: unknown scheme: " + u.Scheme) +} + +var ( + allProxyEnv = &envOnce{ + names: []string{"ALL_PROXY", "all_proxy"}, + } + noProxyEnv = &envOnce{ + names: []string{"NO_PROXY", "no_proxy"}, + } +) + +// envOnce looks up an environment variable (optionally by multiple +// names) once. It mitigates expensive lookups on some platforms +// (e.g. Windows). +// (Borrowed from net/http/transport.go) +type envOnce struct { + names []string + once sync.Once + val string +} + +func (e *envOnce) Get() string { + e.once.Do(e.init) + return e.val +} + +func (e *envOnce) init() { + for _, n := range e.names { + e.val = os.Getenv(n) + if e.val != "" { + return + } + } +} + +// reset is used by tests +func (e *envOnce) reset() { + e.once = sync.Once{} + e.val = "" +} diff --git a/vendor/golang.org/x/net/proxy/socks5.go b/vendor/golang.org/x/net/proxy/socks5.go new file mode 100644 index 000000000..3fed38ef1 --- /dev/null +++ b/vendor/golang.org/x/net/proxy/socks5.go @@ -0,0 +1,214 @@ +// Copyright 2011 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proxy + +import ( + "errors" + "io" + "net" + "strconv" +) + +// SOCKS5 returns a Dialer that makes SOCKSv5 connections to the given address +// with an optional username and password. See RFC 1928 and RFC 1929. +func SOCKS5(network, addr string, auth *Auth, forward Dialer) (Dialer, error) { + s := &socks5{ + network: network, + addr: addr, + forward: forward, + } + if auth != nil { + s.user = auth.User + s.password = auth.Password + } + + return s, nil +} + +type socks5 struct { + user, password string + network, addr string + forward Dialer +} + +const socks5Version = 5 + +const ( + socks5AuthNone = 0 + socks5AuthPassword = 2 +) + +const socks5Connect = 1 + +const ( + socks5IP4 = 1 + socks5Domain = 3 + socks5IP6 = 4 +) + +var socks5Errors = []string{ + "", + "general failure", + "connection forbidden", + "network unreachable", + "host unreachable", + "connection refused", + "TTL expired", + "command not supported", + "address type not supported", +} + +// Dial connects to the address addr on the given network via the SOCKS5 proxy. +func (s *socks5) Dial(network, addr string) (net.Conn, error) { + switch network { + case "tcp", "tcp6", "tcp4": + default: + return nil, errors.New("proxy: no support for SOCKS5 proxy connections of type " + network) + } + + conn, err := s.forward.Dial(s.network, s.addr) + if err != nil { + return nil, err + } + if err := s.connect(conn, addr); err != nil { + conn.Close() + return nil, err + } + return conn, nil +} + +// connect takes an existing connection to a socks5 proxy server, +// and commands the server to extend that connection to target, +// which must be a canonical address with a host and port. +func (s *socks5) connect(conn net.Conn, target string) error { + host, portStr, err := net.SplitHostPort(target) + if err != nil { + return err + } + + port, err := strconv.Atoi(portStr) + if err != nil { + return errors.New("proxy: failed to parse port number: " + portStr) + } + if port < 1 || port > 0xffff { + return errors.New("proxy: port number out of range: " + portStr) + } + + // the size here is just an estimate + buf := make([]byte, 0, 6+len(host)) + + buf = append(buf, socks5Version) + if len(s.user) > 0 && len(s.user) < 256 && len(s.password) < 256 { + buf = append(buf, 2 /* num auth methods */, socks5AuthNone, socks5AuthPassword) + } else { + buf = append(buf, 1 /* num auth methods */, socks5AuthNone) + } + + if _, err := conn.Write(buf); err != nil { + return errors.New("proxy: failed to write greeting to SOCKS5 proxy at " + s.addr + ": " + err.Error()) + } + + if _, err := io.ReadFull(conn, buf[:2]); err != nil { + return errors.New("proxy: failed to read greeting from SOCKS5 proxy at " + s.addr + ": " + err.Error()) + } + if buf[0] != 5 { + return errors.New("proxy: SOCKS5 proxy at " + s.addr + " has unexpected version " + strconv.Itoa(int(buf[0]))) + } + if buf[1] == 0xff { + return errors.New("proxy: SOCKS5 proxy at " + s.addr + " requires authentication") + } + + // See RFC 1929 + if buf[1] == socks5AuthPassword { + buf = buf[:0] + buf = append(buf, 1 /* password protocol version */) + buf = append(buf, uint8(len(s.user))) + buf = append(buf, s.user...) + buf = append(buf, uint8(len(s.password))) + buf = append(buf, s.password...) + + if _, err := conn.Write(buf); err != nil { + return errors.New("proxy: failed to write authentication request to SOCKS5 proxy at " + s.addr + ": " + err.Error()) + } + + if _, err := io.ReadFull(conn, buf[:2]); err != nil { + return errors.New("proxy: failed to read authentication reply from SOCKS5 proxy at " + s.addr + ": " + err.Error()) + } + + if buf[1] != 0 { + return errors.New("proxy: SOCKS5 proxy at " + s.addr + " rejected username/password") + } + } + + buf = buf[:0] + buf = append(buf, socks5Version, socks5Connect, 0 /* reserved */) + + if ip := net.ParseIP(host); ip != nil { + if ip4 := ip.To4(); ip4 != nil { + buf = append(buf, socks5IP4) + ip = ip4 + } else { + buf = append(buf, socks5IP6) + } + buf = append(buf, ip...) + } else { + if len(host) > 255 { + return errors.New("proxy: destination host name too long: " + host) + } + buf = append(buf, socks5Domain) + buf = append(buf, byte(len(host))) + buf = append(buf, host...) + } + buf = append(buf, byte(port>>8), byte(port)) + + if _, err := conn.Write(buf); err != nil { + return errors.New("proxy: failed to write connect request to SOCKS5 proxy at " + s.addr + ": " + err.Error()) + } + + if _, err := io.ReadFull(conn, buf[:4]); err != nil { + return errors.New("proxy: failed to read connect reply from SOCKS5 proxy at " + s.addr + ": " + err.Error()) + } + + failure := "unknown error" + if int(buf[1]) < len(socks5Errors) { + failure = socks5Errors[buf[1]] + } + + if len(failure) > 0 { + return errors.New("proxy: SOCKS5 proxy at " + s.addr + " failed to connect: " + failure) + } + + bytesToDiscard := 0 + switch buf[3] { + case socks5IP4: + bytesToDiscard = net.IPv4len + case socks5IP6: + bytesToDiscard = net.IPv6len + case socks5Domain: + _, err := io.ReadFull(conn, buf[:1]) + if err != nil { + return errors.New("proxy: failed to read domain length from SOCKS5 proxy at " + s.addr + ": " + err.Error()) + } + bytesToDiscard = int(buf[0]) + default: + return errors.New("proxy: got unknown address type " + strconv.Itoa(int(buf[3])) + " from SOCKS5 proxy at " + s.addr) + } + + if cap(buf) < bytesToDiscard { + buf = make([]byte, bytesToDiscard) + } else { + buf = buf[:bytesToDiscard] + } + if _, err := io.ReadFull(conn, buf); err != nil { + return errors.New("proxy: failed to read address from SOCKS5 proxy at " + s.addr + ": " + err.Error()) + } + + // Also need to discard the port number + if _, err := io.ReadFull(conn, buf[:2]); err != nil { + return errors.New("proxy: failed to read port from SOCKS5 proxy at " + s.addr + ": " + err.Error()) + } + + return nil +} diff --git a/vendor/golang.org/x/sys/unix/creds_test.go b/vendor/golang.org/x/sys/unix/creds_test.go deleted file mode 100644 index 1b5083196..000000000 --- a/vendor/golang.org/x/sys/unix/creds_test.go +++ /dev/null @@ -1,134 +0,0 @@ -// Copyright 2012 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build linux - -package unix_test - -import ( - "bytes" - "go/build" - "net" - "os" - "testing" - - "golang.org/x/sys/unix" -) - -// TestSCMCredentials tests the sending and receiving of credentials -// (PID, UID, GID) in an ancillary message between two UNIX -// sockets. The SO_PASSCRED socket option is enabled on the sending -// socket for this to work. -func TestSCMCredentials(t *testing.T) { - socketTypeTests := []struct { - socketType int - dataLen int - }{ - { - unix.SOCK_STREAM, - 1, - }, { - unix.SOCK_DGRAM, - 0, - }, - } - - for _, tt := range socketTypeTests { - if tt.socketType == unix.SOCK_DGRAM && !atLeast1p10() { - t.Log("skipping DGRAM test on pre-1.10") - continue - } - - fds, err := unix.Socketpair(unix.AF_LOCAL, tt.socketType, 0) - if err != nil { - t.Fatalf("Socketpair: %v", err) - } - defer unix.Close(fds[0]) - defer unix.Close(fds[1]) - - err = unix.SetsockoptInt(fds[0], unix.SOL_SOCKET, unix.SO_PASSCRED, 1) - if err != nil { - t.Fatalf("SetsockoptInt: %v", err) - } - - srvFile := os.NewFile(uintptr(fds[0]), "server") - defer srvFile.Close() - srv, err := net.FileConn(srvFile) - if err != nil { - t.Errorf("FileConn: %v", err) - return - } - defer srv.Close() - - cliFile := os.NewFile(uintptr(fds[1]), "client") - defer cliFile.Close() - cli, err := net.FileConn(cliFile) - if err != nil { - t.Errorf("FileConn: %v", err) - return - } - defer cli.Close() - - var ucred unix.Ucred - ucred.Pid = int32(os.Getpid()) - ucred.Uid = uint32(os.Getuid()) - ucred.Gid = uint32(os.Getgid()) - oob := unix.UnixCredentials(&ucred) - - // On SOCK_STREAM, this is internally going to send a dummy byte - n, oobn, err := cli.(*net.UnixConn).WriteMsgUnix(nil, oob, nil) - if err != nil { - t.Fatalf("WriteMsgUnix: %v", err) - } - if n != 0 { - t.Fatalf("WriteMsgUnix n = %d, want 0", n) - } - if oobn != len(oob) { - t.Fatalf("WriteMsgUnix oobn = %d, want %d", oobn, len(oob)) - } - - oob2 := make([]byte, 10*len(oob)) - n, oobn2, flags, _, err := srv.(*net.UnixConn).ReadMsgUnix(nil, oob2) - if err != nil { - t.Fatalf("ReadMsgUnix: %v", err) - } - if flags != 0 { - t.Fatalf("ReadMsgUnix flags = 0x%x, want 0", flags) - } - if n != tt.dataLen { - t.Fatalf("ReadMsgUnix n = %d, want %d", n, tt.dataLen) - } - if oobn2 != oobn { - // without SO_PASSCRED set on the socket, ReadMsgUnix will - // return zero oob bytes - t.Fatalf("ReadMsgUnix oobn = %d, want %d", oobn2, oobn) - } - oob2 = oob2[:oobn2] - if !bytes.Equal(oob, oob2) { - t.Fatal("ReadMsgUnix oob bytes don't match") - } - - scm, err := unix.ParseSocketControlMessage(oob2) - if err != nil { - t.Fatalf("ParseSocketControlMessage: %v", err) - } - newUcred, err := unix.ParseUnixCredentials(&scm[0]) - if err != nil { - t.Fatalf("ParseUnixCredentials: %v", err) - } - if *newUcred != ucred { - t.Fatalf("ParseUnixCredentials = %+v, want %+v", newUcred, ucred) - } - } -} - -// atLeast1p10 reports whether we are running on Go 1.10 or later. -func atLeast1p10() bool { - for _, ver := range build.Default.ReleaseTags { - if ver == "go1.10" { - return true - } - } - return false -} diff --git a/vendor/golang.org/x/sys/unix/dev_linux_test.go b/vendor/golang.org/x/sys/unix/dev_linux_test.go deleted file mode 100644 index 51645289c..000000000 --- a/vendor/golang.org/x/sys/unix/dev_linux_test.go +++ /dev/null @@ -1,56 +0,0 @@ -// Copyright 2017 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build go1.7 - -package unix_test - -import ( - "fmt" - "testing" - - "golang.org/x/sys/unix" -) - -func TestDevices(t *testing.T) { - testCases := []struct { - path string - major uint32 - minor uint32 - }{ - // well known major/minor numbers according to - // https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/Documentation/admin-guide/devices.txt - {"/dev/null", 1, 3}, - {"/dev/zero", 1, 5}, - {"/dev/random", 1, 8}, - {"/dev/full", 1, 7}, - {"/dev/urandom", 1, 9}, - {"/dev/tty", 5, 0}, - } - for _, tc := range testCases { - t.Run(fmt.Sprintf("%s %v:%v", tc.path, tc.major, tc.minor), func(t *testing.T) { - var stat unix.Stat_t - err := unix.Stat(tc.path, &stat) - if err != nil { - if err == unix.EACCES { - t.Skip("no permission to stat device, skipping test") - } - t.Errorf("failed to stat device: %v", err) - return - } - - dev := uint64(stat.Rdev) - if unix.Major(dev) != tc.major { - t.Errorf("for %s Major(%#x) == %d, want %d", tc.path, dev, unix.Major(dev), tc.major) - } - if unix.Minor(dev) != tc.minor { - t.Errorf("for %s Minor(%#x) == %d, want %d", tc.path, dev, unix.Minor(dev), tc.minor) - } - if unix.Mkdev(tc.major, tc.minor) != dev { - t.Errorf("for %s Mkdev(%d, %d) == %#x, want %#x", tc.path, tc.major, tc.minor, unix.Mkdev(tc.major, tc.minor), dev) - } - }) - - } -} diff --git a/vendor/golang.org/x/sys/unix/example_test.go b/vendor/golang.org/x/sys/unix/example_test.go deleted file mode 100644 index 10619afdd..000000000 --- a/vendor/golang.org/x/sys/unix/example_test.go +++ /dev/null @@ -1,19 +0,0 @@ -// Copyright 2018 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build darwin dragonfly freebsd linux netbsd openbsd solaris - -package unix_test - -import ( - "log" - "os" - - "golang.org/x/sys/unix" -) - -func ExampleExec() { - err := unix.Exec("/bin/ls", []string{"ls", "-al"}, os.Environ()) - log.Fatal(err) -} diff --git a/vendor/golang.org/x/sys/unix/export_test.go b/vendor/golang.org/x/sys/unix/export_test.go deleted file mode 100644 index e8024690d..000000000 --- a/vendor/golang.org/x/sys/unix/export_test.go +++ /dev/null @@ -1,9 +0,0 @@ -// Copyright 2015 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build darwin dragonfly freebsd linux netbsd openbsd solaris - -package unix - -var Itoa = itoa diff --git a/vendor/golang.org/x/sys/unix/mkpost.go b/vendor/golang.org/x/sys/unix/mkpost.go deleted file mode 100644 index 7e5c22c47..000000000 --- a/vendor/golang.org/x/sys/unix/mkpost.go +++ /dev/null @@ -1,98 +0,0 @@ -// Copyright 2016 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build ignore - -// mkpost processes the output of cgo -godefs to -// modify the generated types. It is used to clean up -// the sys API in an architecture specific manner. -// -// mkpost is run after cgo -godefs; see README.md. -package main - -import ( - "bytes" - "fmt" - "go/format" - "io/ioutil" - "log" - "os" - "regexp" -) - -func main() { - // Get the OS and architecture (using GOARCH_TARGET if it exists) - goos := os.Getenv("GOOS") - goarch := os.Getenv("GOARCH_TARGET") - if goarch == "" { - goarch = os.Getenv("GOARCH") - } - // Check that we are using the new build system if we should be. - if goos == "linux" && goarch != "sparc64" { - if os.Getenv("GOLANG_SYS_BUILD") != "docker" { - os.Stderr.WriteString("In the new build system, mkpost should not be called directly.\n") - os.Stderr.WriteString("See README.md\n") - os.Exit(1) - } - } - - b, err := ioutil.ReadAll(os.Stdin) - if err != nil { - log.Fatal(err) - } - - // Intentionally export __val fields in Fsid and Sigset_t - valRegex := regexp.MustCompile(`type (Fsid|Sigset_t) struct {(\s+)X__val(\s+\S+\s+)}`) - b = valRegex.ReplaceAll(b, []byte("type $1 struct {${2}Val$3}")) - - // If we have empty Ptrace structs, we should delete them. Only s390x emits - // nonempty Ptrace structs. - ptraceRexexp := regexp.MustCompile(`type Ptrace((Psw|Fpregs|Per) struct {\s*})`) - b = ptraceRexexp.ReplaceAll(b, nil) - - // Replace the control_regs union with a blank identifier for now. - controlRegsRegex := regexp.MustCompile(`(Control_regs)\s+\[0\]uint64`) - b = controlRegsRegex.ReplaceAll(b, []byte("_ [0]uint64")) - - // Remove fields that are added by glibc - // Note that this is unstable as the identifers are private. - removeFieldsRegex := regexp.MustCompile(`X__glibc\S*`) - b = removeFieldsRegex.ReplaceAll(b, []byte("_")) - - // Convert [65]int8 to [65]byte in Utsname members to simplify - // conversion to string; see golang.org/issue/20753 - convertUtsnameRegex := regexp.MustCompile(`((Sys|Node|Domain)name|Release|Version|Machine)(\s+)\[(\d+)\]u?int8`) - b = convertUtsnameRegex.ReplaceAll(b, []byte("$1$3[$4]byte")) - - // Remove spare fields (e.g. in Statx_t) - spareFieldsRegex := regexp.MustCompile(`X__spare\S*`) - b = spareFieldsRegex.ReplaceAll(b, []byte("_")) - - // Remove cgo padding fields - removePaddingFieldsRegex := regexp.MustCompile(`Pad_cgo_\d+`) - b = removePaddingFieldsRegex.ReplaceAll(b, []byte("_")) - - // Remove padding, hidden, or unused fields - removeFieldsRegex = regexp.MustCompile(`\b(X_\S+|Padding)`) - b = removeFieldsRegex.ReplaceAll(b, []byte("_")) - - // Remove the first line of warning from cgo - b = b[bytes.IndexByte(b, '\n')+1:] - // Modify the command in the header to include: - // mkpost, our own warning, and a build tag. - replacement := fmt.Sprintf(`$1 | go run mkpost.go -// Code generated by the command above; see README.md. DO NOT EDIT. - -// +build %s,%s`, goarch, goos) - cgoCommandRegex := regexp.MustCompile(`(cgo -godefs .*)`) - b = cgoCommandRegex.ReplaceAll(b, []byte(replacement)) - - // gofmt - b, err = format.Source(b) - if err != nil { - log.Fatal(err) - } - - os.Stdout.Write(b) -} diff --git a/vendor/golang.org/x/sys/unix/mmap_unix_test.go b/vendor/golang.org/x/sys/unix/mmap_unix_test.go deleted file mode 100644 index 3258ca328..000000000 --- a/vendor/golang.org/x/sys/unix/mmap_unix_test.go +++ /dev/null @@ -1,35 +0,0 @@ -// Copyright 2014 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build darwin dragonfly freebsd linux netbsd openbsd solaris - -package unix_test - -import ( - "testing" - - "golang.org/x/sys/unix" -) - -func TestMmap(t *testing.T) { - b, err := unix.Mmap(-1, 0, unix.Getpagesize(), unix.PROT_NONE, unix.MAP_ANON|unix.MAP_PRIVATE) - if err != nil { - t.Fatalf("Mmap: %v", err) - } - if err := unix.Mprotect(b, unix.PROT_READ|unix.PROT_WRITE); err != nil { - t.Fatalf("Mprotect: %v", err) - } - - b[0] = 42 - - if err := unix.Msync(b, unix.MS_SYNC); err != nil { - t.Fatalf("Msync: %v", err) - } - if err := unix.Madvise(b, unix.MADV_DONTNEED); err != nil { - t.Fatalf("Madvise: %v", err) - } - if err := unix.Munmap(b); err != nil { - t.Fatalf("Munmap: %v", err) - } -} diff --git a/vendor/golang.org/x/sys/unix/openbsd_test.go b/vendor/golang.org/x/sys/unix/openbsd_test.go deleted file mode 100644 index 734d76585..000000000 --- a/vendor/golang.org/x/sys/unix/openbsd_test.go +++ /dev/null @@ -1,113 +0,0 @@ -// Copyright 2016 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build openbsd - -// This, on the face of it, bizarre testing mechanism is necessary because -// the only reliable way to gauge whether or not a pledge(2) call has succeeded -// is that the program has been killed as a result of breaking its pledge. - -package unix_test - -import ( - "flag" - "fmt" - "io/ioutil" - "os" - "os/exec" - "path/filepath" - "testing" - - "golang.org/x/sys/unix" -) - -type testProc struct { - fn func() // should always exit instead of returning - cleanup func() error // for instance, delete coredumps from testing pledge - success bool // whether zero-exit means success or failure -} - -var ( - testProcs = map[string]testProc{} - procName = "" -) - -const ( - optName = "sys-unix-internal-procname" -) - -func init() { - flag.StringVar(&procName, optName, "", "internal use only") -} - -// testCmd generates a proper command that, when executed, runs the test -// corresponding to the given key. -func testCmd(procName string) (*exec.Cmd, error) { - exe, err := filepath.Abs(os.Args[0]) - if err != nil { - return nil, err - } - cmd := exec.Command(exe, "-"+optName+"="+procName) - cmd.Stdout, cmd.Stderr = os.Stdout, os.Stderr - return cmd, nil -} - -// ExitsCorrectly is a comprehensive, one-line-of-use wrapper for testing -// a testProc with a key. -func ExitsCorrectly(procName string, t *testing.T) { - s := testProcs[procName] - c, err := testCmd(procName) - defer func() { - if s.cleanup() != nil { - t.Fatalf("Failed to run cleanup for %s", procName) - } - }() - if err != nil { - t.Fatalf("Failed to construct command for %s", procName) - } - if (c.Run() == nil) != s.success { - result := "succeed" - if !s.success { - result = "fail" - } - t.Fatalf("Process did not %s when it was supposed to", result) - } -} - -func TestMain(m *testing.M) { - flag.Parse() - if procName != "" { - testProcs[procName].fn() - } - os.Exit(m.Run()) -} - -// For example, add a test for pledge. -func init() { - testProcs["pledge"] = testProc{ - func() { - fmt.Println(unix.Pledge("", nil)) - os.Exit(0) - }, - func() error { - files, err := ioutil.ReadDir(".") - if err != nil { - return err - } - for _, file := range files { - if filepath.Ext(file.Name()) == ".core" { - if err := os.Remove(file.Name()); err != nil { - return err - } - } - } - return nil - }, - false, - } -} - -func TestPledge(t *testing.T) { - ExitsCorrectly("pledge", t) -} diff --git a/vendor/golang.org/x/sys/unix/syscall_bsd_test.go b/vendor/golang.org/x/sys/unix/syscall_bsd_test.go deleted file mode 100644 index 6c4e2aca0..000000000 --- a/vendor/golang.org/x/sys/unix/syscall_bsd_test.go +++ /dev/null @@ -1,93 +0,0 @@ -// Copyright 2014 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build darwin dragonfly freebsd openbsd - -package unix_test - -import ( - "os/exec" - "runtime" - "testing" - "time" - - "golang.org/x/sys/unix" -) - -const MNT_WAIT = 1 -const MNT_NOWAIT = 2 - -func TestGetfsstat(t *testing.T) { - const flags = MNT_NOWAIT // see golang.org/issue/16937 - n, err := unix.Getfsstat(nil, flags) - if err != nil { - t.Fatal(err) - } - - data := make([]unix.Statfs_t, n) - n2, err := unix.Getfsstat(data, flags) - if err != nil { - t.Fatal(err) - } - if n != n2 { - t.Errorf("Getfsstat(nil) = %d, but subsequent Getfsstat(slice) = %d", n, n2) - } - for i, stat := range data { - if stat == (unix.Statfs_t{}) { - t.Errorf("index %v is an empty Statfs_t struct", i) - } - } - if t.Failed() { - for i, stat := range data[:n2] { - t.Logf("data[%v] = %+v", i, stat) - } - mount, err := exec.Command("mount").CombinedOutput() - if err != nil { - t.Logf("mount: %v\n%s", err, mount) - } else { - t.Logf("mount: %s", mount) - } - } -} - -func TestSelect(t *testing.T) { - err := unix.Select(0, nil, nil, nil, &unix.Timeval{Sec: 0, Usec: 0}) - if err != nil { - t.Fatalf("Select: %v", err) - } - - dur := 250 * time.Millisecond - tv := unix.NsecToTimeval(int64(dur)) - start := time.Now() - err = unix.Select(0, nil, nil, nil, &tv) - took := time.Since(start) - if err != nil { - t.Fatalf("Select: %v", err) - } - - // On some BSDs the actual timeout might also be slightly less than the requested. - // Add an acceptable margin to avoid flaky tests. - if took < dur*2/3 { - t.Errorf("Select: timeout should have been at least %v, got %v", dur, took) - } -} - -func TestSysctlRaw(t *testing.T) { - if runtime.GOOS == "openbsd" { - t.Skip("kern.proc.pid does not exist on OpenBSD") - } - - _, err := unix.SysctlRaw("kern.proc.pid", unix.Getpid()) - if err != nil { - t.Fatal(err) - } -} - -func TestSysctlUint32(t *testing.T) { - maxproc, err := unix.SysctlUint32("kern.maxproc") - if err != nil { - t.Fatal(err) - } - t.Logf("kern.maxproc: %v", maxproc) -} diff --git a/vendor/golang.org/x/sys/unix/syscall_darwin_test.go b/vendor/golang.org/x/sys/unix/syscall_darwin_test.go deleted file mode 100644 index 65691d5c1..000000000 --- a/vendor/golang.org/x/sys/unix/syscall_darwin_test.go +++ /dev/null @@ -1,19 +0,0 @@ -// Copyright 2018 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -package unix_test - -// stringsFromByteSlice converts a sequence of attributes to a []string. -// On Darwin, each entry is a NULL-terminated string. -func stringsFromByteSlice(buf []byte) []string { - var result []string - off := 0 - for i, b := range buf { - if b == 0 { - result = append(result, string(buf[off:i])) - off = i + 1 - } - } - return result -} diff --git a/vendor/golang.org/x/sys/unix/syscall_freebsd_test.go b/vendor/golang.org/x/sys/unix/syscall_freebsd_test.go deleted file mode 100644 index 0fec1a827..000000000 --- a/vendor/golang.org/x/sys/unix/syscall_freebsd_test.go +++ /dev/null @@ -1,312 +0,0 @@ -// Copyright 2014 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build freebsd - -package unix_test - -import ( - "flag" - "fmt" - "io/ioutil" - "os" - "os/exec" - "path" - "path/filepath" - "runtime" - "testing" - - "golang.org/x/sys/unix" -) - -func TestSysctlUint64(t *testing.T) { - _, err := unix.SysctlUint64("vm.swap_total") - if err != nil { - t.Fatal(err) - } -} - -// FIXME: Infrastructure for launching tests in subprocesses stolen from openbsd_test.go - refactor? -// testCmd generates a proper command that, when executed, runs the test -// corresponding to the given key. - -type testProc struct { - fn func() // should always exit instead of returning - arg func(t *testing.T) string // generate argument for test - cleanup func(arg string) error // for instance, delete coredumps from testing pledge - success bool // whether zero-exit means success or failure -} - -var ( - testProcs = map[string]testProc{} - procName = "" - procArg = "" -) - -const ( - optName = "sys-unix-internal-procname" - optArg = "sys-unix-internal-arg" -) - -func init() { - flag.StringVar(&procName, optName, "", "internal use only") - flag.StringVar(&procArg, optArg, "", "internal use only") - -} - -func testCmd(procName string, procArg string) (*exec.Cmd, error) { - exe, err := filepath.Abs(os.Args[0]) - if err != nil { - return nil, err - } - cmd := exec.Command(exe, "-"+optName+"="+procName, "-"+optArg+"="+procArg) - cmd.Stdout, cmd.Stderr = os.Stdout, os.Stderr - return cmd, nil -} - -// ExitsCorrectly is a comprehensive, one-line-of-use wrapper for testing -// a testProc with a key. -func ExitsCorrectly(t *testing.T, procName string) { - s := testProcs[procName] - arg := "-" - if s.arg != nil { - arg = s.arg(t) - } - c, err := testCmd(procName, arg) - defer func(arg string) { - if err := s.cleanup(arg); err != nil { - t.Fatalf("Failed to run cleanup for %s %s %#v", procName, err, err) - } - }(arg) - if err != nil { - t.Fatalf("Failed to construct command for %s", procName) - } - if (c.Run() == nil) != s.success { - result := "succeed" - if !s.success { - result = "fail" - } - t.Fatalf("Process did not %s when it was supposed to", result) - } -} - -func TestMain(m *testing.M) { - flag.Parse() - if procName != "" { - t := testProcs[procName] - t.fn() - os.Stderr.WriteString("test function did not exit\n") - if t.success { - os.Exit(1) - } else { - os.Exit(0) - } - } - os.Exit(m.Run()) -} - -// end of infrastructure - -const testfile = "gocapmodetest" -const testfile2 = testfile + "2" - -func CapEnterTest() { - _, err := os.OpenFile(path.Join(procArg, testfile), os.O_WRONLY|os.O_CREATE|os.O_APPEND, 0666) - if err != nil { - panic(fmt.Sprintf("OpenFile: %s", err)) - } - - err = unix.CapEnter() - if err != nil { - panic(fmt.Sprintf("CapEnter: %s", err)) - } - - _, err = os.OpenFile(path.Join(procArg, testfile2), os.O_WRONLY|os.O_CREATE|os.O_APPEND, 0666) - if err == nil { - panic("OpenFile works!") - } - if err.(*os.PathError).Err != unix.ECAPMODE { - panic(fmt.Sprintf("OpenFile failed wrong: %s %#v", err, err)) - } - os.Exit(0) -} - -func makeTempDir(t *testing.T) string { - d, err := ioutil.TempDir("", "go_openat_test") - if err != nil { - t.Fatalf("TempDir failed: %s", err) - } - return d -} - -func removeTempDir(arg string) error { - err := os.RemoveAll(arg) - if err != nil && err.(*os.PathError).Err == unix.ENOENT { - return nil - } - return err -} - -func init() { - testProcs["cap_enter"] = testProc{ - CapEnterTest, - makeTempDir, - removeTempDir, - true, - } -} - -func TestCapEnter(t *testing.T) { - if runtime.GOARCH != "amd64" { - t.Skipf("skipping test on %s", runtime.GOARCH) - } - ExitsCorrectly(t, "cap_enter") -} - -func OpenatTest() { - f, err := os.Open(procArg) - if err != nil { - panic(err) - } - - err = unix.CapEnter() - if err != nil { - panic(fmt.Sprintf("CapEnter: %s", err)) - } - - fxx, err := unix.Openat(int(f.Fd()), "xx", os.O_WRONLY|os.O_CREATE|os.O_APPEND, 0666) - if err != nil { - panic(err) - } - unix.Close(fxx) - - // The right to open BASE/xx is not ambient - _, err = os.OpenFile(procArg+"/xx", os.O_WRONLY|os.O_CREATE|os.O_APPEND, 0666) - if err == nil { - panic("OpenFile succeeded") - } - if err.(*os.PathError).Err != unix.ECAPMODE { - panic(fmt.Sprintf("OpenFile failed wrong: %s %#v", err, err)) - } - - // Can't make a new directory either - err = os.Mkdir(procArg+"2", 0777) - if err == nil { - panic("MKdir succeeded") - } - if err.(*os.PathError).Err != unix.ECAPMODE { - panic(fmt.Sprintf("Mkdir failed wrong: %s %#v", err, err)) - } - - // Remove all caps except read and lookup. - r, err := unix.CapRightsInit([]uint64{unix.CAP_READ, unix.CAP_LOOKUP}) - if err != nil { - panic(fmt.Sprintf("CapRightsInit failed: %s %#v", err, err)) - } - err = unix.CapRightsLimit(f.Fd(), r) - if err != nil { - panic(fmt.Sprintf("CapRightsLimit failed: %s %#v", err, err)) - } - - // Check we can get the rights back again - r, err = unix.CapRightsGet(f.Fd()) - if err != nil { - panic(fmt.Sprintf("CapRightsGet failed: %s %#v", err, err)) - } - b, err := unix.CapRightsIsSet(r, []uint64{unix.CAP_READ, unix.CAP_LOOKUP}) - if err != nil { - panic(fmt.Sprintf("CapRightsIsSet failed: %s %#v", err, err)) - } - if !b { - panic(fmt.Sprintf("Unexpected rights")) - } - b, err = unix.CapRightsIsSet(r, []uint64{unix.CAP_READ, unix.CAP_LOOKUP, unix.CAP_WRITE}) - if err != nil { - panic(fmt.Sprintf("CapRightsIsSet failed: %s %#v", err, err)) - } - if b { - panic(fmt.Sprintf("Unexpected rights (2)")) - } - - // Can no longer create a file - _, err = unix.Openat(int(f.Fd()), "xx2", os.O_WRONLY|os.O_CREATE|os.O_APPEND, 0666) - if err == nil { - panic("Openat succeeded") - } - if err != unix.ENOTCAPABLE { - panic(fmt.Sprintf("OpenFileAt failed wrong: %s %#v", err, err)) - } - - // But can read an existing one - _, err = unix.Openat(int(f.Fd()), "xx", os.O_RDONLY, 0666) - if err != nil { - panic(fmt.Sprintf("Openat failed: %s %#v", err, err)) - } - - os.Exit(0) -} - -func init() { - testProcs["openat"] = testProc{ - OpenatTest, - makeTempDir, - removeTempDir, - true, - } -} - -func TestOpenat(t *testing.T) { - if runtime.GOARCH != "amd64" { - t.Skipf("skipping test on %s", runtime.GOARCH) - } - ExitsCorrectly(t, "openat") -} - -func TestCapRightsSetAndClear(t *testing.T) { - r, err := unix.CapRightsInit([]uint64{unix.CAP_READ, unix.CAP_WRITE, unix.CAP_PDWAIT}) - if err != nil { - t.Fatalf("CapRightsInit failed: %s", err) - } - - err = unix.CapRightsSet(r, []uint64{unix.CAP_EVENT, unix.CAP_LISTEN}) - if err != nil { - t.Fatalf("CapRightsSet failed: %s", err) - } - - b, err := unix.CapRightsIsSet(r, []uint64{unix.CAP_READ, unix.CAP_WRITE, unix.CAP_PDWAIT, unix.CAP_EVENT, unix.CAP_LISTEN}) - if err != nil { - t.Fatalf("CapRightsIsSet failed: %s", err) - } - if !b { - t.Fatalf("Wrong rights set") - } - - err = unix.CapRightsClear(r, []uint64{unix.CAP_READ, unix.CAP_PDWAIT}) - if err != nil { - t.Fatalf("CapRightsClear failed: %s", err) - } - - b, err = unix.CapRightsIsSet(r, []uint64{unix.CAP_WRITE, unix.CAP_EVENT, unix.CAP_LISTEN}) - if err != nil { - t.Fatalf("CapRightsIsSet failed: %s", err) - } - if !b { - t.Fatalf("Wrong rights set") - } -} - -// stringsFromByteSlice converts a sequence of attributes to a []string. -// On FreeBSD, each entry consists of a single byte containing the length -// of the attribute name, followed by the attribute name. -// The name is _not_ NULL-terminated. -func stringsFromByteSlice(buf []byte) []string { - var result []string - i := 0 - for i < len(buf) { - next := i + 1 + int(buf[i]) - result = append(result, string(buf[i+1:next])) - i = next - } - return result -} diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_test.go b/vendor/golang.org/x/sys/unix/syscall_linux_test.go deleted file mode 100644 index eed17268b..000000000 --- a/vendor/golang.org/x/sys/unix/syscall_linux_test.go +++ /dev/null @@ -1,421 +0,0 @@ -// Copyright 2016 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build linux - -package unix_test - -import ( - "os" - "runtime" - "runtime/debug" - "testing" - "time" - - "golang.org/x/sys/unix" -) - -func TestIoctlGetInt(t *testing.T) { - f, err := os.Open("/dev/random") - if err != nil { - t.Fatalf("failed to open device: %v", err) - } - defer f.Close() - - v, err := unix.IoctlGetInt(int(f.Fd()), unix.RNDGETENTCNT) - if err != nil { - t.Fatalf("failed to perform ioctl: %v", err) - } - - t.Logf("%d bits of entropy available", v) -} - -func TestPpoll(t *testing.T) { - if runtime.GOOS == "android" { - t.Skip("mkfifo syscall is not available on android, skipping test") - } - - f, cleanup := mktmpfifo(t) - defer cleanup() - - const timeout = 100 * time.Millisecond - - ok := make(chan bool, 1) - go func() { - select { - case <-time.After(10 * timeout): - t.Errorf("Ppoll: failed to timeout after %d", 10*timeout) - case <-ok: - } - }() - - fds := []unix.PollFd{{Fd: int32(f.Fd()), Events: unix.POLLIN}} - timeoutTs := unix.NsecToTimespec(int64(timeout)) - n, err := unix.Ppoll(fds, &timeoutTs, nil) - ok <- true - if err != nil { - t.Errorf("Ppoll: unexpected error: %v", err) - return - } - if n != 0 { - t.Errorf("Ppoll: wrong number of events: got %v, expected %v", n, 0) - return - } -} - -func TestTime(t *testing.T) { - var ut unix.Time_t - ut2, err := unix.Time(&ut) - if err != nil { - t.Fatalf("Time: %v", err) - } - if ut != ut2 { - t.Errorf("Time: return value %v should be equal to argument %v", ut2, ut) - } - - var now time.Time - - for i := 0; i < 10; i++ { - ut, err = unix.Time(nil) - if err != nil { - t.Fatalf("Time: %v", err) - } - - now = time.Now() - - if int64(ut) == now.Unix() { - return - } - } - - t.Errorf("Time: return value %v should be nearly equal to time.Now().Unix() %v", ut, now.Unix()) -} - -func TestUtime(t *testing.T) { - defer chtmpdir(t)() - - touch(t, "file1") - - buf := &unix.Utimbuf{ - Modtime: 12345, - } - - err := unix.Utime("file1", buf) - if err != nil { - t.Fatalf("Utime: %v", err) - } - - fi, err := os.Stat("file1") - if err != nil { - t.Fatal(err) - } - - if fi.ModTime().Unix() != 12345 { - t.Errorf("Utime: failed to change modtime: expected %v, got %v", 12345, fi.ModTime().Unix()) - } -} - -func TestUtimesNanoAt(t *testing.T) { - defer chtmpdir(t)() - - symlink := "symlink1" - os.Remove(symlink) - err := os.Symlink("nonexisting", symlink) - if err != nil { - t.Fatal(err) - } - - ts := []unix.Timespec{ - {Sec: 1111, Nsec: 2222}, - {Sec: 3333, Nsec: 4444}, - } - err = unix.UtimesNanoAt(unix.AT_FDCWD, symlink, ts, unix.AT_SYMLINK_NOFOLLOW) - if err != nil { - t.Fatalf("UtimesNanoAt: %v", err) - } - - var st unix.Stat_t - err = unix.Lstat(symlink, &st) - if err != nil { - t.Fatalf("Lstat: %v", err) - } - - // Only check Mtim, Atim might not be supported by the underlying filesystem - expected := ts[1] - if st.Mtim.Nsec == 0 { - // Some filesystems only support 1-second time stamp resolution - // and will always set Nsec to 0. - expected.Nsec = 0 - } - if st.Mtim != expected { - t.Errorf("UtimesNanoAt: wrong mtime: expected %v, got %v", expected, st.Mtim) - } -} - -func TestRlimitAs(t *testing.T) { - // disable GC during to avoid flaky test - defer debug.SetGCPercent(debug.SetGCPercent(-1)) - - var rlim unix.Rlimit - err := unix.Getrlimit(unix.RLIMIT_AS, &rlim) - if err != nil { - t.Fatalf("Getrlimit: %v", err) - } - var zero unix.Rlimit - if zero == rlim { - t.Fatalf("Getrlimit: got zero value %#v", rlim) - } - set := rlim - set.Cur = uint64(unix.Getpagesize()) - err = unix.Setrlimit(unix.RLIMIT_AS, &set) - if err != nil { - t.Fatalf("Setrlimit: set failed: %#v %v", set, err) - } - - // RLIMIT_AS was set to the page size, so mmap()'ing twice the page size - // should fail. See 'man 2 getrlimit'. - _, err = unix.Mmap(-1, 0, 2*unix.Getpagesize(), unix.PROT_NONE, unix.MAP_ANON|unix.MAP_PRIVATE) - if err == nil { - t.Fatal("Mmap: unexpectedly suceeded after setting RLIMIT_AS") - } - - err = unix.Setrlimit(unix.RLIMIT_AS, &rlim) - if err != nil { - t.Fatalf("Setrlimit: restore failed: %#v %v", rlim, err) - } - - b, err := unix.Mmap(-1, 0, 2*unix.Getpagesize(), unix.PROT_NONE, unix.MAP_ANON|unix.MAP_PRIVATE) - if err != nil { - t.Fatalf("Mmap: %v", err) - } - err = unix.Munmap(b) - if err != nil { - t.Fatalf("Munmap: %v", err) - } -} - -func TestSelect(t *testing.T) { - _, err := unix.Select(0, nil, nil, nil, &unix.Timeval{Sec: 0, Usec: 0}) - if err != nil { - t.Fatalf("Select: %v", err) - } - - dur := 150 * time.Millisecond - tv := unix.NsecToTimeval(int64(dur)) - start := time.Now() - _, err = unix.Select(0, nil, nil, nil, &tv) - took := time.Since(start) - if err != nil { - t.Fatalf("Select: %v", err) - } - - if took < dur { - t.Errorf("Select: timeout should have been at least %v, got %v", dur, took) - } -} - -func TestPselect(t *testing.T) { - _, err := unix.Pselect(0, nil, nil, nil, &unix.Timespec{Sec: 0, Nsec: 0}, nil) - if err != nil { - t.Fatalf("Pselect: %v", err) - } - - dur := 2500 * time.Microsecond - ts := unix.NsecToTimespec(int64(dur)) - start := time.Now() - _, err = unix.Pselect(0, nil, nil, nil, &ts, nil) - took := time.Since(start) - if err != nil { - t.Fatalf("Pselect: %v", err) - } - - if took < dur { - t.Errorf("Pselect: timeout should have been at least %v, got %v", dur, took) - } -} - -func TestSchedSetaffinity(t *testing.T) { - runtime.LockOSThread() - defer runtime.UnlockOSThread() - - var oldMask unix.CPUSet - err := unix.SchedGetaffinity(0, &oldMask) - if err != nil { - t.Fatalf("SchedGetaffinity: %v", err) - } - - var newMask unix.CPUSet - newMask.Zero() - if newMask.Count() != 0 { - t.Errorf("CpuZero: didn't zero CPU set: %v", newMask) - } - cpu := 1 - newMask.Set(cpu) - if newMask.Count() != 1 || !newMask.IsSet(cpu) { - t.Errorf("CpuSet: didn't set CPU %d in set: %v", cpu, newMask) - } - cpu = 5 - newMask.Set(cpu) - if newMask.Count() != 2 || !newMask.IsSet(cpu) { - t.Errorf("CpuSet: didn't set CPU %d in set: %v", cpu, newMask) - } - newMask.Clear(cpu) - if newMask.Count() != 1 || newMask.IsSet(cpu) { - t.Errorf("CpuClr: didn't clear CPU %d in set: %v", cpu, newMask) - } - - if runtime.NumCPU() < 2 { - t.Skip("skipping setaffinity tests on single CPU system") - } - if runtime.GOOS == "android" { - t.Skip("skipping setaffinity tests on android") - } - - err = unix.SchedSetaffinity(0, &newMask) - if err != nil { - t.Fatalf("SchedSetaffinity: %v", err) - } - - var gotMask unix.CPUSet - err = unix.SchedGetaffinity(0, &gotMask) - if err != nil { - t.Fatalf("SchedGetaffinity: %v", err) - } - - if gotMask != newMask { - t.Errorf("SchedSetaffinity: returned affinity mask does not match set affinity mask") - } - - // Restore old mask so it doesn't affect successive tests - err = unix.SchedSetaffinity(0, &oldMask) - if err != nil { - t.Fatalf("SchedSetaffinity: %v", err) - } -} - -func TestStatx(t *testing.T) { - var stx unix.Statx_t - err := unix.Statx(unix.AT_FDCWD, ".", 0, 0, &stx) - if err == unix.ENOSYS || err == unix.EPERM { - t.Skip("statx syscall is not available, skipping test") - } else if err != nil { - t.Fatalf("Statx: %v", err) - } - - defer chtmpdir(t)() - touch(t, "file1") - - var st unix.Stat_t - err = unix.Stat("file1", &st) - if err != nil { - t.Fatalf("Stat: %v", err) - } - - flags := unix.AT_STATX_SYNC_AS_STAT - err = unix.Statx(unix.AT_FDCWD, "file1", flags, unix.STATX_ALL, &stx) - if err != nil { - t.Fatalf("Statx: %v", err) - } - - if uint32(stx.Mode) != st.Mode { - t.Errorf("Statx: returned stat mode does not match Stat") - } - - ctime := unix.StatxTimestamp{Sec: int64(st.Ctim.Sec), Nsec: uint32(st.Ctim.Nsec)} - mtime := unix.StatxTimestamp{Sec: int64(st.Mtim.Sec), Nsec: uint32(st.Mtim.Nsec)} - - if stx.Ctime != ctime { - t.Errorf("Statx: returned stat ctime does not match Stat") - } - if stx.Mtime != mtime { - t.Errorf("Statx: returned stat mtime does not match Stat") - } - - err = os.Symlink("file1", "symlink1") - if err != nil { - t.Fatal(err) - } - - err = unix.Lstat("symlink1", &st) - if err != nil { - t.Fatalf("Lstat: %v", err) - } - - err = unix.Statx(unix.AT_FDCWD, "symlink1", flags, unix.STATX_BASIC_STATS, &stx) - if err != nil { - t.Fatalf("Statx: %v", err) - } - - // follow symlink, expect a regulat file - if stx.Mode&unix.S_IFREG == 0 { - t.Errorf("Statx: didn't follow symlink") - } - - err = unix.Statx(unix.AT_FDCWD, "symlink1", flags|unix.AT_SYMLINK_NOFOLLOW, unix.STATX_ALL, &stx) - if err != nil { - t.Fatalf("Statx: %v", err) - } - - // follow symlink, expect a symlink - if stx.Mode&unix.S_IFLNK == 0 { - t.Errorf("Statx: unexpectedly followed symlink") - } - if uint32(stx.Mode) != st.Mode { - t.Errorf("Statx: returned stat mode does not match Lstat") - } - - ctime = unix.StatxTimestamp{Sec: int64(st.Ctim.Sec), Nsec: uint32(st.Ctim.Nsec)} - mtime = unix.StatxTimestamp{Sec: int64(st.Mtim.Sec), Nsec: uint32(st.Mtim.Nsec)} - - if stx.Ctime != ctime { - t.Errorf("Statx: returned stat ctime does not match Lstat") - } - if stx.Mtime != mtime { - t.Errorf("Statx: returned stat mtime does not match Lstat") - } -} - -// stringsFromByteSlice converts a sequence of attributes to a []string. -// On Linux, each entry is a NULL-terminated string. -func stringsFromByteSlice(buf []byte) []string { - var result []string - off := 0 - for i, b := range buf { - if b == 0 { - result = append(result, string(buf[off:i])) - off = i + 1 - } - } - return result -} - -func TestFaccessat(t *testing.T) { - defer chtmpdir(t)() - touch(t, "file1") - - err := unix.Faccessat(unix.AT_FDCWD, "file1", unix.O_RDONLY, 0) - if err != nil { - t.Errorf("Faccessat: unexpected error: %v", err) - } - - err = unix.Faccessat(unix.AT_FDCWD, "file1", unix.O_RDONLY, 2) - if err != unix.EINVAL { - t.Errorf("Faccessat: unexpected error: %v, want EINVAL", err) - } - - err = unix.Faccessat(unix.AT_FDCWD, "file1", unix.O_RDONLY, unix.AT_EACCESS) - if err != unix.EOPNOTSUPP { - t.Errorf("Faccessat: unexpected error: %v, want EOPNOTSUPP", err) - } - - err = os.Symlink("file1", "symlink1") - if err != nil { - t.Fatal(err) - } - - err = unix.Faccessat(unix.AT_FDCWD, "symlink1", unix.O_RDONLY, unix.AT_SYMLINK_NOFOLLOW) - if err != unix.EOPNOTSUPP { - t.Errorf("Faccessat: unexpected error: %v, want EOPNOTSUPP", err) - } -} diff --git a/vendor/golang.org/x/sys/unix/syscall_solaris_test.go b/vendor/golang.org/x/sys/unix/syscall_solaris_test.go deleted file mode 100644 index 57dba8824..000000000 --- a/vendor/golang.org/x/sys/unix/syscall_solaris_test.go +++ /dev/null @@ -1,55 +0,0 @@ -// Copyright 2017 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build solaris - -package unix_test - -import ( - "os/exec" - "testing" - "time" - - "golang.org/x/sys/unix" -) - -func TestSelect(t *testing.T) { - err := unix.Select(0, nil, nil, nil, &unix.Timeval{Sec: 0, Usec: 0}) - if err != nil { - t.Fatalf("Select: %v", err) - } - - dur := 150 * time.Millisecond - tv := unix.NsecToTimeval(int64(dur)) - start := time.Now() - err = unix.Select(0, nil, nil, nil, &tv) - took := time.Since(start) - if err != nil { - t.Fatalf("Select: %v", err) - } - - if took < dur { - t.Errorf("Select: timeout should have been at least %v, got %v", dur, took) - } -} - -func TestStatvfs(t *testing.T) { - if err := unix.Statvfs("", nil); err == nil { - t.Fatal(`Statvfs("") expected failure`) - } - - statvfs := unix.Statvfs_t{} - if err := unix.Statvfs("/", &statvfs); err != nil { - t.Errorf(`Statvfs("/") failed: %v`, err) - } - - if t.Failed() { - mount, err := exec.Command("mount").CombinedOutput() - if err != nil { - t.Logf("mount: %v\n%s", err, mount) - } else { - t.Logf("mount: %s", mount) - } - } -} diff --git a/vendor/golang.org/x/sys/unix/syscall_test.go b/vendor/golang.org/x/sys/unix/syscall_test.go deleted file mode 100644 index a8eef7cf8..000000000 --- a/vendor/golang.org/x/sys/unix/syscall_test.go +++ /dev/null @@ -1,60 +0,0 @@ -// Copyright 2013 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build darwin dragonfly freebsd linux netbsd openbsd solaris - -package unix_test - -import ( - "fmt" - "testing" - - "golang.org/x/sys/unix" -) - -func testSetGetenv(t *testing.T, key, value string) { - err := unix.Setenv(key, value) - if err != nil { - t.Fatalf("Setenv failed to set %q: %v", value, err) - } - newvalue, found := unix.Getenv(key) - if !found { - t.Fatalf("Getenv failed to find %v variable (want value %q)", key, value) - } - if newvalue != value { - t.Fatalf("Getenv(%v) = %q; want %q", key, newvalue, value) - } -} - -func TestEnv(t *testing.T) { - testSetGetenv(t, "TESTENV", "AVALUE") - // make sure TESTENV gets set to "", not deleted - testSetGetenv(t, "TESTENV", "") -} - -func TestItoa(t *testing.T) { - // Make most negative integer: 0x8000... - i := 1 - for i<<1 != 0 { - i <<= 1 - } - if i >= 0 { - t.Fatal("bad math") - } - s := unix.Itoa(i) - f := fmt.Sprint(i) - if s != f { - t.Fatalf("itoa(%d) = %s, want %s", i, s, f) - } -} - -func TestUname(t *testing.T) { - var utsname unix.Utsname - err := unix.Uname(&utsname) - if err != nil { - t.Fatalf("Uname: %v", err) - } - - t.Logf("OS: %s/%s %s", utsname.Sysname[:], utsname.Machine[:], utsname.Release[:]) -} diff --git a/vendor/golang.org/x/sys/unix/syscall_unix_test.go b/vendor/golang.org/x/sys/unix/syscall_unix_test.go deleted file mode 100644 index d694990db..000000000 --- a/vendor/golang.org/x/sys/unix/syscall_unix_test.go +++ /dev/null @@ -1,657 +0,0 @@ -// Copyright 2013 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build darwin dragonfly freebsd linux netbsd openbsd solaris - -package unix_test - -import ( - "flag" - "fmt" - "io/ioutil" - "net" - "os" - "os/exec" - "path/filepath" - "runtime" - "syscall" - "testing" - "time" - - "golang.org/x/sys/unix" -) - -// Tests that below functions, structures and constants are consistent -// on all Unix-like systems. -func _() { - // program scheduling priority functions and constants - var ( - _ func(int, int, int) error = unix.Setpriority - _ func(int, int) (int, error) = unix.Getpriority - ) - const ( - _ int = unix.PRIO_USER - _ int = unix.PRIO_PROCESS - _ int = unix.PRIO_PGRP - ) - - // termios constants - const ( - _ int = unix.TCIFLUSH - _ int = unix.TCIOFLUSH - _ int = unix.TCOFLUSH - ) - - // fcntl file locking structure and constants - var ( - _ = unix.Flock_t{ - Type: int16(0), - Whence: int16(0), - Start: int64(0), - Len: int64(0), - Pid: int32(0), - } - ) - const ( - _ = unix.F_GETLK - _ = unix.F_SETLK - _ = unix.F_SETLKW - ) -} - -func TestErrnoSignalName(t *testing.T) { - testErrors := []struct { - num syscall.Errno - name string - }{ - {syscall.EPERM, "EPERM"}, - {syscall.EINVAL, "EINVAL"}, - {syscall.ENOENT, "ENOENT"}, - } - - for _, te := range testErrors { - t.Run(fmt.Sprintf("%d/%s", te.num, te.name), func(t *testing.T) { - e := unix.ErrnoName(te.num) - if e != te.name { - t.Errorf("ErrnoName(%d) returned %s, want %s", te.num, e, te.name) - } - }) - } - - testSignals := []struct { - num syscall.Signal - name string - }{ - {syscall.SIGHUP, "SIGHUP"}, - {syscall.SIGPIPE, "SIGPIPE"}, - {syscall.SIGSEGV, "SIGSEGV"}, - } - - for _, ts := range testSignals { - t.Run(fmt.Sprintf("%d/%s", ts.num, ts.name), func(t *testing.T) { - s := unix.SignalName(ts.num) - if s != ts.name { - t.Errorf("SignalName(%d) returned %s, want %s", ts.num, s, ts.name) - } - }) - } -} - -func TestFcntlInt(t *testing.T) { - t.Parallel() - file, err := ioutil.TempFile("", "TestFnctlInt") - if err != nil { - t.Fatal(err) - } - defer os.Remove(file.Name()) - defer file.Close() - f := file.Fd() - flags, err := unix.FcntlInt(f, unix.F_GETFD, 0) - if err != nil { - t.Fatal(err) - } - if flags&unix.FD_CLOEXEC == 0 { - t.Errorf("flags %#x do not include FD_CLOEXEC", flags) - } -} - -// TestFcntlFlock tests whether the file locking structure matches -// the calling convention of each kernel. -func TestFcntlFlock(t *testing.T) { - name := filepath.Join(os.TempDir(), "TestFcntlFlock") - fd, err := unix.Open(name, unix.O_CREAT|unix.O_RDWR|unix.O_CLOEXEC, 0) - if err != nil { - t.Fatalf("Open failed: %v", err) - } - defer unix.Unlink(name) - defer unix.Close(fd) - flock := unix.Flock_t{ - Type: unix.F_RDLCK, - Start: 0, Len: 0, Whence: 1, - } - if err := unix.FcntlFlock(uintptr(fd), unix.F_GETLK, &flock); err != nil { - t.Fatalf("FcntlFlock failed: %v", err) - } -} - -// TestPassFD tests passing a file descriptor over a Unix socket. -// -// This test involved both a parent and child process. The parent -// process is invoked as a normal test, with "go test", which then -// runs the child process by running the current test binary with args -// "-test.run=^TestPassFD$" and an environment variable used to signal -// that the test should become the child process instead. -func TestPassFD(t *testing.T) { - if runtime.GOOS == "darwin" && (runtime.GOARCH == "arm" || runtime.GOARCH == "arm64") { - t.Skip("cannot exec subprocess on iOS, skipping test") - } - - if os.Getenv("GO_WANT_HELPER_PROCESS") == "1" { - passFDChild() - return - } - - tempDir, err := ioutil.TempDir("", "TestPassFD") - if err != nil { - t.Fatal(err) - } - defer os.RemoveAll(tempDir) - - fds, err := unix.Socketpair(unix.AF_LOCAL, unix.SOCK_STREAM, 0) - if err != nil { - t.Fatalf("Socketpair: %v", err) - } - defer unix.Close(fds[0]) - defer unix.Close(fds[1]) - writeFile := os.NewFile(uintptr(fds[0]), "child-writes") - readFile := os.NewFile(uintptr(fds[1]), "parent-reads") - defer writeFile.Close() - defer readFile.Close() - - cmd := exec.Command(os.Args[0], "-test.run=^TestPassFD$", "--", tempDir) - cmd.Env = []string{"GO_WANT_HELPER_PROCESS=1"} - if lp := os.Getenv("LD_LIBRARY_PATH"); lp != "" { - cmd.Env = append(cmd.Env, "LD_LIBRARY_PATH="+lp) - } - cmd.ExtraFiles = []*os.File{writeFile} - - out, err := cmd.CombinedOutput() - if len(out) > 0 || err != nil { - t.Fatalf("child process: %q, %v", out, err) - } - - c, err := net.FileConn(readFile) - if err != nil { - t.Fatalf("FileConn: %v", err) - } - defer c.Close() - - uc, ok := c.(*net.UnixConn) - if !ok { - t.Fatalf("unexpected FileConn type; expected UnixConn, got %T", c) - } - - buf := make([]byte, 32) // expect 1 byte - oob := make([]byte, 32) // expect 24 bytes - closeUnix := time.AfterFunc(5*time.Second, func() { - t.Logf("timeout reading from unix socket") - uc.Close() - }) - _, oobn, _, _, err := uc.ReadMsgUnix(buf, oob) - if err != nil { - t.Fatalf("ReadMsgUnix: %v", err) - } - closeUnix.Stop() - - scms, err := unix.ParseSocketControlMessage(oob[:oobn]) - if err != nil { - t.Fatalf("ParseSocketControlMessage: %v", err) - } - if len(scms) != 1 { - t.Fatalf("expected 1 SocketControlMessage; got scms = %#v", scms) - } - scm := scms[0] - gotFds, err := unix.ParseUnixRights(&scm) - if err != nil { - t.Fatalf("unix.ParseUnixRights: %v", err) - } - if len(gotFds) != 1 { - t.Fatalf("wanted 1 fd; got %#v", gotFds) - } - - f := os.NewFile(uintptr(gotFds[0]), "fd-from-child") - defer f.Close() - - got, err := ioutil.ReadAll(f) - want := "Hello from child process!\n" - if string(got) != want { - t.Errorf("child process ReadAll: %q, %v; want %q", got, err, want) - } -} - -// passFDChild is the child process used by TestPassFD. -func passFDChild() { - defer os.Exit(0) - - // Look for our fd. It should be fd 3, but we work around an fd leak - // bug here (http://golang.org/issue/2603) to let it be elsewhere. - var uc *net.UnixConn - for fd := uintptr(3); fd <= 10; fd++ { - f := os.NewFile(fd, "unix-conn") - var ok bool - netc, _ := net.FileConn(f) - uc, ok = netc.(*net.UnixConn) - if ok { - break - } - } - if uc == nil { - fmt.Println("failed to find unix fd") - return - } - - // Make a file f to send to our parent process on uc. - // We make it in tempDir, which our parent will clean up. - flag.Parse() - tempDir := flag.Arg(0) - f, err := ioutil.TempFile(tempDir, "") - if err != nil { - fmt.Printf("TempFile: %v", err) - return - } - - f.Write([]byte("Hello from child process!\n")) - f.Seek(0, 0) - - rights := unix.UnixRights(int(f.Fd())) - dummyByte := []byte("x") - n, oobn, err := uc.WriteMsgUnix(dummyByte, rights, nil) - if err != nil { - fmt.Printf("WriteMsgUnix: %v", err) - return - } - if n != 1 || oobn != len(rights) { - fmt.Printf("WriteMsgUnix = %d, %d; want 1, %d", n, oobn, len(rights)) - return - } -} - -// TestUnixRightsRoundtrip tests that UnixRights, ParseSocketControlMessage, -// and ParseUnixRights are able to successfully round-trip lists of file descriptors. -func TestUnixRightsRoundtrip(t *testing.T) { - testCases := [...][][]int{ - {{42}}, - {{1, 2}}, - {{3, 4, 5}}, - {{}}, - {{1, 2}, {3, 4, 5}, {}, {7}}, - } - for _, testCase := range testCases { - b := []byte{} - var n int - for _, fds := range testCase { - // Last assignment to n wins - n = len(b) + unix.CmsgLen(4*len(fds)) - b = append(b, unix.UnixRights(fds...)...) - } - // Truncate b - b = b[:n] - - scms, err := unix.ParseSocketControlMessage(b) - if err != nil { - t.Fatalf("ParseSocketControlMessage: %v", err) - } - if len(scms) != len(testCase) { - t.Fatalf("expected %v SocketControlMessage; got scms = %#v", len(testCase), scms) - } - for i, scm := range scms { - gotFds, err := unix.ParseUnixRights(&scm) - if err != nil { - t.Fatalf("ParseUnixRights: %v", err) - } - wantFds := testCase[i] - if len(gotFds) != len(wantFds) { - t.Fatalf("expected %v fds, got %#v", len(wantFds), gotFds) - } - for j, fd := range gotFds { - if fd != wantFds[j] { - t.Fatalf("expected fd %v, got %v", wantFds[j], fd) - } - } - } - } -} - -func TestRlimit(t *testing.T) { - var rlimit, zero unix.Rlimit - err := unix.Getrlimit(unix.RLIMIT_NOFILE, &rlimit) - if err != nil { - t.Fatalf("Getrlimit: save failed: %v", err) - } - if zero == rlimit { - t.Fatalf("Getrlimit: save failed: got zero value %#v", rlimit) - } - set := rlimit - set.Cur = set.Max - 1 - err = unix.Setrlimit(unix.RLIMIT_NOFILE, &set) - if err != nil { - t.Fatalf("Setrlimit: set failed: %#v %v", set, err) - } - var get unix.Rlimit - err = unix.Getrlimit(unix.RLIMIT_NOFILE, &get) - if err != nil { - t.Fatalf("Getrlimit: get failed: %v", err) - } - set = rlimit - set.Cur = set.Max - 1 - if set != get { - // Seems like Darwin requires some privilege to - // increase the soft limit of rlimit sandbox, though - // Setrlimit never reports an error. - switch runtime.GOOS { - case "darwin": - default: - t.Fatalf("Rlimit: change failed: wanted %#v got %#v", set, get) - } - } - err = unix.Setrlimit(unix.RLIMIT_NOFILE, &rlimit) - if err != nil { - t.Fatalf("Setrlimit: restore failed: %#v %v", rlimit, err) - } -} - -func TestSeekFailure(t *testing.T) { - _, err := unix.Seek(-1, 0, 0) - if err == nil { - t.Fatalf("Seek(-1, 0, 0) did not fail") - } - str := err.Error() // used to crash on Linux - t.Logf("Seek: %v", str) - if str == "" { - t.Fatalf("Seek(-1, 0, 0) return error with empty message") - } -} - -func TestDup(t *testing.T) { - file, err := ioutil.TempFile("", "TestDup") - if err != nil { - t.Fatalf("Tempfile failed: %v", err) - } - defer os.Remove(file.Name()) - defer file.Close() - f := int(file.Fd()) - - newFd, err := unix.Dup(f) - if err != nil { - t.Fatalf("Dup: %v", err) - } - - err = unix.Dup2(newFd, newFd+1) - if err != nil { - t.Fatalf("Dup2: %v", err) - } - - b1 := []byte("Test123") - b2 := make([]byte, 7) - _, err = unix.Write(newFd+1, b1) - if err != nil { - t.Fatalf("Write to dup2 fd failed: %v", err) - } - _, err = unix.Seek(f, 0, 0) - if err != nil { - t.Fatalf("Seek failed: %v", err) - } - _, err = unix.Read(f, b2) - if err != nil { - t.Fatalf("Read back failed: %v", err) - } - if string(b1) != string(b2) { - t.Errorf("Dup: stdout write not in file, expected %v, got %v", string(b1), string(b2)) - } -} - -func TestPoll(t *testing.T) { - if runtime.GOOS == "android" || - (runtime.GOOS == "darwin" && (runtime.GOARCH == "arm" || runtime.GOARCH == "arm64")) { - t.Skip("mkfifo syscall is not available on android and iOS, skipping test") - } - - f, cleanup := mktmpfifo(t) - defer cleanup() - - const timeout = 100 - - ok := make(chan bool, 1) - go func() { - select { - case <-time.After(10 * timeout * time.Millisecond): - t.Errorf("Poll: failed to timeout after %d milliseconds", 10*timeout) - case <-ok: - } - }() - - fds := []unix.PollFd{{Fd: int32(f.Fd()), Events: unix.POLLIN}} - n, err := unix.Poll(fds, timeout) - ok <- true - if err != nil { - t.Errorf("Poll: unexpected error: %v", err) - return - } - if n != 0 { - t.Errorf("Poll: wrong number of events: got %v, expected %v", n, 0) - return - } -} - -func TestGetwd(t *testing.T) { - fd, err := os.Open(".") - if err != nil { - t.Fatalf("Open .: %s", err) - } - defer fd.Close() - // These are chosen carefully not to be symlinks on a Mac - // (unlike, say, /var, /etc) - dirs := []string{"/", "/usr/bin"} - switch runtime.GOOS { - case "android": - dirs = []string{"/", "/system/bin"} - case "darwin": - switch runtime.GOARCH { - case "arm", "arm64": - d1, err := ioutil.TempDir("", "d1") - if err != nil { - t.Fatalf("TempDir: %v", err) - } - d2, err := ioutil.TempDir("", "d2") - if err != nil { - t.Fatalf("TempDir: %v", err) - } - dirs = []string{d1, d2} - } - } - oldwd := os.Getenv("PWD") - for _, d := range dirs { - err = os.Chdir(d) - if err != nil { - t.Fatalf("Chdir: %v", err) - } - pwd, err := unix.Getwd() - if err != nil { - t.Fatalf("Getwd in %s: %s", d, err) - } - os.Setenv("PWD", oldwd) - err = fd.Chdir() - if err != nil { - // We changed the current directory and cannot go back. - // Don't let the tests continue; they'll scribble - // all over some other directory. - fmt.Fprintf(os.Stderr, "fchdir back to dot failed: %s\n", err) - os.Exit(1) - } - if pwd != d { - t.Fatalf("Getwd returned %q want %q", pwd, d) - } - } -} - -func TestFstatat(t *testing.T) { - defer chtmpdir(t)() - - touch(t, "file1") - - var st1 unix.Stat_t - err := unix.Stat("file1", &st1) - if err != nil { - t.Fatalf("Stat: %v", err) - } - - var st2 unix.Stat_t - err = unix.Fstatat(unix.AT_FDCWD, "file1", &st2, 0) - if err != nil { - t.Fatalf("Fstatat: %v", err) - } - - if st1 != st2 { - t.Errorf("Fstatat: returned stat does not match Stat") - } - - err = os.Symlink("file1", "symlink1") - if err != nil { - t.Fatal(err) - } - - err = unix.Lstat("symlink1", &st1) - if err != nil { - t.Fatalf("Lstat: %v", err) - } - - err = unix.Fstatat(unix.AT_FDCWD, "symlink1", &st2, unix.AT_SYMLINK_NOFOLLOW) - if err != nil { - t.Fatalf("Fstatat: %v", err) - } - - if st1 != st2 { - t.Errorf("Fstatat: returned stat does not match Lstat") - } -} - -func TestFchmodat(t *testing.T) { - defer chtmpdir(t)() - - touch(t, "file1") - err := os.Symlink("file1", "symlink1") - if err != nil { - t.Fatal(err) - } - - mode := os.FileMode(0444) - err = unix.Fchmodat(unix.AT_FDCWD, "symlink1", uint32(mode), 0) - if err != nil { - t.Fatalf("Fchmodat: unexpected error: %v", err) - } - - fi, err := os.Stat("file1") - if err != nil { - t.Fatal(err) - } - - if fi.Mode() != mode { - t.Errorf("Fchmodat: failed to change file mode: expected %v, got %v", mode, fi.Mode()) - } - - mode = os.FileMode(0644) - didChmodSymlink := true - err = unix.Fchmodat(unix.AT_FDCWD, "symlink1", uint32(mode), unix.AT_SYMLINK_NOFOLLOW) - if err != nil { - if (runtime.GOOS == "android" || runtime.GOOS == "linux" || runtime.GOOS == "solaris") && err == unix.EOPNOTSUPP { - // Linux and Illumos don't support flags != 0 - didChmodSymlink = false - } else { - t.Fatalf("Fchmodat: unexpected error: %v", err) - } - } - - if !didChmodSymlink { - // Didn't change mode of the symlink. On Linux, the permissions - // of a symbolic link are always 0777 according to symlink(7) - mode = os.FileMode(0777) - } - - var st unix.Stat_t - err = unix.Lstat("symlink1", &st) - if err != nil { - t.Fatal(err) - } - - got := os.FileMode(st.Mode & 0777) - if got != mode { - t.Errorf("Fchmodat: failed to change symlink mode: expected %v, got %v", mode, got) - } -} - -func TestMkdev(t *testing.T) { - major := uint32(42) - minor := uint32(7) - dev := unix.Mkdev(major, minor) - - if unix.Major(dev) != major { - t.Errorf("Major(%#x) == %d, want %d", dev, unix.Major(dev), major) - } - if unix.Minor(dev) != minor { - t.Errorf("Minor(%#x) == %d, want %d", dev, unix.Minor(dev), minor) - } -} - -// mktmpfifo creates a temporary FIFO and provides a cleanup function. -func mktmpfifo(t *testing.T) (*os.File, func()) { - err := unix.Mkfifo("fifo", 0666) - if err != nil { - t.Fatalf("mktmpfifo: failed to create FIFO: %v", err) - } - - f, err := os.OpenFile("fifo", os.O_RDWR, 0666) - if err != nil { - os.Remove("fifo") - t.Fatalf("mktmpfifo: failed to open FIFO: %v", err) - } - - return f, func() { - f.Close() - os.Remove("fifo") - } -} - -// utilities taken from os/os_test.go - -func touch(t *testing.T, name string) { - f, err := os.Create(name) - if err != nil { - t.Fatal(err) - } - if err := f.Close(); err != nil { - t.Fatal(err) - } -} - -// chtmpdir changes the working directory to a new temporary directory and -// provides a cleanup function. Used when PWD is read-only. -func chtmpdir(t *testing.T) func() { - oldwd, err := os.Getwd() - if err != nil { - t.Fatalf("chtmpdir: %v", err) - } - d, err := ioutil.TempDir("", "test") - if err != nil { - t.Fatalf("chtmpdir: %v", err) - } - if err := os.Chdir(d); err != nil { - t.Fatalf("chtmpdir: %v", err) - } - return func() { - if err := os.Chdir(oldwd); err != nil { - t.Fatalf("chtmpdir: %v", err) - } - os.RemoveAll(d) - } -} diff --git a/vendor/golang.org/x/sys/unix/timestruct_test.go b/vendor/golang.org/x/sys/unix/timestruct_test.go deleted file mode 100644 index 4215f46d1..000000000 --- a/vendor/golang.org/x/sys/unix/timestruct_test.go +++ /dev/null @@ -1,54 +0,0 @@ -// Copyright 2017 The Go Authors. All right reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build darwin dragonfly freebsd linux netbsd openbsd solaris - -package unix_test - -import ( - "testing" - "time" - "unsafe" - - "golang.org/x/sys/unix" -) - -func TestTimeToTimespec(t *testing.T) { - timeTests := []struct { - time time.Time - valid bool - }{ - {time.Unix(0, 0), true}, - {time.Date(2009, time.November, 10, 23, 0, 0, 0, time.UTC), true}, - {time.Date(2262, time.December, 31, 23, 0, 0, 0, time.UTC), false}, - {time.Unix(0x7FFFFFFF, 0), true}, - {time.Unix(0x80000000, 0), false}, - {time.Unix(0x7FFFFFFF, 1000000000), false}, - {time.Unix(0x7FFFFFFF, 999999999), true}, - {time.Unix(-0x80000000, 0), true}, - {time.Unix(-0x80000001, 0), false}, - {time.Date(2038, time.January, 19, 3, 14, 7, 0, time.UTC), true}, - {time.Date(2038, time.January, 19, 3, 14, 8, 0, time.UTC), false}, - {time.Date(1901, time.December, 13, 20, 45, 52, 0, time.UTC), true}, - {time.Date(1901, time.December, 13, 20, 45, 51, 0, time.UTC), false}, - } - - // Currently all targets have either int32 or int64 for Timespec.Sec. - // If there were a new target with unsigned or floating point type for - // it, this test must be adjusted. - have64BitTime := (unsafe.Sizeof(unix.Timespec{}.Sec) == 8) - for _, tt := range timeTests { - ts, err := unix.TimeToTimespec(tt.time) - tt.valid = tt.valid || have64BitTime - if tt.valid && err != nil { - t.Errorf("TimeToTimespec(%v): %v", tt.time, err) - } - if err == nil { - tstime := time.Unix(int64(ts.Sec), int64(ts.Nsec)) - if !tstime.Equal(tt.time) { - t.Errorf("TimeToTimespec(%v) is the time %v", tt.time, tstime) - } - } - } -} diff --git a/vendor/golang.org/x/sys/unix/types_dragonfly.go b/vendor/golang.org/x/sys/unix/types_dragonfly.go deleted file mode 100644 index 0c6330483..000000000 --- a/vendor/golang.org/x/sys/unix/types_dragonfly.go +++ /dev/null @@ -1,280 +0,0 @@ -// Copyright 2009 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build ignore - -/* -Input to cgo -godefs. See README.md -*/ - -// +godefs map struct_in_addr [4]byte /* in_addr */ -// +godefs map struct_in6_addr [16]byte /* in6_addr */ - -package unix - -/* -#define KERNEL -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include - -enum { - sizeofPtr = sizeof(void*), -}; - -union sockaddr_all { - struct sockaddr s1; // this one gets used for fields - struct sockaddr_in s2; // these pad it out - struct sockaddr_in6 s3; - struct sockaddr_un s4; - struct sockaddr_dl s5; -}; - -struct sockaddr_any { - struct sockaddr addr; - char pad[sizeof(union sockaddr_all) - sizeof(struct sockaddr)]; -}; - -*/ -import "C" - -// Machine characteristics; for internal use. - -const ( - sizeofPtr = C.sizeofPtr - sizeofShort = C.sizeof_short - sizeofInt = C.sizeof_int - sizeofLong = C.sizeof_long - sizeofLongLong = C.sizeof_longlong -) - -// Basic types - -type ( - _C_short C.short - _C_int C.int - _C_long C.long - _C_long_long C.longlong -) - -// Time - -type Timespec C.struct_timespec - -type Timeval C.struct_timeval - -// Processes - -type Rusage C.struct_rusage - -type Rlimit C.struct_rlimit - -type _Gid_t C.gid_t - -// Files - -const ( // Directory mode bits - S_IFMT = C.S_IFMT - S_IFIFO = C.S_IFIFO - S_IFCHR = C.S_IFCHR - S_IFDIR = C.S_IFDIR - S_IFBLK = C.S_IFBLK - S_IFREG = C.S_IFREG - S_IFLNK = C.S_IFLNK - S_IFSOCK = C.S_IFSOCK - S_ISUID = C.S_ISUID - S_ISGID = C.S_ISGID - S_ISVTX = C.S_ISVTX - S_IRUSR = C.S_IRUSR - S_IWUSR = C.S_IWUSR - S_IXUSR = C.S_IXUSR -) - -type Stat_t C.struct_stat - -type Statfs_t C.struct_statfs - -type Flock_t C.struct_flock - -type Dirent C.struct_dirent - -type Fsid C.struct_fsid - -// File system limits - -const ( - PathMax = C.PATH_MAX -) - -// Sockets - -type RawSockaddrInet4 C.struct_sockaddr_in - -type RawSockaddrInet6 C.struct_sockaddr_in6 - -type RawSockaddrUnix C.struct_sockaddr_un - -type RawSockaddrDatalink C.struct_sockaddr_dl - -type RawSockaddr C.struct_sockaddr - -type RawSockaddrAny C.struct_sockaddr_any - -type _Socklen C.socklen_t - -type Linger C.struct_linger - -type Iovec C.struct_iovec - -type IPMreq C.struct_ip_mreq - -type IPv6Mreq C.struct_ipv6_mreq - -type Msghdr C.struct_msghdr - -type Cmsghdr C.struct_cmsghdr - -type Inet6Pktinfo C.struct_in6_pktinfo - -type IPv6MTUInfo C.struct_ip6_mtuinfo - -type ICMPv6Filter C.struct_icmp6_filter - -const ( - SizeofSockaddrInet4 = C.sizeof_struct_sockaddr_in - SizeofSockaddrInet6 = C.sizeof_struct_sockaddr_in6 - SizeofSockaddrAny = C.sizeof_struct_sockaddr_any - SizeofSockaddrUnix = C.sizeof_struct_sockaddr_un - SizeofSockaddrDatalink = C.sizeof_struct_sockaddr_dl - SizeofLinger = C.sizeof_struct_linger - SizeofIPMreq = C.sizeof_struct_ip_mreq - SizeofIPv6Mreq = C.sizeof_struct_ipv6_mreq - SizeofMsghdr = C.sizeof_struct_msghdr - SizeofCmsghdr = C.sizeof_struct_cmsghdr - SizeofInet6Pktinfo = C.sizeof_struct_in6_pktinfo - SizeofIPv6MTUInfo = C.sizeof_struct_ip6_mtuinfo - SizeofICMPv6Filter = C.sizeof_struct_icmp6_filter -) - -// Ptrace requests - -const ( - PTRACE_TRACEME = C.PT_TRACE_ME - PTRACE_CONT = C.PT_CONTINUE - PTRACE_KILL = C.PT_KILL -) - -// Events (kqueue, kevent) - -type Kevent_t C.struct_kevent - -// Select - -type FdSet C.fd_set - -// Routing and interface messages - -const ( - SizeofIfMsghdr = C.sizeof_struct_if_msghdr - SizeofIfData = C.sizeof_struct_if_data - SizeofIfaMsghdr = C.sizeof_struct_ifa_msghdr - SizeofIfmaMsghdr = C.sizeof_struct_ifma_msghdr - SizeofIfAnnounceMsghdr = C.sizeof_struct_if_announcemsghdr - SizeofRtMsghdr = C.sizeof_struct_rt_msghdr - SizeofRtMetrics = C.sizeof_struct_rt_metrics -) - -type IfMsghdr C.struct_if_msghdr - -type IfData C.struct_if_data - -type IfaMsghdr C.struct_ifa_msghdr - -type IfmaMsghdr C.struct_ifma_msghdr - -type IfAnnounceMsghdr C.struct_if_announcemsghdr - -type RtMsghdr C.struct_rt_msghdr - -type RtMetrics C.struct_rt_metrics - -// Berkeley packet filter - -const ( - SizeofBpfVersion = C.sizeof_struct_bpf_version - SizeofBpfStat = C.sizeof_struct_bpf_stat - SizeofBpfProgram = C.sizeof_struct_bpf_program - SizeofBpfInsn = C.sizeof_struct_bpf_insn - SizeofBpfHdr = C.sizeof_struct_bpf_hdr -) - -type BpfVersion C.struct_bpf_version - -type BpfStat C.struct_bpf_stat - -type BpfProgram C.struct_bpf_program - -type BpfInsn C.struct_bpf_insn - -type BpfHdr C.struct_bpf_hdr - -// Terminal handling - -type Termios C.struct_termios - -type Winsize C.struct_winsize - -// fchmodat-like syscalls. - -const ( - AT_FDCWD = C.AT_FDCWD - AT_SYMLINK_NOFOLLOW = C.AT_SYMLINK_NOFOLLOW -) - -// poll - -type PollFd C.struct_pollfd - -const ( - POLLERR = C.POLLERR - POLLHUP = C.POLLHUP - POLLIN = C.POLLIN - POLLNVAL = C.POLLNVAL - POLLOUT = C.POLLOUT - POLLPRI = C.POLLPRI - POLLRDBAND = C.POLLRDBAND - POLLRDNORM = C.POLLRDNORM - POLLWRBAND = C.POLLWRBAND - POLLWRNORM = C.POLLWRNORM -) - -// Uname - -type Utsname C.struct_utsname diff --git a/vendor/golang.org/x/sys/unix/types_freebsd.go b/vendor/golang.org/x/sys/unix/types_freebsd.go deleted file mode 100644 index 4eb02cd4d..000000000 --- a/vendor/golang.org/x/sys/unix/types_freebsd.go +++ /dev/null @@ -1,402 +0,0 @@ -// Copyright 2009 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build ignore - -/* -Input to cgo -godefs. See README.md -*/ - -// +godefs map struct_in_addr [4]byte /* in_addr */ -// +godefs map struct_in6_addr [16]byte /* in6_addr */ - -package unix - -/* -#define KERNEL -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include - -enum { - sizeofPtr = sizeof(void*), -}; - -union sockaddr_all { - struct sockaddr s1; // this one gets used for fields - struct sockaddr_in s2; // these pad it out - struct sockaddr_in6 s3; - struct sockaddr_un s4; - struct sockaddr_dl s5; -}; - -struct sockaddr_any { - struct sockaddr addr; - char pad[sizeof(union sockaddr_all) - sizeof(struct sockaddr)]; -}; - -// This structure is a duplicate of stat on FreeBSD 8-STABLE. -// See /usr/include/sys/stat.h. -struct stat8 { -#undef st_atimespec st_atim -#undef st_mtimespec st_mtim -#undef st_ctimespec st_ctim -#undef st_birthtimespec st_birthtim - __dev_t st_dev; - ino_t st_ino; - mode_t st_mode; - nlink_t st_nlink; - uid_t st_uid; - gid_t st_gid; - __dev_t st_rdev; -#if __BSD_VISIBLE - struct timespec st_atimespec; - struct timespec st_mtimespec; - struct timespec st_ctimespec; -#else - time_t st_atime; - long __st_atimensec; - time_t st_mtime; - long __st_mtimensec; - time_t st_ctime; - long __st_ctimensec; -#endif - off_t st_size; - blkcnt_t st_blocks; - blksize_t st_blksize; - fflags_t st_flags; - __uint32_t st_gen; - __int32_t st_lspare; -#if __BSD_VISIBLE - struct timespec st_birthtimespec; - unsigned int :(8 / 2) * (16 - (int)sizeof(struct timespec)); - unsigned int :(8 / 2) * (16 - (int)sizeof(struct timespec)); -#else - time_t st_birthtime; - long st_birthtimensec; - unsigned int :(8 / 2) * (16 - (int)sizeof(struct __timespec)); - unsigned int :(8 / 2) * (16 - (int)sizeof(struct __timespec)); -#endif -}; - -// This structure is a duplicate of if_data on FreeBSD 8-STABLE. -// See /usr/include/net/if.h. -struct if_data8 { - u_char ifi_type; - u_char ifi_physical; - u_char ifi_addrlen; - u_char ifi_hdrlen; - u_char ifi_link_state; - u_char ifi_spare_char1; - u_char ifi_spare_char2; - u_char ifi_datalen; - u_long ifi_mtu; - u_long ifi_metric; - u_long ifi_baudrate; - u_long ifi_ipackets; - u_long ifi_ierrors; - u_long ifi_opackets; - u_long ifi_oerrors; - u_long ifi_collisions; - u_long ifi_ibytes; - u_long ifi_obytes; - u_long ifi_imcasts; - u_long ifi_omcasts; - u_long ifi_iqdrops; - u_long ifi_noproto; - u_long ifi_hwassist; -// FIXME: these are now unions, so maybe need to change definitions? -#undef ifi_epoch - time_t ifi_epoch; -#undef ifi_lastchange - struct timeval ifi_lastchange; -}; - -// This structure is a duplicate of if_msghdr on FreeBSD 8-STABLE. -// See /usr/include/net/if.h. -struct if_msghdr8 { - u_short ifm_msglen; - u_char ifm_version; - u_char ifm_type; - int ifm_addrs; - int ifm_flags; - u_short ifm_index; - struct if_data8 ifm_data; -}; -*/ -import "C" - -// Machine characteristics; for internal use. - -const ( - sizeofPtr = C.sizeofPtr - sizeofShort = C.sizeof_short - sizeofInt = C.sizeof_int - sizeofLong = C.sizeof_long - sizeofLongLong = C.sizeof_longlong -) - -// Basic types - -type ( - _C_short C.short - _C_int C.int - _C_long C.long - _C_long_long C.longlong -) - -// Time - -type Timespec C.struct_timespec - -type Timeval C.struct_timeval - -// Processes - -type Rusage C.struct_rusage - -type Rlimit C.struct_rlimit - -type _Gid_t C.gid_t - -// Files - -const ( // Directory mode bits - S_IFMT = C.S_IFMT - S_IFIFO = C.S_IFIFO - S_IFCHR = C.S_IFCHR - S_IFDIR = C.S_IFDIR - S_IFBLK = C.S_IFBLK - S_IFREG = C.S_IFREG - S_IFLNK = C.S_IFLNK - S_IFSOCK = C.S_IFSOCK - S_ISUID = C.S_ISUID - S_ISGID = C.S_ISGID - S_ISVTX = C.S_ISVTX - S_IRUSR = C.S_IRUSR - S_IWUSR = C.S_IWUSR - S_IXUSR = C.S_IXUSR -) - -type Stat_t C.struct_stat8 - -type Statfs_t C.struct_statfs - -type Flock_t C.struct_flock - -type Dirent C.struct_dirent - -type Fsid C.struct_fsid - -// File system limits - -const ( - PathMax = C.PATH_MAX -) - -// Advice to Fadvise - -const ( - FADV_NORMAL = C.POSIX_FADV_NORMAL - FADV_RANDOM = C.POSIX_FADV_RANDOM - FADV_SEQUENTIAL = C.POSIX_FADV_SEQUENTIAL - FADV_WILLNEED = C.POSIX_FADV_WILLNEED - FADV_DONTNEED = C.POSIX_FADV_DONTNEED - FADV_NOREUSE = C.POSIX_FADV_NOREUSE -) - -// Sockets - -type RawSockaddrInet4 C.struct_sockaddr_in - -type RawSockaddrInet6 C.struct_sockaddr_in6 - -type RawSockaddrUnix C.struct_sockaddr_un - -type RawSockaddrDatalink C.struct_sockaddr_dl - -type RawSockaddr C.struct_sockaddr - -type RawSockaddrAny C.struct_sockaddr_any - -type _Socklen C.socklen_t - -type Linger C.struct_linger - -type Iovec C.struct_iovec - -type IPMreq C.struct_ip_mreq - -type IPMreqn C.struct_ip_mreqn - -type IPv6Mreq C.struct_ipv6_mreq - -type Msghdr C.struct_msghdr - -type Cmsghdr C.struct_cmsghdr - -type Inet6Pktinfo C.struct_in6_pktinfo - -type IPv6MTUInfo C.struct_ip6_mtuinfo - -type ICMPv6Filter C.struct_icmp6_filter - -const ( - SizeofSockaddrInet4 = C.sizeof_struct_sockaddr_in - SizeofSockaddrInet6 = C.sizeof_struct_sockaddr_in6 - SizeofSockaddrAny = C.sizeof_struct_sockaddr_any - SizeofSockaddrUnix = C.sizeof_struct_sockaddr_un - SizeofSockaddrDatalink = C.sizeof_struct_sockaddr_dl - SizeofLinger = C.sizeof_struct_linger - SizeofIPMreq = C.sizeof_struct_ip_mreq - SizeofIPMreqn = C.sizeof_struct_ip_mreqn - SizeofIPv6Mreq = C.sizeof_struct_ipv6_mreq - SizeofMsghdr = C.sizeof_struct_msghdr - SizeofCmsghdr = C.sizeof_struct_cmsghdr - SizeofInet6Pktinfo = C.sizeof_struct_in6_pktinfo - SizeofIPv6MTUInfo = C.sizeof_struct_ip6_mtuinfo - SizeofICMPv6Filter = C.sizeof_struct_icmp6_filter -) - -// Ptrace requests - -const ( - PTRACE_TRACEME = C.PT_TRACE_ME - PTRACE_CONT = C.PT_CONTINUE - PTRACE_KILL = C.PT_KILL -) - -// Events (kqueue, kevent) - -type Kevent_t C.struct_kevent - -// Select - -type FdSet C.fd_set - -// Routing and interface messages - -const ( - sizeofIfMsghdr = C.sizeof_struct_if_msghdr - SizeofIfMsghdr = C.sizeof_struct_if_msghdr8 - sizeofIfData = C.sizeof_struct_if_data - SizeofIfData = C.sizeof_struct_if_data8 - SizeofIfaMsghdr = C.sizeof_struct_ifa_msghdr - SizeofIfmaMsghdr = C.sizeof_struct_ifma_msghdr - SizeofIfAnnounceMsghdr = C.sizeof_struct_if_announcemsghdr - SizeofRtMsghdr = C.sizeof_struct_rt_msghdr - SizeofRtMetrics = C.sizeof_struct_rt_metrics -) - -type ifMsghdr C.struct_if_msghdr - -type IfMsghdr C.struct_if_msghdr8 - -type ifData C.struct_if_data - -type IfData C.struct_if_data8 - -type IfaMsghdr C.struct_ifa_msghdr - -type IfmaMsghdr C.struct_ifma_msghdr - -type IfAnnounceMsghdr C.struct_if_announcemsghdr - -type RtMsghdr C.struct_rt_msghdr - -type RtMetrics C.struct_rt_metrics - -// Berkeley packet filter - -const ( - SizeofBpfVersion = C.sizeof_struct_bpf_version - SizeofBpfStat = C.sizeof_struct_bpf_stat - SizeofBpfZbuf = C.sizeof_struct_bpf_zbuf - SizeofBpfProgram = C.sizeof_struct_bpf_program - SizeofBpfInsn = C.sizeof_struct_bpf_insn - SizeofBpfHdr = C.sizeof_struct_bpf_hdr - SizeofBpfZbufHeader = C.sizeof_struct_bpf_zbuf_header -) - -type BpfVersion C.struct_bpf_version - -type BpfStat C.struct_bpf_stat - -type BpfZbuf C.struct_bpf_zbuf - -type BpfProgram C.struct_bpf_program - -type BpfInsn C.struct_bpf_insn - -type BpfHdr C.struct_bpf_hdr - -type BpfZbufHeader C.struct_bpf_zbuf_header - -// Terminal handling - -type Termios C.struct_termios - -type Winsize C.struct_winsize - -// fchmodat-like syscalls. - -const ( - AT_FDCWD = C.AT_FDCWD - AT_REMOVEDIR = C.AT_REMOVEDIR - AT_SYMLINK_FOLLOW = C.AT_SYMLINK_FOLLOW - AT_SYMLINK_NOFOLLOW = C.AT_SYMLINK_NOFOLLOW -) - -// poll - -type PollFd C.struct_pollfd - -const ( - POLLERR = C.POLLERR - POLLHUP = C.POLLHUP - POLLIN = C.POLLIN - POLLINIGNEOF = C.POLLINIGNEOF - POLLNVAL = C.POLLNVAL - POLLOUT = C.POLLOUT - POLLPRI = C.POLLPRI - POLLRDBAND = C.POLLRDBAND - POLLRDNORM = C.POLLRDNORM - POLLWRBAND = C.POLLWRBAND - POLLWRNORM = C.POLLWRNORM -) - -// Capabilities - -type CapRights C.struct_cap_rights - -// Uname - -type Utsname C.struct_utsname diff --git a/vendor/golang.org/x/sys/unix/types_netbsd.go b/vendor/golang.org/x/sys/unix/types_netbsd.go deleted file mode 100644 index 1494aafcb..000000000 --- a/vendor/golang.org/x/sys/unix/types_netbsd.go +++ /dev/null @@ -1,281 +0,0 @@ -// Copyright 2009 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build ignore - -/* -Input to cgo -godefs. See README.md -*/ - -// +godefs map struct_in_addr [4]byte /* in_addr */ -// +godefs map struct_in6_addr [16]byte /* in6_addr */ - -package unix - -/* -#define KERNEL -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include - -enum { - sizeofPtr = sizeof(void*), -}; - -union sockaddr_all { - struct sockaddr s1; // this one gets used for fields - struct sockaddr_in s2; // these pad it out - struct sockaddr_in6 s3; - struct sockaddr_un s4; - struct sockaddr_dl s5; -}; - -struct sockaddr_any { - struct sockaddr addr; - char pad[sizeof(union sockaddr_all) - sizeof(struct sockaddr)]; -}; - -*/ -import "C" - -// Machine characteristics; for internal use. - -const ( - sizeofPtr = C.sizeofPtr - sizeofShort = C.sizeof_short - sizeofInt = C.sizeof_int - sizeofLong = C.sizeof_long - sizeofLongLong = C.sizeof_longlong -) - -// Basic types - -type ( - _C_short C.short - _C_int C.int - _C_long C.long - _C_long_long C.longlong -) - -// Time - -type Timespec C.struct_timespec - -type Timeval C.struct_timeval - -// Processes - -type Rusage C.struct_rusage - -type Rlimit C.struct_rlimit - -type _Gid_t C.gid_t - -// Files - -type Stat_t C.struct_stat - -type Statfs_t C.struct_statfs - -type Flock_t C.struct_flock - -type Dirent C.struct_dirent - -type Fsid C.fsid_t - -// File system limits - -const ( - PathMax = C.PATH_MAX -) - -// Advice to Fadvise - -const ( - FADV_NORMAL = C.POSIX_FADV_NORMAL - FADV_RANDOM = C.POSIX_FADV_RANDOM - FADV_SEQUENTIAL = C.POSIX_FADV_SEQUENTIAL - FADV_WILLNEED = C.POSIX_FADV_WILLNEED - FADV_DONTNEED = C.POSIX_FADV_DONTNEED - FADV_NOREUSE = C.POSIX_FADV_NOREUSE -) - -// Sockets - -type RawSockaddrInet4 C.struct_sockaddr_in - -type RawSockaddrInet6 C.struct_sockaddr_in6 - -type RawSockaddrUnix C.struct_sockaddr_un - -type RawSockaddrDatalink C.struct_sockaddr_dl - -type RawSockaddr C.struct_sockaddr - -type RawSockaddrAny C.struct_sockaddr_any - -type _Socklen C.socklen_t - -type Linger C.struct_linger - -type Iovec C.struct_iovec - -type IPMreq C.struct_ip_mreq - -type IPv6Mreq C.struct_ipv6_mreq - -type Msghdr C.struct_msghdr - -type Cmsghdr C.struct_cmsghdr - -type Inet6Pktinfo C.struct_in6_pktinfo - -type IPv6MTUInfo C.struct_ip6_mtuinfo - -type ICMPv6Filter C.struct_icmp6_filter - -const ( - SizeofSockaddrInet4 = C.sizeof_struct_sockaddr_in - SizeofSockaddrInet6 = C.sizeof_struct_sockaddr_in6 - SizeofSockaddrAny = C.sizeof_struct_sockaddr_any - SizeofSockaddrUnix = C.sizeof_struct_sockaddr_un - SizeofSockaddrDatalink = C.sizeof_struct_sockaddr_dl - SizeofLinger = C.sizeof_struct_linger - SizeofIPMreq = C.sizeof_struct_ip_mreq - SizeofIPv6Mreq = C.sizeof_struct_ipv6_mreq - SizeofMsghdr = C.sizeof_struct_msghdr - SizeofCmsghdr = C.sizeof_struct_cmsghdr - SizeofInet6Pktinfo = C.sizeof_struct_in6_pktinfo - SizeofIPv6MTUInfo = C.sizeof_struct_ip6_mtuinfo - SizeofICMPv6Filter = C.sizeof_struct_icmp6_filter -) - -// Ptrace requests - -const ( - PTRACE_TRACEME = C.PT_TRACE_ME - PTRACE_CONT = C.PT_CONTINUE - PTRACE_KILL = C.PT_KILL -) - -// Events (kqueue, kevent) - -type Kevent_t C.struct_kevent - -// Select - -type FdSet C.fd_set - -// Routing and interface messages - -const ( - SizeofIfMsghdr = C.sizeof_struct_if_msghdr - SizeofIfData = C.sizeof_struct_if_data - SizeofIfaMsghdr = C.sizeof_struct_ifa_msghdr - SizeofIfAnnounceMsghdr = C.sizeof_struct_if_announcemsghdr - SizeofRtMsghdr = C.sizeof_struct_rt_msghdr - SizeofRtMetrics = C.sizeof_struct_rt_metrics -) - -type IfMsghdr C.struct_if_msghdr - -type IfData C.struct_if_data - -type IfaMsghdr C.struct_ifa_msghdr - -type IfAnnounceMsghdr C.struct_if_announcemsghdr - -type RtMsghdr C.struct_rt_msghdr - -type RtMetrics C.struct_rt_metrics - -type Mclpool C.struct_mclpool - -// Berkeley packet filter - -const ( - SizeofBpfVersion = C.sizeof_struct_bpf_version - SizeofBpfStat = C.sizeof_struct_bpf_stat - SizeofBpfProgram = C.sizeof_struct_bpf_program - SizeofBpfInsn = C.sizeof_struct_bpf_insn - SizeofBpfHdr = C.sizeof_struct_bpf_hdr -) - -type BpfVersion C.struct_bpf_version - -type BpfStat C.struct_bpf_stat - -type BpfProgram C.struct_bpf_program - -type BpfInsn C.struct_bpf_insn - -type BpfHdr C.struct_bpf_hdr - -type BpfTimeval C.struct_bpf_timeval - -// Terminal handling - -type Termios C.struct_termios - -type Winsize C.struct_winsize - -// fchmodat-like syscalls. - -const ( - AT_FDCWD = C.AT_FDCWD - AT_SYMLINK_NOFOLLOW = C.AT_SYMLINK_NOFOLLOW -) - -// poll - -type PollFd C.struct_pollfd - -const ( - POLLERR = C.POLLERR - POLLHUP = C.POLLHUP - POLLIN = C.POLLIN - POLLNVAL = C.POLLNVAL - POLLOUT = C.POLLOUT - POLLPRI = C.POLLPRI - POLLRDBAND = C.POLLRDBAND - POLLRDNORM = C.POLLRDNORM - POLLWRBAND = C.POLLWRBAND - POLLWRNORM = C.POLLWRNORM -) - -// Sysctl - -type Sysctlnode C.struct_sysctlnode - -// Uname - -type Utsname C.struct_utsname diff --git a/vendor/golang.org/x/sys/unix/types_openbsd.go b/vendor/golang.org/x/sys/unix/types_openbsd.go deleted file mode 100644 index 649e55998..000000000 --- a/vendor/golang.org/x/sys/unix/types_openbsd.go +++ /dev/null @@ -1,282 +0,0 @@ -// Copyright 2009 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build ignore - -/* -Input to cgo -godefs. See README.md -*/ - -// +godefs map struct_in_addr [4]byte /* in_addr */ -// +godefs map struct_in6_addr [16]byte /* in6_addr */ - -package unix - -/* -#define KERNEL -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include - -enum { - sizeofPtr = sizeof(void*), -}; - -union sockaddr_all { - struct sockaddr s1; // this one gets used for fields - struct sockaddr_in s2; // these pad it out - struct sockaddr_in6 s3; - struct sockaddr_un s4; - struct sockaddr_dl s5; -}; - -struct sockaddr_any { - struct sockaddr addr; - char pad[sizeof(union sockaddr_all) - sizeof(struct sockaddr)]; -}; - -*/ -import "C" - -// Machine characteristics; for internal use. - -const ( - sizeofPtr = C.sizeofPtr - sizeofShort = C.sizeof_short - sizeofInt = C.sizeof_int - sizeofLong = C.sizeof_long - sizeofLongLong = C.sizeof_longlong -) - -// Basic types - -type ( - _C_short C.short - _C_int C.int - _C_long C.long - _C_long_long C.longlong -) - -// Time - -type Timespec C.struct_timespec - -type Timeval C.struct_timeval - -// Processes - -type Rusage C.struct_rusage - -type Rlimit C.struct_rlimit - -type _Gid_t C.gid_t - -// Files - -const ( // Directory mode bits - S_IFMT = C.S_IFMT - S_IFIFO = C.S_IFIFO - S_IFCHR = C.S_IFCHR - S_IFDIR = C.S_IFDIR - S_IFBLK = C.S_IFBLK - S_IFREG = C.S_IFREG - S_IFLNK = C.S_IFLNK - S_IFSOCK = C.S_IFSOCK - S_ISUID = C.S_ISUID - S_ISGID = C.S_ISGID - S_ISVTX = C.S_ISVTX - S_IRUSR = C.S_IRUSR - S_IWUSR = C.S_IWUSR - S_IXUSR = C.S_IXUSR -) - -type Stat_t C.struct_stat - -type Statfs_t C.struct_statfs - -type Flock_t C.struct_flock - -type Dirent C.struct_dirent - -type Fsid C.fsid_t - -// File system limits - -const ( - PathMax = C.PATH_MAX -) - -// Sockets - -type RawSockaddrInet4 C.struct_sockaddr_in - -type RawSockaddrInet6 C.struct_sockaddr_in6 - -type RawSockaddrUnix C.struct_sockaddr_un - -type RawSockaddrDatalink C.struct_sockaddr_dl - -type RawSockaddr C.struct_sockaddr - -type RawSockaddrAny C.struct_sockaddr_any - -type _Socklen C.socklen_t - -type Linger C.struct_linger - -type Iovec C.struct_iovec - -type IPMreq C.struct_ip_mreq - -type IPv6Mreq C.struct_ipv6_mreq - -type Msghdr C.struct_msghdr - -type Cmsghdr C.struct_cmsghdr - -type Inet6Pktinfo C.struct_in6_pktinfo - -type IPv6MTUInfo C.struct_ip6_mtuinfo - -type ICMPv6Filter C.struct_icmp6_filter - -const ( - SizeofSockaddrInet4 = C.sizeof_struct_sockaddr_in - SizeofSockaddrInet6 = C.sizeof_struct_sockaddr_in6 - SizeofSockaddrAny = C.sizeof_struct_sockaddr_any - SizeofSockaddrUnix = C.sizeof_struct_sockaddr_un - SizeofSockaddrDatalink = C.sizeof_struct_sockaddr_dl - SizeofLinger = C.sizeof_struct_linger - SizeofIPMreq = C.sizeof_struct_ip_mreq - SizeofIPv6Mreq = C.sizeof_struct_ipv6_mreq - SizeofMsghdr = C.sizeof_struct_msghdr - SizeofCmsghdr = C.sizeof_struct_cmsghdr - SizeofInet6Pktinfo = C.sizeof_struct_in6_pktinfo - SizeofIPv6MTUInfo = C.sizeof_struct_ip6_mtuinfo - SizeofICMPv6Filter = C.sizeof_struct_icmp6_filter -) - -// Ptrace requests - -const ( - PTRACE_TRACEME = C.PT_TRACE_ME - PTRACE_CONT = C.PT_CONTINUE - PTRACE_KILL = C.PT_KILL -) - -// Events (kqueue, kevent) - -type Kevent_t C.struct_kevent - -// Select - -type FdSet C.fd_set - -// Routing and interface messages - -const ( - SizeofIfMsghdr = C.sizeof_struct_if_msghdr - SizeofIfData = C.sizeof_struct_if_data - SizeofIfaMsghdr = C.sizeof_struct_ifa_msghdr - SizeofIfAnnounceMsghdr = C.sizeof_struct_if_announcemsghdr - SizeofRtMsghdr = C.sizeof_struct_rt_msghdr - SizeofRtMetrics = C.sizeof_struct_rt_metrics -) - -type IfMsghdr C.struct_if_msghdr - -type IfData C.struct_if_data - -type IfaMsghdr C.struct_ifa_msghdr - -type IfAnnounceMsghdr C.struct_if_announcemsghdr - -type RtMsghdr C.struct_rt_msghdr - -type RtMetrics C.struct_rt_metrics - -type Mclpool C.struct_mclpool - -// Berkeley packet filter - -const ( - SizeofBpfVersion = C.sizeof_struct_bpf_version - SizeofBpfStat = C.sizeof_struct_bpf_stat - SizeofBpfProgram = C.sizeof_struct_bpf_program - SizeofBpfInsn = C.sizeof_struct_bpf_insn - SizeofBpfHdr = C.sizeof_struct_bpf_hdr -) - -type BpfVersion C.struct_bpf_version - -type BpfStat C.struct_bpf_stat - -type BpfProgram C.struct_bpf_program - -type BpfInsn C.struct_bpf_insn - -type BpfHdr C.struct_bpf_hdr - -type BpfTimeval C.struct_bpf_timeval - -// Terminal handling - -type Termios C.struct_termios - -type Winsize C.struct_winsize - -// fchmodat-like syscalls. - -const ( - AT_FDCWD = C.AT_FDCWD - AT_SYMLINK_NOFOLLOW = C.AT_SYMLINK_NOFOLLOW -) - -// poll - -type PollFd C.struct_pollfd - -const ( - POLLERR = C.POLLERR - POLLHUP = C.POLLHUP - POLLIN = C.POLLIN - POLLNVAL = C.POLLNVAL - POLLOUT = C.POLLOUT - POLLPRI = C.POLLPRI - POLLRDBAND = C.POLLRDBAND - POLLRDNORM = C.POLLRDNORM - POLLWRBAND = C.POLLWRBAND - POLLWRNORM = C.POLLWRNORM -) - -// Uname - -type Utsname C.struct_utsname diff --git a/vendor/golang.org/x/sys/unix/xattr_test.go b/vendor/golang.org/x/sys/unix/xattr_test.go deleted file mode 100644 index b8b28d0c1..000000000 --- a/vendor/golang.org/x/sys/unix/xattr_test.go +++ /dev/null @@ -1,119 +0,0 @@ -// Copyright 2018 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// +build darwin freebsd linux - -package unix_test - -import ( - "os" - "runtime" - "strings" - "testing" - - "golang.org/x/sys/unix" -) - -func TestXattr(t *testing.T) { - defer chtmpdir(t)() - - f := "xattr1" - touch(t, f) - - xattrName := "user.test" - xattrDataSet := "gopher" - err := unix.Setxattr(f, xattrName, []byte(xattrDataSet), 0) - if err == unix.ENOTSUP || err == unix.EOPNOTSUPP { - t.Skip("filesystem does not support extended attributes, skipping test") - } else if err != nil { - t.Fatalf("Setxattr: %v", err) - } - - // find size - size, err := unix.Listxattr(f, nil) - if err != nil { - t.Fatalf("Listxattr: %v", err) - } - - if size <= 0 { - t.Fatalf("Listxattr returned an empty list of attributes") - } - - buf := make([]byte, size) - read, err := unix.Listxattr(f, buf) - if err != nil { - t.Fatalf("Listxattr: %v", err) - } - - xattrs := stringsFromByteSlice(buf[:read]) - - xattrWant := xattrName - if runtime.GOOS == "freebsd" { - // On FreeBSD, the namespace is stored separately from the xattr - // name and Listxattr doesn't return the namespace prefix. - xattrWant = strings.TrimPrefix(xattrWant, "user.") - } - found := false - for _, name := range xattrs { - if name == xattrWant { - found = true - } - } - - if !found { - t.Errorf("Listxattr did not return previously set attribute '%s'", xattrName) - } - - // find size - size, err = unix.Getxattr(f, xattrName, nil) - if err != nil { - t.Fatalf("Getxattr: %v", err) - } - - if size <= 0 { - t.Fatalf("Getxattr returned an empty attribute") - } - - xattrDataGet := make([]byte, size) - _, err = unix.Getxattr(f, xattrName, xattrDataGet) - if err != nil { - t.Fatalf("Getxattr: %v", err) - } - - got := string(xattrDataGet) - if got != xattrDataSet { - t.Errorf("Getxattr: expected attribute value %s, got %s", xattrDataSet, got) - } - - err = unix.Removexattr(f, xattrName) - if err != nil { - t.Fatalf("Removexattr: %v", err) - } - - n := "nonexistent" - err = unix.Lsetxattr(n, xattrName, []byte(xattrDataSet), 0) - if err != unix.ENOENT { - t.Errorf("Lsetxattr: expected %v on non-existent file, got %v", unix.ENOENT, err) - } - - _, err = unix.Lgetxattr(n, xattrName, nil) - if err != unix.ENOENT { - t.Errorf("Lgetxattr: %v", err) - } - - s := "symlink1" - err = os.Symlink(n, s) - if err != nil { - t.Fatal(err) - } - - err = unix.Lsetxattr(s, xattrName, []byte(xattrDataSet), 0) - if err != nil { - // Linux and Android doen't support xattrs on symlinks according - // to xattr(7), so just test that we get the proper error. - if (runtime.GOOS != "linux" && runtime.GOOS != "android") || err != unix.EPERM { - t.Fatalf("Lsetxattr: %v", err) - } - } -} diff --git a/vendor/vendor.json b/vendor/vendor.json index 779cb773f..059c9cd56 100644 --- a/vendor/vendor.json +++ b/vendor/vendor.json @@ -2,19 +2,17 @@ "comment": "", "ignore": "test", "package": [ - {"path":"context","revision":""}, {"path":"github.com/Azure/azure-sdk-for-go/services/network/mgmt/2015-06-15/network","checksumSHA1":"Fs6Gcl0nhC0FJ6MsB+Ck7r4huYo=","revision":"767429fcb996dad413936d682c28301e6739bade","revisionTime":"2018-05-01T22:35:11Z"}, {"path":"github.com/Azure/azure-sdk-for-go/version","checksumSHA1":"FAw+h8wS2QiQEIVC3z/8R2q1bvY=","revision":"767429fcb996dad413936d682c28301e6739bade","revisionTime":"2018-05-01T22:35:11Z"}, - {"path":"github.com/Azure/go-ansiterm","checksumSHA1":"23FJUX+AInYeEM2hoUMvYZtXZd4=","revision":"fa152c58bc15761d0200cb75fe958b89a9d4888e","revisionTime":"2016-06-22T17:32:16Z"}, - {"path":"github.com/Azure/go-ansiterm/winterm","checksumSHA1":"jBimnggjIiFUjaImNoJhSVLtdzw=","revision":"fa152c58bc15761d0200cb75fe958b89a9d4888e","revisionTime":"2016-06-22T17:32:16Z"}, + {"path":"github.com/Azure/go-ansiterm","checksumSHA1":"9NFR6RG8H2fNyKHscGmuGLQhRm4=","revision":"d6e3b3328b783f23731bc4d058875b0371ff8109","revisionTime":"2017-09-29T23:40:23Z","version":"master","versionExact":"master"}, + {"path":"github.com/Azure/go-ansiterm/winterm","checksumSHA1":"3/UphB+6Hbx5otA4PjFjvObT+L4=","revision":"d6e3b3328b783f23731bc4d058875b0371ff8109","revisionTime":"2017-09-29T23:40:23Z","version":"master","versionExact":"master"}, {"path":"github.com/DataDog/datadog-go/statsd","checksumSHA1":"WvApwvvSe3i/3KO8300dyeFmkbI=","revision":"b10af4b12965a1ad08d164f57d14195b4140d8de","revisionTime":"2017-08-09T10:47:06Z"}, {"path":"github.com/Microsoft/go-winio","checksumSHA1":"AzjRkOQtVBTwIw4RJLTygFhJs3s=","revision":"f533f7a102197536779ea3a8cb881d639e21ec5a","revisionTime":"2017-05-24T00:36:31Z"}, {"path":"github.com/NVIDIA/gpu-monitoring-tools","checksumSHA1":"kF1vk+8Xvb3nGBiw9+qbUc0SZ4M=","revision":"86f2a9fac6c5b597dc494420005144b8ef7ec9fb","revisionTime":"2018-08-29T22:20:09Z"}, {"path":"github.com/NVIDIA/gpu-monitoring-tools/bindings/go/nvml","checksumSHA1":"P8FATSSgpe5A17FyPrGpsX95Xw8=","revision":"86f2a9fac6c5b597dc494420005144b8ef7ec9fb","revisionTime":"2018-08-29T22:20:09Z"}, {"path":"github.com/NYTimes/gziphandler","checksumSHA1":"jktW57+vJsziNVPeXMCoujTzdW4=","revision":"97ae7fbaf81620fe97840685304a78a306a39c64","revisionTime":"2017-09-16T00:36:49Z"}, {"path":"github.com/Nvveen/Gotty","checksumSHA1":"Aqy8/FoAIidY/DeQ5oTYSZ4YFVc=","revision":"cd527374f1e5bff4938207604a14f2e38a9cf512","revisionTime":"2012-06-04T00:48:16Z"}, - {"path":"github.com/RackSec/srslog","checksumSHA1":"OTN4c1F0p+mEG2CpkU1Kuavupf0=","revision":"259aed10dfa74ea2961eddd1d9847619f6e98837","revisionTime":"2016-01-20T22:33:50Z"}, - {"path":"github.com/Sirupsen/logrus","comment":"v0.8.7-87-g4b6ea73","revision":"4b6ea7319e214d98c938f12692336f7ca9348d6b"}, + {"path":"github.com/RackSec/srslog","checksumSHA1":"GkeYcZh0B14rtg15JAVDPVwjNoY=","revision":"259aed10dfa74ea2961eddd1d9847619f6e98837","revisionTime":"2016-01-20T22:33:50Z"}, {"path":"github.com/StackExchange/wmi","checksumSHA1":"qtjd74+bErubh+qyv3s+lWmn9wc=","revision":"ea383cf3ba6ec950874b8486cd72356d007c768f","revisionTime":"2017-04-10T19:29:09Z"}, {"path":"github.com/agext/levenshtein","checksumSHA1":"jQh1fnoKPKMURvKkpdRjN695nAQ=","revision":"5f10fee965225ac1eecdc234c09daf5cd9e7f7b6","revisionTime":"2017-02-17T06:30:20Z"}, {"path":"github.com/apparentlymart/go-textseg/textseg","checksumSHA1":"Ffhtm8iHH7l2ynVVOIGJE3eiuLA=","revision":"b836f5c4d331d1945a2fead7188db25432d73b69","revisionTime":"2017-05-31T20:39:52Z"}, @@ -22,37 +20,37 @@ {"path":"github.com/appc/spec/schema/common","checksumSHA1":"Q47G6996hbfQaNp/8CFkGWTVQpw=","revision":"cbe99b7160b1397bf89f9c8bb1418f69c9424049","revisionTime":"2017-09-19T09:55:19Z"}, {"path":"github.com/appc/spec/schema/types","checksumSHA1":"kYXCle7Ikc8WqiMs7NXz99bUWqo=","revision":"cbe99b7160b1397bf89f9c8bb1418f69c9424049","revisionTime":"2017-09-19T09:55:19Z"}, {"path":"github.com/appc/spec/schema/types/resource","checksumSHA1":"VgPsPj5PH7LKXMa3ZLe5/+Avydo=","revision":"cbe99b7160b1397bf89f9c8bb1418f69c9424049","revisionTime":"2017-09-19T09:55:19Z"}, - {"path":"github.com/armon/circbuf","revision":"bbbad097214e2918d8543d5201d12bfd7bca254d"}, + {"path":"github.com/armon/circbuf","checksumSHA1":"l0iFqayYAaEip6Olaq3/LCOa/Sg=","revision":"bbbad097214e2918d8543d5201d12bfd7bca254d"}, {"path":"github.com/armon/go-metrics","checksumSHA1":"xp/2s4XclLL17DThGBI7jXZ4Crs=","revision":"6c3acc97c61d04290a8ba2e54640151f54c1546a","revisionTime":"2017-11-16T18:41:20Z"}, {"path":"github.com/armon/go-metrics/circonus","checksumSHA1":"xCsGGM9TKBogZDfSN536KtQdLko=","revision":"6c3acc97c61d04290a8ba2e54640151f54c1546a","revisionTime":"2017-11-16T18:41:20Z"}, {"path":"github.com/armon/go-metrics/datadog","checksumSHA1":"Dt0n1sSivvvdZQdzc4Hu/yOG+T0=","revision":"6c3acc97c61d04290a8ba2e54640151f54c1546a","revisionTime":"2017-11-16T18:41:20Z"}, {"path":"github.com/armon/go-metrics/prometheus","checksumSHA1":"XfPPXw55zKziOWnZbkEGEJ96O9c=","revision":"6c3acc97c61d04290a8ba2e54640151f54c1546a","revisionTime":"2017-11-16T18:41:20Z"}, {"path":"github.com/armon/go-radix","checksumSHA1":"gNO0JNpLzYOdInGeq7HqMZUzx9M=","revision":"4239b77079c7b5d1243b7b4736304ce8ddb6f0f2","revisionTime":"2016-01-15T23:47:25Z"}, - {"path":"github.com/aws/aws-sdk-go/aws","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/aws/awserr","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/aws/awsutil","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/aws/client","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/aws/client/metadata","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/aws/corehandlers","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/aws/credentials","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/aws/credentials/ec2rolecreds","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/aws/defaults","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/aws/ec2metadata","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/aws/request","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/aws/session","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/private/endpoints","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/private/protocol/query","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/private/protocol/query/queryutil","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/private/protocol/rest","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/private/protocol/restxml","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/private/protocol/xml/xmlutil","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/private/signer/v4","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/private/waiter","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, - {"path":"github.com/aws/aws-sdk-go/service/s3","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws","checksumSHA1":"gMMx/JVSQVE4KHKoJQC7cqjTIZc=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws/awserr","checksumSHA1":"OlpdlsDV2Qv50MuHlpH9heaHjQc=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws/awsutil","checksumSHA1":"MoxolxrlsmtDri7ndvx5gkXI2hU=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws/client","checksumSHA1":"qZUhjIZT8zU/xdQJ3La948GqEgU=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws/client/metadata","checksumSHA1":"ieAJ+Cvp/PKv1LpUEnUXpc3OI6E=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws/corehandlers","checksumSHA1":"VAb9dwYeOW1iViqztJq8DGtlrV4=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws/credentials","checksumSHA1":"x1grxMebz9A06jOsLieoExjhltU=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws/credentials/ec2rolecreds","checksumSHA1":"nCeVh7E8twNLP887Zanei1wd6ks=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws/defaults","checksumSHA1":"9c2Th6//lUUZPehmPkHXJh4hE/s=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws/ec2metadata","checksumSHA1":"xyUdqBgj3ltt9LXnY4UvEct2izQ=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws/request","checksumSHA1":"0i8r38TS6u3uT2HBM+8tNsozY6w=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/aws/session","checksumSHA1":"S46ej8/Fd7YnczK42I1UnjoeqqM=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/private/endpoints","checksumSHA1":"Rv6k85b5aRb0la07XzRPsrzwrHE=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/private/protocol/query","checksumSHA1":"ndEFDDt1gd5taOeeUrf0T66H+bg=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/private/protocol/query/queryutil","checksumSHA1":"j8UDnDnB59+u1w/qBHKG0PSWH8U=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/private/protocol/rest","checksumSHA1":"WRZcRqV95LFgq5V09PmDmSmNTEw=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/private/protocol/restxml","checksumSHA1":"KVdpXAG2PmIyopNC4jJd/JdjefA=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/private/protocol/xml/xmlutil","checksumSHA1":"xmKYerjqq1FO8kJK0/4Ds8iGyWg=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/private/signer/v4","checksumSHA1":"H0s7iZn2nk/fq52QaaTeW+gSGgA=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/private/waiter","checksumSHA1":"Pfoou5jtRj8A0SuCl8toNbXQlfw=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, + {"path":"github.com/aws/aws-sdk-go/service/s3","checksumSHA1":"NAitp0A1zE3Zry+Z41LOlCr+9lQ=","comment":"v1.0.6-2-g80dd495","revision":"80dd4951fdb3f711d31843b8d87871130ef2df67"}, {"path":"github.com/beorn7/perks/quantile","checksumSHA1":"spyv5/YFBjYyZLZa1U2LBfDR8PM=","revision":"4c0e84591b9aa9e6dcfdf3e020114cd81f89d5f9","revisionTime":"2016-08-04T10:47:26Z"}, {"path":"github.com/bgentry/go-netrc/netrc","checksumSHA1":"nqw2Qn5xUklssHTubS5HDvEL9L4=","revision":"9fd32a8b3d3d3f9d43c341bfe098430e07609480","revisionTime":"2014-04-22T17:41:19Z"}, - {"path":"github.com/bgentry/speakeasy","revision":"36e9cfdd690967f4f690c6edcc9ffacd006014a0"}, - {"path":"github.com/bgentry/speakeasy/example","revision":"36e9cfdd690967f4f690c6edcc9ffacd006014a0"}, + {"path":"github.com/bgentry/speakeasy","checksumSHA1":"7SbTaY0kaYxgQrG3/9cjrI+BcyU=","revision":"36e9cfdd690967f4f690c6edcc9ffacd006014a0"}, + {"path":"github.com/bgentry/speakeasy/example","checksumSHA1":"twtRfb6484vfr2qqjiFkLThTjcQ=","revision":"36e9cfdd690967f4f690c6edcc9ffacd006014a0"}, {"path":"github.com/boltdb/bolt","checksumSHA1":"R1Q34Pfnt197F/nCOO9kG8c+Z90=","comment":"v1.2.0","revision":"2f1ce7a837dcb8da3ec595b1dac9d0632f0f99e8","revisionTime":"2017-07-17T17:11:48Z","version":"v1.3.1","versionExact":"v1.3.1"}, {"path":"github.com/burntsushi/toml","checksumSHA1":"InIrfOI7Ys1QqZpCgTB4yW1G32M=","revision":"99064174e013895bbd9b025c31100bd1d9b590ca","revisionTime":"2016-07-17T15:07:09Z"}, {"path":"github.com/circonus-labs/circonus-gometrics","checksumSHA1":"/qvtQq5y0RZCsRyOOsan87V2AL0=","revision":"dd698dc110872f6e554abf74a7740fc363354086","revisionTime":"2018-08-20T20:09:38Z"}, @@ -60,57 +58,77 @@ {"path":"github.com/circonus-labs/circonus-gometrics/api/config","checksumSHA1":"bQhz/fcyZPmuHSH2qwC4ZtATy5c=","revision":"dd698dc110872f6e554abf74a7740fc363354086","revisionTime":"2018-08-20T20:09:38Z"}, {"path":"github.com/circonus-labs/circonus-gometrics/checkmgr","checksumSHA1":"Ij8yB33E0Kk+GfTkNRoF1mG26dc=","revision":"dd698dc110872f6e554abf74a7740fc363354086","revisionTime":"2018-08-20T20:09:38Z"}, {"path":"github.com/circonus-labs/circonusllhist","checksumSHA1":"VbfeVqeOM+dTNxCmpvmYS0LwQn0=","revision":"7d649b46cdc2cd2ed102d350688a75a4fd7778c6","revisionTime":"2016-11-21T13:51:53Z"}, + {"path":"github.com/containerd/console","checksumSHA1":"IGtuR58l2zmYRcNf8sPDlCSgovE=","origin":"github.com/opencontainers/runc/vendor/github.com/containerd/console","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/containerd/continuity/pathdriver","checksumSHA1":"GqIrOttKaO7k6HIaHQLPr3cY7rY=","origin":"github.com/docker/docker/vendor/github.com/containerd/continuity/pathdriver","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, {"path":"github.com/containernetworking/cni/pkg/types","checksumSHA1":"NeAp/3+Hedu9tnMai+LihERPj84=","revision":"5c3c17164270150467498a32c71436c7cd5501be","revisionTime":"2016-06-02T16:00:07Z"}, {"path":"github.com/coreos/go-semver/semver","checksumSHA1":"97BsbXOiZ8+Kr+LIuZkQFtSj7H4=","revision":"1817cd4bea52af76542157eeabd74b057d1a199e","revisionTime":"2017-06-13T09:22:38Z"}, + {"path":"github.com/coreos/go-systemd/dbus","checksumSHA1":"/zxxFPYjUB7Wowz33r5AhTDvoz0=","origin":"github.com/opencontainers/runc/vendor/github.com/coreos/go-systemd/dbus","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/coreos/go-systemd/util","checksumSHA1":"e8qgBHxXbij3RVspqrkeBzMZ564=","origin":"github.com/opencontainers/runc/vendor/github.com/coreos/go-systemd/util","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/coreos/pkg/dlopen","checksumSHA1":"O8c/VKtW34XPJNNlyeb/im8vWSI=","origin":"github.com/opencontainers/runc/vendor/github.com/coreos/pkg/dlopen","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/cyphar/filepath-securejoin","checksumSHA1":"4Cq4wS4l0O8WlugamGEPvooJPAk=","origin":"github.com/opencontainers/runc/vendor/github.com/cyphar/filepath-securejoin","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, {"path":"github.com/davecgh/go-spew/spew","checksumSHA1":"mrz/kicZiUaHxkyfvC/DyQcr8Do=","revision":"ecdeabc65495df2dec95d7c4a4c3e021903035e5","revisionTime":"2017-10-02T20:02:53Z"}, - {"path":"github.com/docker/distribution/digestset","checksumSHA1":"Gj+xR1VgFKKmFXYOJMnAczC3Znk=","revision":"69c7f303d511a5777e79505b23340d37d532d7ac","revisionTime":"2017-01-11T18:37:34Z"}, - {"path":"github.com/docker/distribution/reference","checksumSHA1":"FqADPtHUqEtmfC7Zf+0pI++Kb6g=","revision":"69c7f303d511a5777e79505b23340d37d532d7ac","revisionTime":"2017-01-11T18:37:34Z"}, - {"path":"github.com/docker/docker/api/types","checksumSHA1":"HoN/78ovv3/DC+kDKF7IENEc40g=","revision":"e4d0fe84f9ea88b0e0cfd847412c9f29442cc62d","revisionTime":"2017-11-14T19:25:14Z"}, - {"path":"github.com/docker/docker/api/types/blkiodev","checksumSHA1":"jVJDbe0IcyjoKc2xbohwzQr+FF0=","revision":"3dc87714d1bfcaf0620fabec3bfea89620b59337","revisionTime":"2017-02-14T23:27:04Z"}, - {"path":"github.com/docker/docker/api/types/container","checksumSHA1":"Vso/6NenP1G74lQjvyvzOdfIZ28=","revision":"3dc87714d1bfcaf0620fabec3bfea89620b59337","revisionTime":"2017-02-14T23:27:04Z"}, - {"path":"github.com/docker/docker/api/types/filters","checksumSHA1":"J2OKngfI3vgswudr9PZVUFcRRu0=","revision":"fe9ab0588606a5566d065bc68ae68f3926ddaa72","revisionTime":"2017-02-28T18:28:19Z"}, - {"path":"github.com/docker/docker/api/types/mount","checksumSHA1":"OXsrx4ynzLV+6/6vUeyru0Fprx8=","revision":"3dc87714d1bfcaf0620fabec3bfea89620b59337","revisionTime":"2017-02-14T23:27:04Z"}, - {"path":"github.com/docker/docker/api/types/network","checksumSHA1":"Gskp+nvbVe8Gk1xPLHylZvNmqTg=","revision":"e4d0fe84f9ea88b0e0cfd847412c9f29442cc62d","revisionTime":"2017-11-14T19:25:14Z"}, - {"path":"github.com/docker/docker/api/types/registry","checksumSHA1":"r2vWq7Uc3ExKzMqYgH0b4AKjLKY=","revision":"e4d0fe84f9ea88b0e0cfd847412c9f29442cc62d","revisionTime":"2017-11-14T19:25:14Z"}, - {"path":"github.com/docker/docker/api/types/strslice","checksumSHA1":"VTxWyFud/RedrpllGdQonVtGM/A=","revision":"3dc87714d1bfcaf0620fabec3bfea89620b59337","revisionTime":"2017-02-14T23:27:04Z"}, - {"path":"github.com/docker/docker/api/types/swarm","checksumSHA1":"Q0U3queMsCw+rPPztXnRHwAxQEc=","revision":"e4d0fe84f9ea88b0e0cfd847412c9f29442cc62d","revisionTime":"2017-11-14T19:25:14Z"}, - {"path":"github.com/docker/docker/api/types/swarm/runtime","checksumSHA1":"mi8EDCDjtrZEONRXPG7VHJosDwY=","revision":"e4d0fe84f9ea88b0e0cfd847412c9f29442cc62d","revisionTime":"2017-11-14T19:25:14Z"}, - {"path":"github.com/docker/docker/api/types/versions","checksumSHA1":"uDPQ3nHsrvGQc9tg/J9OSC4N5dQ=","revision":"fe9ab0588606a5566d065bc68ae68f3926ddaa72","revisionTime":"2017-02-28T18:28:19Z"}, - {"path":"github.com/docker/docker/cli/config/configfile","checksumSHA1":"l8Re54Tp3x8kYWPoD2jXBOWloSY=","revision":"518945b6bd30dc535addf2aac9d5eeefbfc5e523","revisionTime":"2017-01-10T19:35:11Z"}, - {"path":"github.com/docker/docker/opts","checksumSHA1":"mswe275heIklTKj7mPTnVzAFoMk=","revision":"da39e9a4f920a15683dd0f23923c302d4db6eed5","revisionTime":"2016-05-28T08:11:04Z"}, - {"path":"github.com/docker/docker/pkg/archive","checksumSHA1":"OShlvanyvyee8I0/kdmLuOEOF5w=","revision":"da39e9a4f920a15683dd0f23923c302d4db6eed5","revisionTime":"2016-05-28T08:11:04Z"}, - {"path":"github.com/docker/docker/pkg/fileutils","checksumSHA1":"EONnM7E8xCzJCAbX1rhayK6knwM=","revision":"da39e9a4f920a15683dd0f23923c302d4db6eed5","revisionTime":"2016-05-28T08:11:04Z"}, - {"path":"github.com/docker/docker/pkg/homedir","checksumSHA1":"p6Ud4Yf1ywWy20YxXF1RU4yhTio=","revision":"da39e9a4f920a15683dd0f23923c302d4db6eed5","revisionTime":"2016-05-28T08:11:04Z"}, - {"path":"github.com/docker/docker/pkg/idtools","checksumSHA1":"iP5slJJPRZUm0rfdII8OiATAACA=","revision":"02caa73df411debed164f520a6a1304778f8b88c","revisionTime":"2016-05-28T10:48:36Z"}, - {"path":"github.com/docker/docker/pkg/idtools","checksumSHA1":"iP5slJJPRZUm0rfdII8OiATAACA=","revision":"52debcd58ac91bf68503ce60561536911b74ff05","revisionTime":"2016-05-20T15:17:10Z"}, - {"path":"github.com/docker/docker/pkg/ioutils","checksumSHA1":"tdhmIGUaoOMEDymMC23qTS7bt0g=","revision":"52debcd58ac91bf68503ce60561536911b74ff05","revisionTime":"2016-05-20T15:17:10Z"}, - {"path":"github.com/docker/docker/pkg/ioutils","checksumSHA1":"tdhmIGUaoOMEDymMC23qTS7bt0g=","revision":"da39e9a4f920a15683dd0f23923c302d4db6eed5","revisionTime":"2016-05-28T08:11:04Z"}, - {"path":"github.com/docker/docker/pkg/jsonlog","checksumSHA1":"BlFSSK7zUjPzPuxkLmM/0wpvku8=","revision":"3dc87714d1bfcaf0620fabec3bfea89620b59337","revisionTime":"2017-02-14T23:27:04Z"}, - {"path":"github.com/docker/docker/pkg/jsonmessage","checksumSHA1":"IpcW+FAHu0DmbvbhqXuP42f4FCo=","revision":"3dc87714d1bfcaf0620fabec3bfea89620b59337","revisionTime":"2017-02-14T23:27:04Z"}, - {"path":"github.com/docker/docker/pkg/longpath","checksumSHA1":"ndnAFCfsGC3upNQ6jAEwzxcurww=","revision":"da39e9a4f920a15683dd0f23923c302d4db6eed5","revisionTime":"2016-05-28T08:11:04Z"}, - {"path":"github.com/docker/docker/pkg/pools","checksumSHA1":"rArZ5mYIe9I1L5PRQOJu8BwafFw=","revision":"52debcd58ac91bf68503ce60561536911b74ff05","revisionTime":"2016-05-20T15:17:10Z"}, - {"path":"github.com/docker/docker/pkg/pools","checksumSHA1":"rArZ5mYIe9I1L5PRQOJu8BwafFw=","revision":"da39e9a4f920a15683dd0f23923c302d4db6eed5","revisionTime":"2016-05-28T08:11:04Z"}, - {"path":"github.com/docker/docker/pkg/promise","checksumSHA1":"txf3EORYff4hO6PEvwBm2lyh1MU=","revision":"52debcd58ac91bf68503ce60561536911b74ff05","revisionTime":"2016-05-20T15:17:10Z"}, - {"path":"github.com/docker/docker/pkg/promise","checksumSHA1":"txf3EORYff4hO6PEvwBm2lyh1MU=","revision":"da39e9a4f920a15683dd0f23923c302d4db6eed5","revisionTime":"2016-05-28T08:11:04Z"}, - {"path":"github.com/docker/docker/pkg/random","checksumSHA1":"lThih54jzz9A4zHKEFb9SIV3Ed0=","revision":"518945b6bd30dc535addf2aac9d5eeefbfc5e523","revisionTime":"2017-01-10T19:35:11Z"}, - {"path":"github.com/docker/docker/pkg/stdcopy","checksumSHA1":"YDYbS5U2mDwfcOUJ6M09cP6Bubg=","revision":"da39e9a4f920a15683dd0f23923c302d4db6eed5","revisionTime":"2016-05-28T08:11:04Z"}, - {"path":"github.com/docker/docker/pkg/stringid","checksumSHA1":"YGZBluVbky9i5i+BgM+RTc0NEtc=","revision":"518945b6bd30dc535addf2aac9d5eeefbfc5e523","revisionTime":"2017-01-10T19:35:11Z"}, - {"path":"github.com/docker/docker/pkg/system","checksumSHA1":"NGqzajRG0Vri8ConQuHY3cCs1RA=","revision":"fe9ab0588606a5566d065bc68ae68f3926ddaa72","revisionTime":"2017-02-28T18:28:19Z"}, - {"path":"github.com/docker/docker/pkg/system","checksumSHA1":"jtruj9m4YJtPaXqUZ5PXdgFQjRA=","revision":"47cd4bf526723100bdbbd5a6bb983cdd668882e9","revisionTime":"2017-03-13T17:42:22Z"}, - {"path":"github.com/docker/docker/pkg/term","checksumSHA1":"Nfp/0W+HK8ZEgklbSWmjJGXTJew=","revision":"3dc87714d1bfcaf0620fabec3bfea89620b59337","revisionTime":"2017-02-14T23:27:04Z"}, - {"path":"github.com/docker/docker/pkg/term/windows","checksumSHA1":"EMjTiUTHNxqSsmdMB1V29hRvJbQ=","revision":"3dc87714d1bfcaf0620fabec3bfea89620b59337","revisionTime":"2017-02-14T23:27:04Z"}, - {"path":"github.com/docker/docker/reference","checksumSHA1":"bASLGmo2Ae+10Pc2zVCpXbBcPag=","revision":"518945b6bd30dc535addf2aac9d5eeefbfc5e523","revisionTime":"2017-01-10T19:35:11Z"}, - {"path":"github.com/docker/docker/registry","checksumSHA1":"AlBQLsNEpPJIe7OxsqN1P3BKi/M=","revision":"518945b6bd30dc535addf2aac9d5eeefbfc5e523","revisionTime":"2017-01-10T19:35:11Z"}, + {"path":"github.com/docker/cli/cli/config/configfile","checksumSHA1":"wf9Rn3a9cPag5B9Dd+qHHEink+I=","revision":"67f9a3912cf944cf71b31f3fc14e3f2a18d95802","revisionTime":"2018-08-14T14:54:37Z","version":"v18.06.1-ce","versionExact":"v18.06.1-ce"}, + {"path":"github.com/docker/cli/cli/config/credentials","checksumSHA1":"fJpuGdxgATGNHm+INOPNVIhBnj0=","revision":"deb84a9e4e10b590e6de6aa6081532c87a5a2cfe","revisionTime":"2018-08-29T13:09:58Z"}, + {"path":"github.com/docker/cli/opts","checksumSHA1":"+yq5Rc1QTapDrr151x0m5ANZZeY=","revision":"67f9a3912cf944cf71b31f3fc14e3f2a18d95802","revisionTime":"2018-08-14T14:54:37Z","version":"v18.06.1-ce","versionExact":"v18.06.1-ce"}, + {"path":"github.com/docker/distribution","checksumSHA1":"ae06MP/1OVwQ/s/PsEp9wxfnBXM=","revision":"b12bd4004afc203f1cbd2072317c8fda30b89710","revisionTime":"2018-08-28T23:03:05Z"}, + {"path":"github.com/docker/distribution/digestset","checksumSHA1":"Gj+xR1VgFKKmFXYOJMnAczC3Znk=","revision":"83389a148052d74ac602f5f1d62f86ff2f3c4aa5","revisionTime":"2018-03-27T20:24:08Z"}, + {"path":"github.com/docker/distribution/metrics","checksumSHA1":"yqCaL8oUi3OlR/Mr4oHB5HKpstc=","revision":"b12bd4004afc203f1cbd2072317c8fda30b89710","revisionTime":"2018-08-28T23:03:05Z"}, + {"path":"github.com/docker/distribution/reference","checksumSHA1":"2Fe4D6PGaVE2he4fUeenLmhC1lE=","revision":"83389a148052d74ac602f5f1d62f86ff2f3c4aa5","revisionTime":"2018-03-27T20:24:08Z"}, + {"path":"github.com/docker/distribution/registry/api/errcode","checksumSHA1":"oFgg0LXTCZuYeI0/eEatdTyLexg=","revision":"83389a148052d74ac602f5f1d62f86ff2f3c4aa5","revisionTime":"2018-03-27T20:24:08Z"}, + {"path":"github.com/docker/distribution/registry/api/v2","checksumSHA1":"tjPyswv8NGYxreknmylv5r5tjt4=","revision":"b12bd4004afc203f1cbd2072317c8fda30b89710","revisionTime":"2018-08-28T23:03:05Z"}, + {"path":"github.com/docker/distribution/registry/client","checksumSHA1":"72zK3wP1n7UTcSFKZZz77sKXZiU=","revision":"b12bd4004afc203f1cbd2072317c8fda30b89710","revisionTime":"2018-08-28T23:03:05Z"}, + {"path":"github.com/docker/distribution/registry/client/auth","checksumSHA1":"UeouykquJzdjJv1+Vv0ikpe7Yvo=","revision":"83389a148052d74ac602f5f1d62f86ff2f3c4aa5","revisionTime":"2018-03-27T20:24:08Z"}, + {"path":"github.com/docker/distribution/registry/client/auth/challenge","checksumSHA1":"GWNgNhqeZST7+rgQdC06yEaLuQg=","revision":"83389a148052d74ac602f5f1d62f86ff2f3c4aa5","revisionTime":"2018-03-27T20:24:08Z"}, + {"path":"github.com/docker/distribution/registry/client/transport","checksumSHA1":"KjpG7FYMU5ugtc/fTfL1YqhdaV4=","revision":"83389a148052d74ac602f5f1d62f86ff2f3c4aa5","revisionTime":"2018-03-27T20:24:08Z"}, + {"path":"github.com/docker/distribution/registry/storage/cache","checksumSHA1":"RjRJSz/ISJEi0uWh5FlXMQetRcg=","revision":"b12bd4004afc203f1cbd2072317c8fda30b89710","revisionTime":"2018-08-28T23:03:05Z"}, + {"path":"github.com/docker/distribution/registry/storage/cache/memory","checksumSHA1":"T8G3A63WALmJ3JT/A0r01LG4KI0=","revision":"b12bd4004afc203f1cbd2072317c8fda30b89710","revisionTime":"2018-08-28T23:03:05Z"}, + {"path":"github.com/docker/docker-credential-helpers/client","checksumSHA1":"zcDmNPSzI1wVokOiHis5+JSg2Rk=","revision":"73e5f5dbfea31ee3b81111ebbf189785fa69731c","revisionTime":"2018-07-19T07:47:51Z"}, + {"path":"github.com/docker/docker-credential-helpers/credentials","checksumSHA1":"4u6EMQqD1zIqOHp76zQFLVH5V8U=","revision":"73e5f5dbfea31ee3b81111ebbf189785fa69731c","revisionTime":"2018-07-19T07:47:51Z"}, + {"path":"github.com/docker/docker/api/types","checksumSHA1":"hh2frMl2OiLNAzAdz48xzzzyGlw=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/api/types/blkiodev","checksumSHA1":"/jF0HVFiLzUUuywSjp4F/piM7BM=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/api/types/container","checksumSHA1":"lsxFU6qegOtXClSTthOvfPtly5I=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/api/types/filters","checksumSHA1":"y9EA6+kZQLx6kCM277CFHTm4eiw=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/api/types/mount","checksumSHA1":"k9CaJVvYL7SxcIP72ng/YcOuF9k=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/api/types/network","checksumSHA1":"IggTTG/yCDaV9SjtQz5SSarqUtc=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/api/types/registry","checksumSHA1":"m4Jg5WnW75I65nvkEno8PElSXik=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/api/types/strslice","checksumSHA1":"OQEUS/2J2xVHpfvcsxcXzYqBSeY=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/api/types/swarm","checksumSHA1":"qlMl0DqET1NC5GFyTSM5b305DL4=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/api/types/swarm/runtime","checksumSHA1":"txs5EKTbKgVyKmKKSnaH3fr+odA=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/api/types/versions","checksumSHA1":"MZsgRjJJ0D/gAsXfKiEys+op6dE=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/errdefs","checksumSHA1":"dF9WiXuwISBPd8bQfqpuoWtB3jk=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/opts","checksumSHA1":"u6EOrZRfhdjr4up14b2JJ7MMMaY=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/archive","checksumSHA1":"6jrp+51MWrqN07RIEXQfkwDzjbU=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/fileutils","checksumSHA1":"CYkS5Yh4GZ80KS+n/o+c5d0ktsA=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/homedir","checksumSHA1":"y37I+5AS96wQSiAxOayiMgnZawA=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/idtools","checksumSHA1":"10t/hGlBdat1QuSmLJ59ynG62BM=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/ioutils","checksumSHA1":"Ybq78CnAoQWVBk+lkh3zykmcSjs=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/jsonmessage","checksumSHA1":"KQflv+x9hoJywgvxMwWcJqrmdkQ=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/longpath","checksumSHA1":"EXiIm2xIL7Ds+YsQUx8Z3eUYPtI=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/mount","checksumSHA1":"IaLUlgL27e2W5LVWvncHgPWKffg=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/pools","checksumSHA1":"dj8atalGWftfM9vdzCsh9YF1Seg=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/stdcopy","checksumSHA1":"w0waeTRJ1sFygI0dZXH6l9E1c60=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/stringid","checksumSHA1":"gAUCA6R7F9kObegRGGNX5PzJahE=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/system","checksumSHA1":"W2vhTmDvG1eY4axzQjCjBVzo5Ms=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/tarsum","checksumSHA1":"I6mTgOFa7NeZpYw2S5342eenRLY=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/term","checksumSHA1":"WjvsjU1rtFjD3S0MmzKi5M08zjc=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/pkg/term/windows","checksumSHA1":"TeOtxuBbbZtp6wDK/t4DdaGGSC0=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/registry","checksumSHA1":"7YhsfoA07O7/TLou5q72Y/2sUsw=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/docker/registry/resumable","checksumSHA1":"jH7uQnDehFQygPP3zLC/mLSqgOk=","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, {"path":"github.com/docker/go-connections/nat","checksumSHA1":"JbiWTzH699Sqz25XmDlsARpMN9w=","revision":"7da10c8c50cad14494ec818dcdfb6506265c0086","revisionTime":"2017-02-03T23:56:24Z"}, - {"path":"github.com/docker/go-units","comment":"v0.1.0-23-g5d2041e","revision":"5d2041e26a699eaca682e2ea41c8f891e1060444"}, - {"path":"github.com/dustin/go-humanize","revision":"8929fe90cee4b2cb9deb468b51fb34eba64d1bf0"}, + {"path":"github.com/docker/go-connections/sockets","checksumSHA1":"jUfDG3VQsA2UZHvvIXncgiddpYA=","origin":"github.com/docker/docker/vendor/github.com/docker/go-connections/sockets","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/go-connections/tlsconfig","checksumSHA1":"zUG/J7nb6PWxfSXOoET6NePfyc0=","origin":"github.com/docker/docker/vendor/github.com/docker/go-connections/tlsconfig","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/docker/go-metrics","checksumSHA1":"kHVt4M5Pfby2dhurp+hZJfQhzVU=","revision":"399ea8c73916000c64c2c76e8da00ca82f8387ab","revisionTime":"2018-02-09T01:25:29Z"}, + {"path":"github.com/docker/go-units","checksumSHA1":"18hmvak2Dc9x5cgKeZ2iApviT7w=","comment":"v0.1.0-23-g5d2041e","revision":"5d2041e26a699eaca682e2ea41c8f891e1060444"}, + {"path":"github.com/docker/libnetwork/ipamutils","checksumSHA1":"vcP3kQNGWKHenrvQxfu4FZkB468=","origin":"github.com/docker/docker/vendor/github.com/docker/libnetwork/ipamutils","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/dustin/go-humanize","checksumSHA1":"xteP9Px90oMrg/HZuqvZkpXCR+s=","revision":"8929fe90cee4b2cb9deb468b51fb34eba64d1bf0"}, {"path":"github.com/elazarl/go-bindata-assetfs","checksumSHA1":"7DxViusFRJ7UPH0jZqYatwDrOkY=","revision":"30f82fa23fd844bd5bb1e5f216db87fd77b5eb43","revisionTime":"2017-02-27T21:27:28Z"}, {"path":"github.com/fatih/color","checksumSHA1":"VsE3zx2d8kpwj97TWhYddzAwBrY=","revision":"507f6050b8568533fb3f5504de8e5205fa62a114","revisionTime":"2018-02-13T13:34:03Z"}, {"path":"github.com/fatih/structs","checksumSHA1":"QBkOnLnM6zZ158NJSVLqoE4V6fI=","revision":"14f46232cd7bc732dc67313a9e4d3d210e082587","revisionTime":"2016-07-19T20:45:16Z"}, {"path":"github.com/fsouza/go-dockerclient","checksumSHA1":"E+wmSQrc/BXzgITrbNAbUzljoiM=","revision":"5ffdfff51ec0eba739b1039e65ba3625ef25f7c0","revisionTime":"2017-11-23T03:37:03Z"}, - {"path":"github.com/go-ini/ini","comment":"v1.8.5-2-g6ec4abd","revision":"6ec4abd8f8d587536da56f730858f0e27aeb4126"}, + {"path":"github.com/go-ini/ini","checksumSHA1":"U4k9IYSBQcW5DW5QDc44n5dddxs=","comment":"v1.8.5-2-g6ec4abd","revision":"6ec4abd8f8d587536da56f730858f0e27aeb4126"}, {"path":"github.com/go-ole/go-ole","checksumSHA1":"IvHj/4iR2nYa/S3cB2GXoyDG/xQ=","comment":"v1.2.0-4-g5005588","revision":"085abb85892dc1949567b726dff00fa226c60c45","revisionTime":"2017-07-12T17:44:59Z"}, - {"path":"github.com/go-ole/go-ole/oleutil","comment":"v1.2.0-4-g5005588","revision":"50055884d646dd9434f16bbb5c9801749b9bafe4"}, + {"path":"github.com/go-ole/go-ole/oleutil","checksumSHA1":"qLYVTQDhgrVIeZ2KI9eZV51mmug=","comment":"v1.2.0-4-g5005588","revision":"50055884d646dd9434f16bbb5c9801749b9bafe4"}, + {"path":"github.com/godbus/dbus","checksumSHA1":"bFplS7sPkJNtlKKCIszFQkAsmGI=","origin":"github.com/opencontainers/runc/vendor/github.com/godbus/dbus","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, {"path":"github.com/gogo/protobuf/proto","checksumSHA1":"I460dM/HmGE9DWimQvd1hJkYosU=","revision":"616a82ed12d78d24d4839363e8f3c5d3f20627cf","revisionTime":"2017-11-09T18:15:19Z"}, {"path":"github.com/golang/protobuf/proto","checksumSHA1":"Pyou8mceOASSFxc7GeXZuVdSMi0=","revision":"b4deda0973fb4c70b50d226b1af49f3da59f5265","revisionTime":"2018-04-30T18:52:41Z","version":"v1","versionExact":"v1.1.0"}, {"path":"github.com/golang/protobuf/ptypes","checksumSHA1":"/s0InJhSrxhTpqw5FIKgSMknCfM=","revision":"b4deda0973fb4c70b50d226b1af49f3da59f5265","revisionTime":"2018-04-30T18:52:41Z","version":"v1","versionExact":"v1.1.0"}, @@ -125,7 +143,9 @@ {"path":"github.com/google/go-cmp/cmp/internal/function","checksumSHA1":"kYtvRhMjM0X4bvEjR3pqEHLw1qo=","revision":"d5735f74713c51f7450a43d0a98d41ce2c1db3cb","revisionTime":"2017-09-01T21:42:48Z"}, {"path":"github.com/google/go-cmp/cmp/internal/value","checksumSHA1":"f+mgZLvc4VITtMmBv0bmew4rL2Y=","revision":"d5735f74713c51f7450a43d0a98d41ce2c1db3cb","revisionTime":"2017-09-01T21:42:48Z"}, {"path":"github.com/gorhill/cronexpr","checksumSHA1":"m8B3L3qJ3tFfP6BI9pIFr9oal3w=","comment":"1.0.0","origin":"github.com/dadgar/cronexpr","revision":"675cac9b2d182dccb5ba8d5f8a0d5988df8a4394","revisionTime":"2017-09-15T18:30:32Z"}, - {"path":"github.com/gorhill/cronexpr/cronexpr","comment":"1.0.0","revision":"a557574d6c024ed6e36acc8b610f5f211c91568a"}, + {"path":"github.com/gorhill/cronexpr/cronexpr","checksumSHA1":"Nd/7mZb0T6Gj6+AymyOPsNCQSJs=","comment":"1.0.0","revision":"a557574d6c024ed6e36acc8b610f5f211c91568a"}, + {"path":"github.com/gorilla/context","checksumSHA1":"g/V4qrXjUGG9B+e3hB+4NAYJ5Gs=","revision":"08b5f424b9271eedf6f9f0ce86cb9396ed337a42","revisionTime":"2016-08-17T18:46:32Z"}, + {"path":"github.com/gorilla/mux","checksumSHA1":"STQSdSj2FcpCf0NLfdsKhNutQT0=","revision":"e48e440e4c92e3251d812f8ce7858944dfa3331c","revisionTime":"2018-08-07T07:52:56Z"}, {"path":"github.com/hashicorp/consul-template/child","checksumSHA1":"Nu2j1GusM7ZH0uYrGzqr1K7yH7I=","revision":"26d029ad37335b3827a9fde5569b2c5e10dcac8f","revisionTime":"2017-10-31T14:25:17Z"}, {"path":"github.com/hashicorp/consul-template/config","checksumSHA1":"qKAxyhYnUpKzZ5KpA6aOiIHHqqg=","revision":"26d029ad37335b3827a9fde5569b2c5e10dcac8f","revisionTime":"2017-10-31T14:25:17Z"}, {"path":"github.com/hashicorp/consul-template/dependency","checksumSHA1":"gZUb/+jEn+2hdO/lmQSKcYuOB/o=","revision":"26d029ad37335b3827a9fde5569b2c5e10dcac8f","revisionTime":"2017-10-31T14:25:17Z"}, @@ -142,7 +162,7 @@ {"path":"github.com/hashicorp/consul/test/porter","checksumSHA1":"5XjgqE4UIfwXvkq5VssGNc7uPhQ=","revision":"ad9425ca6353b8afcfebd19130a8cf768f7eac30","revisionTime":"2017-10-21T00:05:25Z"}, {"path":"github.com/hashicorp/consul/testutil","checksumSHA1":"T4CeQD+QRsjf1BJ1n7FSojS5zDQ=","revision":"fb848fc48818f58690db09d14640513aa6bf3c02","revisionTime":"2018-04-13T17:05:42Z"}, {"path":"github.com/hashicorp/consul/testutil/retry","checksumSHA1":"SCb2b91UYiB/23+SNDBlU5OZfFA=","revision":"fb848fc48818f58690db09d14640513aa6bf3c02","revisionTime":"2018-04-13T17:05:42Z"}, - {"path":"github.com/hashicorp/errwrap","revision":"7554cd9344cec97297fa6649b055a8c98c2a1e55"}, + {"path":"github.com/hashicorp/errwrap","checksumSHA1":"cdOCt0Yb+hdErz8NAQqayxPmRsY=","revision":"7554cd9344cec97297fa6649b055a8c98c2a1e55"}, {"path":"github.com/hashicorp/go-checkpoint","checksumSHA1":"D267IUMW2rcb+vNe3QU+xhfSrgY=","revision":"1545e56e46dec3bba264e41fde2c1e2aa65b5dd4","revisionTime":"2017-10-09T17:35:28Z"}, {"path":"github.com/hashicorp/go-cleanhttp","checksumSHA1":"6ihdHMkDfFx/rJ1A36com2F6bQk=","revision":"a45970658e51fea2c41445ff0f7e07106d007617","revisionTime":"2017-02-11T00:33:01Z"}, {"path":"github.com/hashicorp/go-discover","checksumSHA1":"CnY2iYK47tGA9wsFMfH04PpmSoI=","revision":"40ccfdee6c0d7136f98f2b54882b86aaf0250d2f","revisionTime":"2018-05-03T15:30:45Z"}, @@ -161,19 +181,19 @@ {"path":"github.com/hashicorp/go-hclog","checksumSHA1":"dOP7kCX3dACHc9mU79826N411QA=","revision":"ff2cf002a8dd750586d91dddd4470c341f981fe1","revisionTime":"2018-07-09T16:53:50Z"}, {"path":"github.com/hashicorp/go-immutable-radix","checksumSHA1":"Cas2nprG6pWzf05A2F/OlnjUu2Y=","revision":"8aac2701530899b64bdea735a1de8da899815220","revisionTime":"2017-07-25T22:12:15Z"}, {"path":"github.com/hashicorp/go-memdb","checksumSHA1":"FMAvwDar2bQyYAW4XMFhAt0J5xA=","revision":"20ff6434c1cc49b80963d45bf5c6aa89c78d8d57","revisionTime":"2017-08-31T20:15:40Z"}, - {"path":"github.com/hashicorp/go-msgpack/codec","revision":"fa3f63826f7c23912c15263591e65d54d080b458"}, - {"path":"github.com/hashicorp/go-multierror","revision":"d30f09973e19c1dfcd120b2d9c4f168e68d6b5d5"}, + {"path":"github.com/hashicorp/go-msgpack/codec","checksumSHA1":"TNlVzNR1OaajcNi3CbQ3bGbaLGU=","revision":"fa3f63826f7c23912c15263591e65d54d080b458"}, + {"path":"github.com/hashicorp/go-multierror","checksumSHA1":"lrSl49G23l6NhfilxPM0XFs5rZo=","revision":"d30f09973e19c1dfcd120b2d9c4f168e68d6b5d5"}, {"path":"github.com/hashicorp/go-plugin","checksumSHA1":"lbG9uwM7qJlTIBg+8mjCC88sCPc=","revision":"e8d22c780116115ae5624720c9af0c97afe4f551","revisionTime":"2018-03-31T00:25:53Z"}, {"path":"github.com/hashicorp/go-retryablehttp","checksumSHA1":"/yKfFSspjuDzyOe/bBslrPzyfYM=","revision":"e651d75abec6fbd4f2c09508f72ae7af8a8b7171","revisionTime":"2018-07-18T19:50:05Z"}, {"path":"github.com/hashicorp/go-rootcerts","checksumSHA1":"A1PcINvF3UiwHRKn8UcgARgvGRs=","revision":"6bb64b370b90e7ef1fa532be9e591a81c3493e00","revisionTime":"2016-05-03T14:34:40Z"}, {"path":"github.com/hashicorp/go-safetemp","checksumSHA1":"CduvzBFfTv77nhjtXPGdIjQQLMI=","revision":"b1a1dbde6fdc11e3ae79efd9039009e22d4ae240","revisionTime":"2018-03-26T21:11:50Z"}, {"path":"github.com/hashicorp/go-sockaddr","checksumSHA1":"J47ySO1q0gcnmoMnir1q1loKzCk=","revision":"6d291a969b86c4b633730bfc6b8b9d64c3aafed9","revisionTime":"2018-03-20T11:50:54Z"}, {"path":"github.com/hashicorp/go-sockaddr/template","checksumSHA1":"PDp9DVLvf3KWxhs4G4DpIwauMSU=","revision":"6d291a969b86c4b633730bfc6b8b9d64c3aafed9","revisionTime":"2018-03-20T11:50:54Z"}, - {"path":"github.com/hashicorp/go-syslog","revision":"42a2b573b664dbf281bd48c3cc12c086b17a39ba"}, + {"path":"github.com/hashicorp/go-syslog","checksumSHA1":"xZ7Ban1x//6uUIU1xtrTbCYNHBc=","revision":"42a2b573b664dbf281bd48c3cc12c086b17a39ba"}, {"path":"github.com/hashicorp/go-uuid","checksumSHA1":"mAkPa/RLuIwN53GbwIEMATexams=","revision":"64130c7a86d732268a38cb04cfbaf0cc987fda98","revisionTime":"2016-07-17T02:21:40Z"}, {"path":"github.com/hashicorp/go-version","checksumSHA1":"tUGxc7rfX0cmhOOUDhMuAZ9rWsA=","revision":"03c5bf6be031b6dd45afec16b1cf94fc8938bc77","revisionTime":"2017-02-02T08:07:59Z"}, {"path":"github.com/hashicorp/golang-lru","checksumSHA1":"d9PxF1XQGLMJZRct2R8qVM/eYlE=","revision":"a0d98a5f288019575c6d1f4bb1573fef2d1fcdc4","revisionTime":"2016-02-07T21:47:19Z"}, - {"path":"github.com/hashicorp/golang-lru/simplelru","revision":"a0d98a5f288019575c6d1f4bb1573fef2d1fcdc4"}, + {"path":"github.com/hashicorp/golang-lru/simplelru","checksumSHA1":"2nOpYjx8Sn57bqlZq17yM4YJuM4=","revision":"a0d98a5f288019575c6d1f4bb1573fef2d1fcdc4"}, {"path":"github.com/hashicorp/hcl","checksumSHA1":"8OPDk+bKyRGJoKcS4QNw9F7dpE8=","revision":"6e968a3fcdcbab092f5307fd0d85479d5af1e4dc","revisionTime":"2016-11-01T18:00:25Z"}, {"path":"github.com/hashicorp/hcl/hcl/ast","checksumSHA1":"XQmjDva9JCGGkIecOgwtBEMCJhU=","revision":"6e968a3fcdcbab092f5307fd0d85479d5af1e4dc","revisionTime":"2016-11-01T18:00:25Z"}, {"path":"github.com/hashicorp/hcl/hcl/parser","checksumSHA1":"croNloscHsjX87X+4/cKOURf1EY=","revision":"6e968a3fcdcbab092f5307fd0d85479d5af1e4dc","revisionTime":"2016-11-01T18:00:25Z"}, @@ -190,9 +210,9 @@ {"path":"github.com/hashicorp/hcl2/hcl/json","checksumSHA1":"4Cr8I/nepYf4eRCl5hiazPf+afs=","revision":"6743a2254ba3d642b7d3a0be506259a0842819ac","revisionTime":"2018-08-10T01:10:00Z"}, {"path":"github.com/hashicorp/hcl2/hcldec","checksumSHA1":"wQ3hLj4s+5jN6LePSpT0XTTvdXA=","revision":"6743a2254ba3d642b7d3a0be506259a0842819ac","revisionTime":"2018-08-10T01:10:00Z"}, {"path":"github.com/hashicorp/hcl2/hclparse","checksumSHA1":"IzmftuG99BqNhbFGhxZaGwtiMtM=","revision":"6743a2254ba3d642b7d3a0be506259a0842819ac","revisionTime":"2018-08-10T01:10:00Z"}, - {"path":"github.com/hashicorp/logutils","revision":"0dc08b1671f34c4250ce212759ebd880f743d883"}, + {"path":"github.com/hashicorp/logutils","checksumSHA1":"vt+P9D2yWDO3gdvdgCzwqunlhxU=","revision":"0dc08b1671f34c4250ce212759ebd880f743d883"}, {"path":"github.com/hashicorp/memberlist","checksumSHA1":"1zk7IeGClUqBo+Phsx89p7fQ/rQ=","revision":"23ad4b7d7b38496cd64c241dfd4c60b7794c254a","revisionTime":"2017-02-08T21:15:06Z"}, - {"path":"github.com/hashicorp/net-rpc-msgpackrpc","revision":"a14192a58a694c123d8fe5481d4a4727d6ae82f3"}, + {"path":"github.com/hashicorp/net-rpc-msgpackrpc","checksumSHA1":"qnlqWJYV81ENr61SZk9c65R1mDo=","revision":"a14192a58a694c123d8fe5481d4a4727d6ae82f3"}, {"path":"github.com/hashicorp/raft","checksumSHA1":"zkA9uvbj1BdlveyqXpVTh1N6ers=","revision":"077966dbc90f342107eb723ec52fdb0463ec789b","revisionTime":"2018-01-17T20:29:25Z","version":"master","versionExact":"master"}, {"path":"github.com/hashicorp/raft-boltdb","checksumSHA1":"QAxukkv54/iIvLfsUP6IK4R0m/A=","revision":"d1e82c1ec3f15ee991f7cc7ffd5b67ff6f5bbaee","revisionTime":"2015-02-01T20:08:39Z"}, {"path":"github.com/hashicorp/serf/coordinate","checksumSHA1":"0PeWsO2aI+2PgVYlYlDPKfzCLEQ=","revision":"80ab48778deee28e4ea2dc4ef1ebb2c5f4063996","revisionTime":"2018-05-07T23:19:28Z"}, @@ -207,14 +227,14 @@ {"path":"github.com/hashicorp/yamux","checksumSHA1":"m9OKUPd/iliwKxs+LCSmAGpDJOs=","revision":"7221087c3d281fda5f794e28c2ea4c6e4d5c4558","revisionTime":"2018-09-17T20:50:41Z"}, {"path":"github.com/hpcloud/tail/util","checksumSHA1":"0xM336Lb25URO/1W1/CtGoRygVU=","revision":"37f4271387456dd1bf82ab1ad9229f060cc45386","revisionTime":"2017-08-14T16:06:53Z"}, {"path":"github.com/hpcloud/tail/watch","checksumSHA1":"TP4OAv5JMtzj2TB6OQBKqauaKDc=","revision":"37f4271387456dd1bf82ab1ad9229f060cc45386","revisionTime":"2017-08-14T16:06:53Z"}, - {"path":"github.com/jmespath/go-jmespath","comment":"0.2.2-2-gc01cf91","revision":"c01cf91b011868172fdcd9f41838e80c9d716264"}, + {"path":"github.com/jmespath/go-jmespath","checksumSHA1":"3/Bhy+ua/DCv2ElMD5GzOYSGN6g=","comment":"0.2.2-2-gc01cf91","revision":"c01cf91b011868172fdcd9f41838e80c9d716264"}, {"path":"github.com/kr/pretty","checksumSHA1":"eOXF2PEvYLMeD8DSzLZJWbjYzco=","revision":"cfb55aafdaf3ec08f0db22699ab822c50091b1c4","revisionTime":"2016-08-23T17:07:15Z"}, {"path":"github.com/kr/text","checksumSHA1":"uulQHQ7IsRKqDudBC8Go9J0gtAc=","revision":"7cafcd837844e784b526369c9bce262804aebc60","revisionTime":"2016-05-04T02:26:26Z"}, {"path":"github.com/mattn/go-colorable","checksumSHA1":"iRRp6PUlGXJgudZ3FD5jySI0Ukk=","revision":"efa589957cd060542a26d2dd7832fd6a6c6c3ade","revisionTime":"2018-03-10T13:32:14Z"}, {"path":"github.com/mattn/go-isatty","checksumSHA1":"AZO2VGorXTMDiSVUih3k73vORHY=","revision":"6ca4dbf54d38eea1a992b3c722a76a5d1c4cb25c","revisionTime":"2017-11-07T05:05:31Z"}, {"path":"github.com/mattn/go-shellwords","checksumSHA1":"ajImwVZHzsI+aNwsgzCSFSbmJqs=","revision":"f4e566c536cf69158e808ec28ef4182a37fdc981","revisionTime":"2015-03-21T17:42:21Z"}, {"path":"github.com/matttproud/golang_protobuf_extensions/pbutil","checksumSHA1":"bKMZjd2wPw13VwoE7mBeSv5djFA=","revision":"c12348ce28de40eed0136aa2b644d0ee0650e56c","revisionTime":"2016-04-24T11:30:07Z"}, - {"path":"github.com/miekg/dns","revision":"7e024ce8ce18b21b475ac6baf8fa3c42536bf2fa"}, + {"path":"github.com/miekg/dns","checksumSHA1":"7C76urB5PLSeqMeydxiUGjX5xMI=","revision":"7e024ce8ce18b21b475ac6baf8fa3c42536bf2fa"}, {"path":"github.com/mitchellh/cli","checksumSHA1":"+o0siVvR8q36mKCpT5F/Sn2T7xo=","revision":"c54c85e9bd492bdba226ffdda55d4e293b79f8e8","revisionTime":"2018-04-06T01:10:36Z"}, {"path":"github.com/mitchellh/colorstring","checksumSHA1":"ttEN1Aupb7xpPMkQLqb3tzLFdXs=","revision":"8631ce90f28644f54aeedcb3e389a85174e067d1","revisionTime":"2015-09-17T21:48:07Z"}, {"path":"github.com/mitchellh/copystructure","checksumSHA1":"+p4JY4wmFQAppCdlrJ8Kxybmht8=","revision":"d23ffcb85de31694d6ccaa23ccb4a03e55c1303f","revisionTime":"2017-05-25T01:39:02Z"}, @@ -222,10 +242,11 @@ {"path":"github.com/mitchellh/go-ps","checksumSHA1":"DcYIZnMiq/Tj22/ge9os3NwOhs4=","revision":"4fdf99ab29366514c69ccccddab5dc58b8d84062","revisionTime":"2017-03-09T13:30:38Z"}, {"path":"github.com/mitchellh/go-testing-interface","checksumSHA1":"bDdhmDk8q6utWrccBhEOa6IoGkE=","revision":"a61a99592b77c9ba629d254a693acffaeb4b7e28","revisionTime":"2017-10-04T22:19:16Z"}, {"path":"github.com/mitchellh/go-wordwrap","checksumSHA1":"L3leymg2RT8hFl5uL+5KP/LpBkg=","revision":"ad45545899c7b13c020ea92b2072220eefad42b8","revisionTime":"2015-03-14T17:03:34Z"}, - {"path":"github.com/mitchellh/hashstructure","revision":"1ef5c71b025aef149d12346356ac5973992860bc"}, - {"path":"github.com/mitchellh/mapstructure","revision":"281073eb9eb092240d33ef253c404f1cca550309"}, + {"path":"github.com/mitchellh/hashstructure","checksumSHA1":"Z3FoiV93oUfDoQYMMiHxWCQPlBw=","revision":"1ef5c71b025aef149d12346356ac5973992860bc"}, + {"path":"github.com/mitchellh/mapstructure","checksumSHA1":"4Js6Jlu93Wa0o6Kjt393L9Z7diE=","revision":"281073eb9eb092240d33ef253c404f1cca550309"}, {"path":"github.com/mitchellh/reflectwalk","checksumSHA1":"KqsMqI+Y+3EFYPhyzafpIneaVCM=","revision":"8d802ff4ae93611b807597f639c19f76074df5c6","revisionTime":"2017-05-08T17:38:06Z"}, {"path":"github.com/moby/moby/daemon/caps","checksumSHA1":"FoDTHct8ocl470GYc0i+JRWfrys=","revision":"39377bb96d459d2ef59bd2bad75468638a7f86a3","revisionTime":"2018-01-18T19:02:33Z"}, + {"path":"github.com/mrunalp/fileutils","checksumSHA1":"EKGlMEHq/nwBXQGi9JN/y+H7YMU=","origin":"github.com/opencontainers/runc/vendor/github.com/mrunalp/fileutils","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, {"path":"github.com/oklog/run","checksumSHA1":"nf3UoPNBIut7BL9nWE8Fw2X2j+Q=","revision":"6934b124db28979da51d3470dadfa34d73d72652","revisionTime":"2018-03-08T00:51:04Z"}, {"path":"github.com/onsi/ginkgo","checksumSHA1":"cwbidLG1ET7YSqlwca+nSfYxIbg=","revision":"ba8e856bb854d6771a72ddf6497a42dad3a0c971","revisionTime":"2018-03-12T10:34:14Z"}, {"path":"github.com/onsi/ginkgo/config","checksumSHA1":"Tarhbqac6rFsGPugPoQ4lyhfc7Q=","revision":"9008c7b79f9636c46a0a945141020124702f0ecf","revisionTime":"2018-02-16T17:00:43Z"}, @@ -267,12 +288,25 @@ {"path":"github.com/opencontainers/go-digest","checksumSHA1":"NTperEHVh1uBqfTy9+oKceN4tKI=","revision":"21dfd564fd89c944783d00d069f33e3e7123c448","revisionTime":"2017-01-11T18:16:59Z"}, {"path":"github.com/opencontainers/image-spec/specs-go","checksumSHA1":"ZGlIwSRjdLYCUII7JLE++N4w7Xc=","revision":"89b51c794e9113108a2914e38e66c826a649f2b5","revisionTime":"2017-11-03T11:36:04Z"}, {"path":"github.com/opencontainers/image-spec/specs-go/v1","checksumSHA1":"jdbXRRzeu0njLE9/nCEZG+Yg/Jk=","revision":"89b51c794e9113108a2914e38e66c826a649f2b5","revisionTime":"2017-11-03T11:36:04Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/cgroups","checksumSHA1":"UPfYFznGeLEloCkGG8KKgsKjtPU=","comment":"v0.0.9-108-g89ab7f2","revision":"89ab7f2ccc1e45ddf6485eaa802c35dcf321dfc8","revisionTime":"2016-03-31T09:02:02Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/cgroups/fs","checksumSHA1":"xIrNApwB9+yQBhu2p6bt7blV+x4=","comment":"v0.0.9-108-g89ab7f2","revision":"89ab7f2ccc1e45ddf6485eaa802c35dcf321dfc8","revisionTime":"2016-03-31T09:02:02Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/configs","checksumSHA1":"CdhRZZmDeY+t3PN4Qpi1ojy50pE=","comment":"v0.0.9-108-g89ab7f2","revision":"89ab7f2ccc1e45ddf6485eaa802c35dcf321dfc8","revisionTime":"2016-03-31T09:02:02Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/system","checksumSHA1":"xo2LfoL27cGF6cChWvew2KfhpNI=","comment":"v0.0.9-108-g89ab7f2","revision":"89ab7f2ccc1e45ddf6485eaa802c35dcf321dfc8","revisionTime":"2016-03-31T09:02:02Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/user","checksumSHA1":"3AoPMXlmVq2+iWMpsdJZkcUKHB8=","revision":"89ab7f2ccc1e45ddf6485eaa802c35dcf321dfc8","revisionTime":"2016-03-31T09:02:02Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/utils","checksumSHA1":"PCfoPliJrUhzLP4arNs2KjEb2Bc=","comment":"v0.0.9-108-g89ab7f2","revision":"89ab7f2ccc1e45ddf6485eaa802c35dcf321dfc8","revisionTime":"2016-03-31T09:02:02Z"}, + {"path":"github.com/opencontainers/runc/libcontainer","checksumSHA1":"cqrf+HmlptrK2lRhQ0PkYdOe5cQ=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/apparmor","checksumSHA1":"gVVY8k2G3ws+V1czsfxfuRs8log=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/cgroups","checksumSHA1":"eA9qNw7Tkpi1GLojT/Vxa99aL00=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/cgroups/fs","checksumSHA1":"YgPihDRi3OxI0qv7CxTxrqZuvfU=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/cgroups/systemd","checksumSHA1":"RVQixM4pF56oCNymhNY67I5eS0Y=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/configs","checksumSHA1":"PUv5rdj6oEGJoBij/9Elx8VO6bs=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/configs/validate","checksumSHA1":"YJq+f9izqizSzG/OuVFUOfloNEk=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/criurpc","checksumSHA1":"I1iwXoDUJeDXviilCtkvDpEF/So=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/intelrdt","checksumSHA1":"bAWJX1XUDMd4GqPLSrCkUcdiTbg=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/keys","checksumSHA1":"QXuHZwxlqhoq/oHRJFbTi6+AWLY=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/mount","checksumSHA1":"MJiogPDUU2nFr1fzQU6T+Ry1W8o=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/seccomp","checksumSHA1":"I1Qw/btE1twMqKHpYNsC98cteak=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/stacktrace","checksumSHA1":"yp/kYBgVqKtxlnpq4CmyxLFMAE4=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/system","checksumSHA1":"cjg/UcueM1/2/ExZ3N7010sa+hI=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/user","checksumSHA1":"XtLpcP6ca9SQG218re7E7UcOj3Y=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer/utils","checksumSHA1":"Sb296YW4V+esieanrx4Nzt2L5lU=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runtime-spec/specs-go","checksumSHA1":"AMYc2X2O/IL6EGrq6lTl5vEhLiY=","origin":"github.com/opencontainers/runc/vendor/github.com/opencontainers/runtime-spec/specs-go","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/selinux/go-selinux","checksumSHA1":"j9efF9bPmCCag+LzqwjyB8a44B8=","origin":"github.com/opencontainers/runc/vendor/github.com/opencontainers/selinux/go-selinux","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/selinux/go-selinux/label","checksumSHA1":"QbeVoKIoaJWZDH8V/588i8/Pjjs=","origin":"github.com/opencontainers/runc/vendor/github.com/opencontainers/selinux/go-selinux/label","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, {"path":"github.com/pkg/errors","checksumSHA1":"ynJSWoF6v+3zMnh9R0QmmG6iGV8=","revision":"248dadf4e9068a0b3e79f02ed0a610d935de5302","revisionTime":"2016-10-29T09:36:37Z"}, {"path":"github.com/pmezard/go-difflib/difflib","checksumSHA1":"LuFv4/jlrmFNnDb/5SCSEPAM9vU=","revision":"792786c7400a136282c1664665ae0a8db921c6c2","revisionTime":"2016-01-10T10:55:54Z"}, {"path":"github.com/posener/complete","checksumSHA1":"rTNABfFJ9wtLQRH8uYNkEZGQOrY=","revision":"9f41f7636a724791a3b8b1d35e84caa1124f0d3c","revisionTime":"2017-08-29T17:11:12Z"}, @@ -293,6 +327,7 @@ {"path":"github.com/ryanuber/columnize","checksumSHA1":"M57Rrfc8Z966p+IBtQ91QOcUtcg=","comment":"v2.0.1-8-g983d3a5","revision":"abc90934186a77966e2beeac62ed966aac0561d5","revisionTime":"2017-07-03T20:58:27Z"}, {"path":"github.com/ryanuber/go-glob","checksumSHA1":"6JP37UqrI0H80Gpk0Y2P+KXgn5M=","revision":"256dc444b735e061061cf46c809487313d5b0065","revisionTime":"2017-01-28T01:21:29Z"}, {"path":"github.com/sean-/seed","checksumSHA1":"tnMZLo/kR9Kqx6GtmWwowtTLlA8=","revision":"e2103e2c35297fb7e17febb81e49b312087a2372","revisionTime":"2017-03-13T16:33:22Z"}, + {"path":"github.com/seccomp/libseccomp-golang","checksumSHA1":"6Z/chtTVA74eUZTlG5VRDy59K1M=","origin":"github.com/opencontainers/runc/vendor/github.com/seccomp/libseccomp-golang","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, {"path":"github.com/sethgrid/pester","checksumSHA1":"8Lm8nsMCFz4+gr9EvQLqK8+w+Ks=","revision":"8053687f99650573b28fb75cddf3f295082704d7","revisionTime":"2016-04-29T17:20:22Z"}, {"path":"github.com/shirou/gopsutil/cpu","checksumSHA1":"k+PmW/6PFt0FVFTTnfMbWwrm9hU=","revision":"5776ff9c7c5d063d574ef53d740f75c68b448e53","revisionTime":"2018-02-27T22:58:47Z","tree":true}, {"path":"github.com/shirou/gopsutil/disk","checksumSHA1":"4DZwA8Xf2Zs8vhIc9kx8TIBMmSY=","revision":"5776ff9c7c5d063d574ef53d740f75c68b448e53","revisionTime":"2018-02-27T22:58:47Z","tree":true}, @@ -302,6 +337,7 @@ {"path":"github.com/shirou/gopsutil/net","checksumSHA1":"ME2P9hiaHO/YdVrNInDmb/dB6us=","revision":"5776ff9c7c5d063d574ef53d740f75c68b448e53","revisionTime":"2018-02-27T22:58:47Z","tree":true}, {"path":"github.com/shirou/gopsutil/process","checksumSHA1":"JuBAKUuSyR7RdJELt6tQMn79Y6w=","revision":"5776ff9c7c5d063d574ef53d740f75c68b448e53","revisionTime":"2018-02-27T22:58:47Z","tree":true}, {"path":"github.com/shirou/w32","checksumSHA1":"Nve7SpDmjsv6+rhkXAkfg/UQx94=","revision":"bb4de0191aa41b5507caa14b0650cdbddcd9280b","revisionTime":"2016-09-30T03:27:40Z"}, + {"path":"github.com/sirupsen/logrus","checksumSHA1":"KXUHFmNxt6vTQkUJ00c+M/+dV58=","revision":"51df1d314861d341546c26e4354ce760fe4bb02b","revisionTime":"2018-08-27T05:22:11Z"}, {"path":"github.com/skratchdot/open-golang/open","checksumSHA1":"h/HMhokbQHTdLUbruoBBTee+NYw=","revision":"75fb7ed4208cf72d323d7d02fd1a5964a7a9073c","revisionTime":"2016-03-02T14:40:31Z"}, {"path":"github.com/spf13/pflag","checksumSHA1":"Q52Y7t0lEtk/wcDn5q7tS7B+jqs=","revision":"7aff26db30c1be810f9de5038ec5ef96ac41fd7c","revisionTime":"2017-08-24T17:57:12Z"}, {"path":"github.com/stretchr/objx","checksumSHA1":"K0crHygPTP42i1nLKWphSlvOQJw=","revision":"1a9d0bb9f541897e62256577b352fdbc1fb4fd94","revisionTime":"2015-09-28T12:21:52Z"}, @@ -313,10 +349,12 @@ {"path":"github.com/tv42/httpunix","checksumSHA1":"2xcr/mhxBFlDjpxe/Mc2Wb4RGR8=","revision":"b75d8614f926c077e48d85f1f8f7885b758c6225","revisionTime":"2015-04-27T01:28:21Z"}, {"path":"github.com/ugorji/go/codec","checksumSHA1":"8G1zvpE4gTtWQRuP/x2HPVDmflo=","revision":"0053ebfd9d0ee06ccefbfe17072021e1d4acebee","revisionTime":"2017-06-20T06:01:02Z"}, {"path":"github.com/ugorji/go/codec/codecgen","checksumSHA1":"OgParimNuU2CJqr3pcTympeQZUc=","revision":"5efa3251c7f7d05e5d9704a69a984ec9f1386a40","revisionTime":"2017-06-20T10:48:52Z"}, - {"path":"github.com/ulikunitz/xz","checksumSHA1":"z2kAtVle4NFV2OExI85fZoTcsI4=","revision":"0c6b41e72360850ca4f98dc341fd999726ea007f","revisionTime":"2017-06-05T21:53:11Z"}, + {"path":"github.com/ulikunitz/xz","checksumSHA1":"qgMa75aMGbkFY0jIqqqgVnCUoNA=","revision":"0c6b41e72360850ca4f98dc341fd999726ea007f","revisionTime":"2017-06-05T21:53:11Z"}, {"path":"github.com/ulikunitz/xz/internal/hash","checksumSHA1":"vjnTkzNrMs5Xj6so/fq0mQ6dT1c=","revision":"0c6b41e72360850ca4f98dc341fd999726ea007f","revisionTime":"2017-06-05T21:53:11Z"}, {"path":"github.com/ulikunitz/xz/internal/xlog","checksumSHA1":"m0pm57ASBK/CTdmC0ppRHO17mBs=","revision":"0c6b41e72360850ca4f98dc341fd999726ea007f","revisionTime":"2017-06-05T21:53:11Z"}, {"path":"github.com/ulikunitz/xz/lzma","checksumSHA1":"2vZw6zc8xuNlyVz2QKvdlNSZQ1U=","revision":"0c6b41e72360850ca4f98dc341fd999726ea007f","revisionTime":"2017-06-05T21:53:11Z"}, + {"path":"github.com/vishvananda/netlink","checksumSHA1":"cIkE6EIE7A0IzdhR/Yes8Nzyqtk=","origin":"github.com/opencontainers/runc/vendor/github.com/vishvananda/netlink","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/vishvananda/netlink/nl","checksumSHA1":"r/vcO8YkOWNHKX5HKCukaU4Xzlg=","origin":"github.com/opencontainers/runc/vendor/github.com/vishvananda/netlink/nl","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, {"path":"github.com/vmihailenco/msgpack","checksumSHA1":"t9A/EE2GhHFPHzK+ksAKgKW9ZC8=","revision":"b5e691b1eb52a28c05e67ab9df303626c095c23b","revisionTime":"2018-06-13T09:15:15Z"}, {"path":"github.com/vmihailenco/msgpack/codes","checksumSHA1":"OcTSGT2v7/2saIGq06nDhEZwm8I=","revision":"b5e691b1eb52a28c05e67ab9df303626c095c23b","revisionTime":"2018-06-13T09:15:15Z"}, {"path":"github.com/zclconf/go-cty/cty","checksumSHA1":"Ej+3WWvyjn0xg3aujsyT+yvvmdc=","revision":"02bd58e97b5759d478019c5a6333edbfdfed16a0","revisionTime":"2018-07-18T22:05:26Z"}, @@ -340,10 +378,11 @@ {"path":"golang.org/x/net/idna","checksumSHA1":"g/Z/Ka0VgJESgQK7/SJCjm/aX0I=","revision":"ab5485076ff3407ad2d02db054635913f017b0ed","revisionTime":"2017-07-19T21:11:51Z"}, {"path":"golang.org/x/net/internal/timeseries","checksumSHA1":"UxahDzW2v4mf/+aFxruuupaoIwo=","revision":"ab5485076ff3407ad2d02db054635913f017b0ed","revisionTime":"2017-07-19T21:11:51Z"}, {"path":"golang.org/x/net/lex/httplex","checksumSHA1":"3xyuaSNmClqG4YWC7g0isQIbUTc=","revision":"ab5485076ff3407ad2d02db054635913f017b0ed","revisionTime":"2017-07-19T21:11:51Z"}, + {"path":"golang.org/x/net/proxy","checksumSHA1":"r9l4r3H6FOLQ0c2JaoXpopFjpnw=","origin":"github.com/docker/docker/vendor/golang.org/x/net/proxy","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, {"path":"golang.org/x/net/trace","checksumSHA1":"u/r66lwYfgg682u5hZG7/E7+VCY=","revision":"ab5485076ff3407ad2d02db054635913f017b0ed","revisionTime":"2017-07-19T21:11:51Z"}, {"path":"golang.org/x/sync/errgroup","checksumSHA1":"S0DP7Pn7sZUmXc55IzZnNvERu6s=","revision":"316e794f7b5e3df4e95175a45a5fb8b12f85cb4f","revisionTime":"2016-07-15T18:54:39Z"}, {"path":"golang.org/x/sys/cpu","checksumSHA1":"REkmyB368pIiip76LiqMLspgCRk=","revision":"1b2967e3c290b7c545b3db0deeda16e9be4f98a2","revisionTime":"2018-07-06T09:56:39Z"}, - {"path":"golang.org/x/sys/unix","checksumSHA1":"O2WxXeMRgYFsVABlG1SANNjAJE8=","revision":"ac767d655b305d4e9612f5f6e33120b9176c4ad4","revisionTime":"2018-07-08T03:57:06Z"}, + {"path":"golang.org/x/sys/unix","checksumSHA1":"su2QDjUzrUO0JnOH9m0cNg0QqsM=","revision":"ac767d655b305d4e9612f5f6e33120b9176c4ad4","revisionTime":"2018-07-08T03:57:06Z"}, {"path":"golang.org/x/sys/windows","checksumSHA1":"l1jIhK9Y33obLipDvmjVPCdYtJI=","revision":"1b2967e3c290b7c545b3db0deeda16e9be4f98a2","revisionTime":"2018-07-06T09:56:39Z"}, {"path":"golang.org/x/text/encoding","checksumSHA1":"Mr4ur60bgQJnQFfJY0dGtwWwMPE=","revision":"e113a52b01bdd1744681b6ce70c2e3d26b58d389","revisionTime":"2017-08-30T18:54:29Z"}, {"path":"golang.org/x/text/encoding/charmap","checksumSHA1":"HgcUFTOQF5jOYtTIj5obR3GVN9A=","revision":"e113a52b01bdd1744681b6ce70c2e3d26b58d389","revisionTime":"2017-08-30T18:54:29Z"}, From 8504505c0d2c5ccb6725e4a76c9e1d3338cbf754 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Sat, 15 Sep 2018 17:48:59 -0700 Subject: [PATCH 064/152] client uses passed logger and fix fingerprinters --- client/client.go | 11 ++-------- client/client_test.go | 2 +- client/fingerprint/arch.go | 8 +++---- client/fingerprint/arch_test.go | 2 +- client/fingerprint/cgroup.go | 8 +++---- client/fingerprint/cgroup_linux.go | 4 ++-- client/fingerprint/consul.go | 25 +++++++++++----------- client/fingerprint/consul_test.go | 4 ++-- client/fingerprint/cpu.go | 16 +++++++------- client/fingerprint/cpu_test.go | 4 ++-- client/fingerprint/env_aws.go | 23 ++++++++++---------- client/fingerprint/env_aws_test.go | 10 ++++----- client/fingerprint/env_gce.go | 34 +++++++++++++++++------------- client/fingerprint/env_gce_test.go | 4 ++-- client/fingerprint/fingerprint.go | 6 +++--- client/fingerprint/host.go | 10 ++++----- client/fingerprint/host_test.go | 2 +- client/fingerprint/memory.go | 10 ++++----- client/fingerprint/memory_test.go | 4 ++-- client/fingerprint/network.go | 18 ++++++++-------- client/fingerprint/network_test.go | 12 +++++------ client/fingerprint/nomad.go | 9 ++++---- client/fingerprint/nomad_test.go | 2 +- client/fingerprint/signal.go | 8 +++---- client/fingerprint/signal_test.go | 2 +- client/fingerprint/storage.go | 8 +++---- client/fingerprint/storage_test.go | 2 +- client/fingerprint/vault.go | 12 +++++------ client/fingerprint/vault_test.go | 2 +- client/fingerprint_manager.go | 22 ++++++------------- client/fingerprint_manager_test.go | 28 ++++++++++++------------ client/fs_endpoint_test.go | 2 +- 32 files changed, 151 insertions(+), 163 deletions(-) diff --git a/client/client.go b/client/client.go index afca7d1a8..eef3f4e16 100644 --- a/client/client.go +++ b/client/client.go @@ -229,7 +229,7 @@ func NewClient(cfg *config.Config, consulCatalog consul.CatalogAPI, consulServic } // Create the logger - logger := cfg.Logger.Named("client") + logger := cfg.Logger.ResetNamed("client") // Create the client c := &Client{ @@ -1922,17 +1922,10 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error // Copy the config since the node can be swapped out as it is being updated. // The long term fix is to pass in the config and node separately and then // we don't have to do a copy. - //XXX FIXME create a root logger - logger := hclog.New(&hclog.LoggerOptions{ - Name: "nomad", - Level: hclog.LevelFromString(c.configCopy.LogLevel), - TimeFormat: time.RFC3339, - }) - c.configLock.RLock() arConf := &allocrunnerv2.Config{ Alloc: alloc, - Logger: logger, + Logger: c.logger, ClientConfig: c.config, StateDB: c.stateDB, Consul: c.consulService, diff --git a/client/client_test.go b/client/client_test.go index 06d1afcb7..a8a6c6df4 100644 --- a/client/client_test.go +++ b/client/client_test.go @@ -650,7 +650,7 @@ func TestClient_Init(t *testing.T) { config: &config.Config{ AllocDir: allocDir, }, - logger: testlog.Logger(t), + logger: testlog.HCLogger(t), } if err := client.init(); err != nil { t.Fatalf("err: %s", err) diff --git a/client/fingerprint/arch.go b/client/fingerprint/arch.go index 3277822bc..7f2083f5a 100644 --- a/client/fingerprint/arch.go +++ b/client/fingerprint/arch.go @@ -1,21 +1,21 @@ package fingerprint import ( - "log" "runtime" + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" ) // ArchFingerprint is used to fingerprint the architecture type ArchFingerprint struct { StaticFingerprinter - logger *log.Logger + logger log.Logger } // NewArchFingerprint is used to create an OS fingerprint -func NewArchFingerprint(logger *log.Logger) Fingerprint { - f := &ArchFingerprint{logger: logger} +func NewArchFingerprint(logger log.Logger) Fingerprint { + f := &ArchFingerprint{logger: logger.Named("arch")} return f } diff --git a/client/fingerprint/arch_test.go b/client/fingerprint/arch_test.go index 58c600b6a..3d23b7a84 100644 --- a/client/fingerprint/arch_test.go +++ b/client/fingerprint/arch_test.go @@ -10,7 +10,7 @@ import ( ) func TestArchFingerprint(t *testing.T) { - f := NewArchFingerprint(testlog.Logger(t)) + f := NewArchFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } diff --git a/client/fingerprint/cgroup.go b/client/fingerprint/cgroup.go index d6affbc59..e399e7d08 100644 --- a/client/fingerprint/cgroup.go +++ b/client/fingerprint/cgroup.go @@ -1,9 +1,9 @@ package fingerprint import ( - "log" "time" + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" ) @@ -13,7 +13,7 @@ const ( ) type CGroupFingerprint struct { - logger *log.Logger + logger log.Logger lastState string mountPointDetector MountPointDetector } @@ -35,9 +35,9 @@ func (b *DefaultMountPointDetector) MountPoint() (string, error) { } // NewCGroupFingerprint returns a new cgroup fingerprinter -func NewCGroupFingerprint(logger *log.Logger) Fingerprint { +func NewCGroupFingerprint(logger log.Logger) Fingerprint { f := &CGroupFingerprint{ - logger: logger, + logger: logger.Named("cgroup"), lastState: cgroupUnavailable, mountPointDetector: &DefaultMountPointDetector{}, } diff --git a/client/fingerprint/cgroup_linux.go b/client/fingerprint/cgroup_linux.go index c5eed8f02..467904ebd 100644 --- a/client/fingerprint/cgroup_linux.go +++ b/client/fingerprint/cgroup_linux.go @@ -44,7 +44,7 @@ func (f *CGroupFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp * f.clearCGroupAttributes(resp) if f.lastState == cgroupAvailable { - f.logger.Printf("[INFO] fingerprint.cgroups: cgroups are unavailable") + f.logger.Info("cgroups are unavailable") } f.lastState = cgroupUnavailable return nil @@ -54,7 +54,7 @@ func (f *CGroupFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp * resp.Detected = true if f.lastState == cgroupUnavailable { - f.logger.Printf("[INFO] fingerprint.cgroups: cgroups are available") + f.logger.Info("cgroups are available") } f.lastState = cgroupAvailable return nil diff --git a/client/fingerprint/consul.go b/client/fingerprint/consul.go index 84c6a97d6..66e86a86e 100644 --- a/client/fingerprint/consul.go +++ b/client/fingerprint/consul.go @@ -2,12 +2,11 @@ package fingerprint import ( "fmt" - "log" "strconv" "time" consul "github.com/hashicorp/consul/api" - + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" ) @@ -18,14 +17,14 @@ const ( // ConsulFingerprint is used to fingerprint for Consul type ConsulFingerprint struct { - logger *log.Logger + logger log.Logger client *consul.Client lastState string } // NewConsulFingerprint is used to create a Consul fingerprint -func NewConsulFingerprint(logger *log.Logger) Fingerprint { - return &ConsulFingerprint{logger: logger, lastState: consulUnavailable} +func NewConsulFingerprint(logger log.Logger) Fingerprint { + return &ConsulFingerprint{logger: logger.Named("consul"), lastState: consulUnavailable} } func (f *ConsulFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp *cstructs.FingerprintResponse) error { @@ -52,7 +51,7 @@ func (f *ConsulFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp * // Print a message indicating that the Consul Agent is not available // anymore if f.lastState == consulAvailable { - f.logger.Printf("[INFO] fingerprint.consul: consul agent is unavailable") + f.logger.Info("consul agent is unavailable") } f.lastState = consulUnavailable return nil @@ -61,27 +60,27 @@ func (f *ConsulFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp * if s, ok := info["Config"]["Server"].(bool); ok { resp.AddAttribute("consul.server", strconv.FormatBool(s)) } else { - f.logger.Printf("[WARN] fingerprint.consul: unable to fingerprint consul.server") + f.logger.Warn("unable to fingerprint consul.server") } if v, ok := info["Config"]["Version"].(string); ok { resp.AddAttribute("consul.version", v) } else { - f.logger.Printf("[WARN] fingerprint.consul: unable to fingerprint consul.version") + f.logger.Warn("unable to fingerprint consul.version") } if r, ok := info["Config"]["Revision"].(string); ok { resp.AddAttribute("consul.revision", r) } else { - f.logger.Printf("[WARN] fingerprint.consul: unable to fingerprint consul.revision") + f.logger.Warn("unable to fingerprint consul.revision") } if n, ok := info["Config"]["NodeName"].(string); ok { resp.AddAttribute("unique.consul.name", n) } else { - f.logger.Printf("[WARN] fingerprint.consul: unable to fingerprint unique.consul.name") + f.logger.Warn("unable to fingerprint unique.consul.name") } if d, ok := info["Config"]["Datacenter"].(string); ok { resp.AddAttribute("consul.datacenter", d) } else { - f.logger.Printf("[WARN] fingerprint.consul: unable to fingerprint consul.datacenter") + f.logger.Warn("unable to fingerprint consul.datacenter") } if dc, ok := resp.Attributes["consul.datacenter"]; ok { @@ -89,13 +88,13 @@ func (f *ConsulFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp * resp.AddLink("consul", fmt.Sprintf("%s.%s", dc, name)) } } else { - f.logger.Printf("[WARN] fingerprint.consul: malformed Consul response prevented linking") + f.logger.Warn("malformed Consul response prevented linking") } // If the Consul Agent was previously unavailable print a message to // indicate the Agent is available now if f.lastState == consulUnavailable { - f.logger.Printf("[INFO] fingerprint.consul: consul agent is available") + f.logger.Info("consul agent is available") } f.lastState = consulAvailable resp.Detected = true diff --git a/client/fingerprint/consul_test.go b/client/fingerprint/consul_test.go index a96c73b07..c260fd493 100644 --- a/client/fingerprint/consul_test.go +++ b/client/fingerprint/consul_test.go @@ -15,7 +15,7 @@ import ( ) func TestConsulFingerprint(t *testing.T) { - fp := NewConsulFingerprint(testlog.Logger(t)) + fp := NewConsulFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } @@ -171,7 +171,7 @@ const mockConsulResponse = ` // See https://github.com/hashicorp/nomad/issues/3326 func TestConsulFingerprint_UnexpectedResponse(t *testing.T) { assert := assert.New(t) - fp := NewConsulFingerprint(testlog.Logger(t)) + fp := NewConsulFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } diff --git a/client/fingerprint/cpu.go b/client/fingerprint/cpu.go index e317b7e06..68f01589f 100644 --- a/client/fingerprint/cpu.go +++ b/client/fingerprint/cpu.go @@ -2,8 +2,8 @@ package fingerprint import ( "fmt" - "log" + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/helper/stats" "github.com/hashicorp/nomad/nomad/structs" @@ -12,12 +12,12 @@ import ( // CPUFingerprint is used to fingerprint the CPU type CPUFingerprint struct { StaticFingerprinter - logger *log.Logger + logger log.Logger } // NewCPUFingerprint is used to create a CPU fingerprint -func NewCPUFingerprint(logger *log.Logger) Fingerprint { - f := &CPUFingerprint{logger: logger} +func NewCPUFingerprint(logger log.Logger) Fingerprint { + f := &CPUFingerprint{logger: logger.Named("cpu")} return f } @@ -37,7 +37,7 @@ func (f *CPUFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp *cst } if err := stats.Init(); err != nil { - f.logger.Printf("[WARN] fingerprint.cpu: %v", err) + f.logger.Warn("failed initializing stats collector", "error", err) } if cfg.CpuCompute != 0 { @@ -51,17 +51,17 @@ func (f *CPUFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp *cst if mhz := stats.CPUMHzPerCore(); mhz > 0 { resp.AddAttribute("cpu.frequency", fmt.Sprintf("%.0f", mhz)) - f.logger.Printf("[DEBUG] fingerprint.cpu: frequency: %.0f MHz", mhz) + f.logger.Debug("detected cpu frequency", "MHz", log.Fmt("%.0f", mhz)) } if numCores := stats.CPUNumCores(); numCores > 0 { resp.AddAttribute("cpu.numcores", fmt.Sprintf("%d", numCores)) - f.logger.Printf("[DEBUG] fingerprint.cpu: core count: %d", numCores) + f.logger.Debug("detected core count", "cores", numCores) } tt := int(stats.TotalTicksAvailable()) if cfg.CpuCompute > 0 { - f.logger.Printf("[DEBUG] fingerprint.cpu: Using specified cpu compute %d", cfg.CpuCompute) + f.logger.Debug("using user specified cpu compute", "cpu_compute", cfg.CpuCompute) tt = cfg.CpuCompute } diff --git a/client/fingerprint/cpu_test.go b/client/fingerprint/cpu_test.go index 1088ab95e..81862d08f 100644 --- a/client/fingerprint/cpu_test.go +++ b/client/fingerprint/cpu_test.go @@ -10,7 +10,7 @@ import ( ) func TestCPUFingerprint(t *testing.T) { - f := NewCPUFingerprint(testlog.Logger(t)) + f := NewCPUFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } @@ -58,7 +58,7 @@ func TestCPUFingerprint(t *testing.T) { // TestCPUFingerprint_OverrideCompute asserts that setting cpu_total_compute in // the client config overrides the detected CPU freq (if any). func TestCPUFingerprint_OverrideCompute(t *testing.T) { - f := NewCPUFingerprint(testlog.Logger(t)) + f := NewCPUFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } diff --git a/client/fingerprint/env_aws.go b/client/fingerprint/env_aws.go index 94274a454..6b1749303 100644 --- a/client/fingerprint/env_aws.go +++ b/client/fingerprint/env_aws.go @@ -3,7 +3,6 @@ package fingerprint import ( "fmt" "io/ioutil" - "log" "net/http" "net/url" "os" @@ -11,8 +10,10 @@ import ( "strings" "time" - "github.com/hashicorp/go-cleanhttp" + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" + + "github.com/hashicorp/go-cleanhttp" "github.com/hashicorp/nomad/nomad/structs" ) @@ -51,13 +52,13 @@ var ec2InstanceSpeedMap = map[*regexp.Regexp]int{ type EnvAWSFingerprint struct { StaticFingerprinter timeout time.Duration - logger *log.Logger + logger log.Logger } // NewEnvAWSFingerprint is used to create a fingerprint from AWS metadata -func NewEnvAWSFingerprint(logger *log.Logger) Fingerprint { +func NewEnvAWSFingerprint(logger log.Logger) Fingerprint { f := &EnvAWSFingerprint{ - logger: logger, + logger: logger.Named("env_aws"), timeout: AwsMetadataTimeout, } return f @@ -107,7 +108,7 @@ func (f *EnvAWSFingerprint) Fingerprint(request *cstructs.FingerprintRequest, re for k, unique := range keys { res, err := client.Get(metadataURL + k) if res.StatusCode != http.StatusOK { - f.logger.Printf("[DEBUG]: fingerprint.env_aws: Could not read value for attribute %q", k) + f.logger.Debug("could not read attribute value", "attribute", k) continue } if err != nil { @@ -122,7 +123,7 @@ func (f *EnvAWSFingerprint) Fingerprint(request *cstructs.FingerprintRequest, re resp, err := ioutil.ReadAll(res.Body) res.Body.Close() if err != nil { - f.logger.Printf("[ERR]: fingerprint.env_aws: Error reading response body for AWS %s", k) + f.logger.Error("error reading response body for AWS attribute", "attribute", k, "error", err) } // assume we want blank entries @@ -194,7 +195,7 @@ func (f *EnvAWSFingerprint) isAWS() bool { // Query the metadata url for the ami-id, to verify we're on AWS resp, err := client.Get(metadataURL + "ami-id") if err != nil { - f.logger.Printf("[DEBUG] fingerprint.env_aws: Error querying AWS Metadata URL, skipping") + f.logger.Debug("error querying AWS Metadata URL, skipping") return false } defer resp.Body.Close() @@ -206,7 +207,7 @@ func (f *EnvAWSFingerprint) isAWS() bool { instanceID, err := ioutil.ReadAll(resp.Body) if err != nil { - f.logger.Printf("[DEBUG] fingerprint.env_aws: Error reading AWS Instance ID, skipping") + f.logger.Debug("error reading AWS Instance ID, skipping") return false } @@ -235,14 +236,14 @@ func (f *EnvAWSFingerprint) linkSpeed() int { res, err := client.Get(metadataURL + "instance-type") if err != nil { - f.logger.Printf("[ERR]: fingerprint.env_aws: Error reading instance-type: %v", err) + f.logger.Error("error reading instance-type", "error", err) return 0 } body, err := ioutil.ReadAll(res.Body) res.Body.Close() if err != nil { - f.logger.Printf("[ERR]: fingerprint.env_aws: Error reading response body for instance-type: %v", err) + f.logger.Error("error reading response body for instance-type", "error", err) return 0 } diff --git a/client/fingerprint/env_aws_test.go b/client/fingerprint/env_aws_test.go index 14699b845..3f16a3e3d 100644 --- a/client/fingerprint/env_aws_test.go +++ b/client/fingerprint/env_aws_test.go @@ -16,7 +16,7 @@ import ( func TestEnvAWSFingerprint_nonAws(t *testing.T) { os.Setenv("AWS_ENV_URL", "http://127.0.0.1/latest/meta-data/") - f := NewEnvAWSFingerprint(testlog.Logger(t)) + f := NewEnvAWSFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } @@ -34,7 +34,7 @@ func TestEnvAWSFingerprint_nonAws(t *testing.T) { } func TestEnvAWSFingerprint_aws(t *testing.T) { - f := NewEnvAWSFingerprint(testlog.Logger(t)) + f := NewEnvAWSFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } @@ -168,7 +168,7 @@ func TestNetworkFingerprint_AWS(t *testing.T) { defer ts.Close() os.Setenv("AWS_ENV_URL", ts.URL+"/latest/meta-data/") - f := NewEnvAWSFingerprint(testlog.Logger(t)) + f := NewEnvAWSFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } @@ -217,7 +217,7 @@ func TestNetworkFingerprint_AWS_network(t *testing.T) { defer ts.Close() os.Setenv("AWS_ENV_URL", ts.URL+"/latest/meta-data/") - f := NewEnvAWSFingerprint(testlog.Logger(t)) + f := NewEnvAWSFingerprint(testlog.HCLogger(t)) { node := &structs.Node{ Attributes: make(map[string]string), @@ -298,7 +298,7 @@ func TestNetworkFingerprint_AWS_network(t *testing.T) { func TestNetworkFingerprint_notAWS(t *testing.T) { os.Setenv("AWS_ENV_URL", "http://127.0.0.1/latest/meta-data/") - f := NewEnvAWSFingerprint(testlog.Logger(t)) + f := NewEnvAWSFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } diff --git a/client/fingerprint/env_gce.go b/client/fingerprint/env_gce.go index e404e2415..c3ca35638 100644 --- a/client/fingerprint/env_gce.go +++ b/client/fingerprint/env_gce.go @@ -4,7 +4,6 @@ import ( "encoding/json" "fmt" "io/ioutil" - "log" "net/http" "net/url" "os" @@ -13,8 +12,10 @@ import ( "strings" "time" - "github.com/hashicorp/go-cleanhttp" + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" + + "github.com/hashicorp/go-cleanhttp" "github.com/hashicorp/nomad/helper/useragent" "github.com/hashicorp/nomad/nomad/structs" ) @@ -56,12 +57,12 @@ func lastToken(s string) string { type EnvGCEFingerprint struct { StaticFingerprinter client *http.Client - logger *log.Logger + logger log.Logger metadataURL string } // NewEnvGCEFingerprint is used to create a fingerprint from GCE metadata -func NewEnvGCEFingerprint(logger *log.Logger) Fingerprint { +func NewEnvGCEFingerprint(logger log.Logger) Fingerprint { // Read the internal metadata URL from the environment, allowing test files to // provide their own metadataURL := os.Getenv("GCE_ENV_URL") @@ -77,7 +78,7 @@ func NewEnvGCEFingerprint(logger *log.Logger) Fingerprint { return &EnvGCEFingerprint{ client: client, - logger: logger, + logger: logger.Named("env_gce"), metadataURL: metadataURL, } } @@ -103,15 +104,18 @@ func (f *EnvGCEFingerprint) Get(attribute string, recursive bool) (string, error } res, err := f.client.Do(req) - if err != nil || res.StatusCode != http.StatusOK { - f.logger.Printf("[DEBUG] fingerprint.env_gce: Could not read value for attribute %q", attribute) + if err != nil { + f.logger.Debug("could not read value for attribute", "attribute", attribute, "error", err) + return "", err + } else if res.StatusCode != http.StatusOK { + f.logger.Debug("could not read value for attribute", "attribute", attribute, "resp_code", res.StatusCode) return "", err } resp, err := ioutil.ReadAll(res.Body) res.Body.Close() if err != nil { - f.logger.Printf("[ERR] fingerprint.env_gce: Error reading response body for GCE %s", attribute) + f.logger.Error("error reading response body for GCE attribute", "attribute", attribute, "error", err) return "", err } @@ -122,11 +126,11 @@ func (f *EnvGCEFingerprint) Get(attribute string, recursive bool) (string, error return string(resp), nil } -func checkError(err error, logger *log.Logger, desc string) error { +func checkError(err error, logger log.Logger, desc string) error { // If it's a URL error, assume we're not actually in an GCE environment. // To the outer layers, this isn't an error so return nil. if _, ok := err.(*url.Error); ok { - logger.Printf("[DEBUG] fingerprint.env_gce: Error querying GCE " + desc + ", skipping") + logger.Debug("error querying GCE attribute; skipping", "attribute", desc) return nil } // Otherwise pass the error through. @@ -191,12 +195,12 @@ func (f *EnvGCEFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp * // Get internal and external IPs (if they exist) value, err := f.Get("network-interfaces/", true) if err != nil { - f.logger.Printf("[WARN] fingerprint.env_gce: Error retrieving network interface information: %s", err) + f.logger.Warn("error retrieving network interface information", "error", err) } else { var interfaces []GCEMetadataNetworkInterface if err := json.Unmarshal([]byte(value), &interfaces); err != nil { - f.logger.Printf("[WARN] fingerprint.env_gce: Error decoding network interface information: %s", err.Error()) + f.logger.Warn("error decoding network interface information", "error", err) } for _, intf := range interfaces { @@ -216,7 +220,7 @@ func (f *EnvGCEFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp * return checkError(err, f.logger, "tags") } if err := json.Unmarshal([]byte(value), &tagList); err != nil { - f.logger.Printf("[WARN] fingerprint.env_gce: Error decoding instance tags: %s", err.Error()) + f.logger.Warn("error decoding instance tags", "error", err) } for _, tag := range tagList { attr := "platform.gce.tag." @@ -240,7 +244,7 @@ func (f *EnvGCEFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp * return checkError(err, f.logger, "attributes/") } if err := json.Unmarshal([]byte(value), &attrDict); err != nil { - f.logger.Printf("[WARN] fingerprint.env_gce: Error decoding instance attributes: %s", err.Error()) + f.logger.Warn("error decoding instance attributes", "error", err) } for k, v := range attrDict { attr := "platform.gce.attr." @@ -276,7 +280,7 @@ func (f *EnvGCEFingerprint) isGCE() bool { if err != nil { if re, ok := err.(ReqError); !ok || re.StatusCode != 404 { // If it wasn't a 404 error, print an error message. - f.logger.Printf("[DEBUG] fingerprint.env_gce: Error querying GCE Metadata URL, skipping") + f.logger.Debug("error querying GCE Metadata URL, skipping") } return false } diff --git a/client/fingerprint/env_gce_test.go b/client/fingerprint/env_gce_test.go index 445643765..82a596787 100644 --- a/client/fingerprint/env_gce_test.go +++ b/client/fingerprint/env_gce_test.go @@ -17,7 +17,7 @@ import ( func TestGCEFingerprint_nonGCE(t *testing.T) { os.Setenv("GCE_ENV_URL", "http://127.0.0.1/computeMetadata/v1/instance/") - f := NewEnvGCEFingerprint(testlog.Logger(t)) + f := NewEnvGCEFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } @@ -91,7 +91,7 @@ func testFingerprint_GCE(t *testing.T, withExternalIp bool) { })) defer ts.Close() os.Setenv("GCE_ENV_URL", ts.URL+"/computeMetadata/v1/instance/") - f := NewEnvGCEFingerprint(testlog.Logger(t)) + f := NewEnvGCEFingerprint(testlog.HCLogger(t)) request := &cstructs.FingerprintRequest{Config: &config.Config{}, Node: node} var response cstructs.FingerprintResponse diff --git a/client/fingerprint/fingerprint.go b/client/fingerprint/fingerprint.go index d6746a03d..1a5f6f41c 100644 --- a/client/fingerprint/fingerprint.go +++ b/client/fingerprint/fingerprint.go @@ -2,10 +2,10 @@ package fingerprint import ( "fmt" - "log" "sort" "time" + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" ) @@ -67,7 +67,7 @@ func BuiltinFingerprints() []string { // NewFingerprint is used to instantiate and return a new fingerprint // given the name and a logger -func NewFingerprint(name string, logger *log.Logger) (Fingerprint, error) { +func NewFingerprint(name string, logger log.Logger) (Fingerprint, error) { // Lookup the factory function factory, ok := hostFingerprinters[name] if !ok { @@ -83,7 +83,7 @@ func NewFingerprint(name string, logger *log.Logger) (Fingerprint, error) { } // Factory is used to instantiate a new Fingerprint -type Factory func(*log.Logger) Fingerprint +type Factory func(log.Logger) Fingerprint // HealthCheck is used for doing periodic health checks. On a given time // interfal, a health check will be called by the fingerprint manager of the diff --git a/client/fingerprint/host.go b/client/fingerprint/host.go index cfeabd4ac..d0e69163f 100644 --- a/client/fingerprint/host.go +++ b/client/fingerprint/host.go @@ -1,9 +1,9 @@ package fingerprint import ( - "log" "runtime" + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/shirou/gopsutil/host" ) @@ -11,19 +11,19 @@ import ( // HostFingerprint is used to fingerprint the host type HostFingerprint struct { StaticFingerprinter - logger *log.Logger + logger log.Logger } // NewHostFingerprint is used to create a Host fingerprint -func NewHostFingerprint(logger *log.Logger) Fingerprint { - f := &HostFingerprint{logger: logger} +func NewHostFingerprint(logger log.Logger) Fingerprint { + f := &HostFingerprint{logger: logger.Named("host")} return f } func (f *HostFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp *cstructs.FingerprintResponse) error { hostInfo, err := host.Info() if err != nil { - f.logger.Println("[WARN] Error retrieving host information: ", err) + f.logger.Warn("error retrieving host information", "error", err) return err } diff --git a/client/fingerprint/host_test.go b/client/fingerprint/host_test.go index 6fafca0d9..2fe3a635f 100644 --- a/client/fingerprint/host_test.go +++ b/client/fingerprint/host_test.go @@ -10,7 +10,7 @@ import ( ) func TestHostFingerprint(t *testing.T) { - f := NewHostFingerprint(testlog.Logger(t)) + f := NewHostFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } diff --git a/client/fingerprint/memory.go b/client/fingerprint/memory.go index 938310232..f31ded76b 100644 --- a/client/fingerprint/memory.go +++ b/client/fingerprint/memory.go @@ -2,8 +2,8 @@ package fingerprint import ( "fmt" - "log" + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" "github.com/shirou/gopsutil/mem" @@ -14,13 +14,13 @@ const bytesInMB = 1024 * 1024 // MemoryFingerprint is used to fingerprint the available memory on the node type MemoryFingerprint struct { StaticFingerprinter - logger *log.Logger + logger log.Logger } // NewMemoryFingerprint is used to create a Memory fingerprint -func NewMemoryFingerprint(logger *log.Logger) Fingerprint { +func NewMemoryFingerprint(logger log.Logger) Fingerprint { f := &MemoryFingerprint{ - logger: logger, + logger: logger.Named("memory"), } return f } @@ -33,7 +33,7 @@ func (f *MemoryFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp * } else { memInfo, err := mem.VirtualMemory() if err != nil { - f.logger.Printf("[WARN] Error reading memory information: %s", err) + f.logger.Warn("error reading memory information", "error", err) return err } if memInfo.Total > 0 { diff --git a/client/fingerprint/memory_test.go b/client/fingerprint/memory_test.go index e01502205..b1e0c14ac 100644 --- a/client/fingerprint/memory_test.go +++ b/client/fingerprint/memory_test.go @@ -12,7 +12,7 @@ import ( ) func TestMemoryFingerprint(t *testing.T) { - f := NewMemoryFingerprint(testlog.Logger(t)) + f := NewMemoryFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } @@ -41,7 +41,7 @@ func TestMemoryFingerprint(t *testing.T) { } func TestMemoryFingerprint_Override(t *testing.T) { - f := NewMemoryFingerprint(testlog.Logger(t)) + f := NewMemoryFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } diff --git a/client/fingerprint/network.go b/client/fingerprint/network.go index b83599b9c..f42b31c37 100644 --- a/client/fingerprint/network.go +++ b/client/fingerprint/network.go @@ -2,9 +2,9 @@ package fingerprint import ( "fmt" - "log" "net" + log "github.com/hashicorp/go-hclog" sockaddr "github.com/hashicorp/go-sockaddr" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" @@ -25,7 +25,7 @@ const ( // NetworkFingerprint is used to fingerprint the Network capabilities of a node type NetworkFingerprint struct { StaticFingerprinter - logger *log.Logger + logger log.Logger interfaceDetector NetworkInterfaceDetector } @@ -56,8 +56,8 @@ func (b *DefaultNetworkInterfaceDetector) Addrs(intf *net.Interface) ([]net.Addr // NewNetworkFingerprint returns a new NetworkFingerprinter with the given // logger -func NewNetworkFingerprint(logger *log.Logger) Fingerprint { - f := &NetworkFingerprint{logger: logger, interfaceDetector: &DefaultNetworkInterfaceDetector{}} +func NewNetworkFingerprint(logger log.Logger) Fingerprint { + f := &NetworkFingerprint{logger: logger.Named("network"), interfaceDetector: &DefaultNetworkInterfaceDetector{}} return f } @@ -79,13 +79,13 @@ func (f *NetworkFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp throughput := f.linkSpeed(intf.Name) if cfg.NetworkSpeed != 0 { mbits = cfg.NetworkSpeed - f.logger.Printf("[DEBUG] fingerprint.network: setting link speed to user configured speed: %d", mbits) + f.logger.Debug("setting link speed to user configured speed", "mbits", mbits) } else if throughput != 0 { mbits = throughput - f.logger.Printf("[DEBUG] fingerprint.network: link speed for %v set to %v", intf.Name, mbits) + f.logger.Debug("link speed detected", "interface", intf.Name, "mbits", mbits) } else { mbits = defaultNetworkSpeed - f.logger.Printf("[DEBUG] fingerprint.network: link speed could not be detected and no speed specified by user. Defaulting to %d", defaultNetworkSpeed) + f.logger.Debug("link speed could not be detected and no speed specified by user, falling back to default speed", "mbits", defaultNetworkSpeed) } // Create the network resources from the interface @@ -105,7 +105,7 @@ func (f *NetworkFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp } for _, nwResource := range nwResources { - f.logger.Printf("[DEBUG] fingerprint.network: Detected interface %v with IP: %v", intf.Name, nwResource.IP) + f.logger.Debug("detected interface IP", "interface", intf.Name, "IP", nwResource.IP) } // Deprecated, setting the first IP as unique IP for the node @@ -163,7 +163,7 @@ func (f *NetworkFingerprint) createNetworkResources(throughput int, intf *net.In if len(nwResources) == 0 && len(linkLocals) != 0 { if disallowLinkLocal { - f.logger.Printf("[DEBUG] fingerprint.network: ignoring detected link-local address on interface %v", intf.Name) + f.logger.Debug("ignoring detected link-local address on interface", "interface", intf.Name) return nwResources, nil } diff --git a/client/fingerprint/network_test.go b/client/fingerprint/network_test.go index 527865ed6..5d69710a0 100644 --- a/client/fingerprint/network_test.go +++ b/client/fingerprint/network_test.go @@ -185,7 +185,7 @@ func TestNetworkFingerprint_basic(t *testing.T) { t.Skipf("Environment variable %+q not empty, skipping test", skipOnlineTestsEnvVar) } - f := &NetworkFingerprint{logger: testlog.Logger(t), interfaceDetector: &DefaultNetworkInterfaceDetector{}} + f := &NetworkFingerprint{logger: testlog.HCLogger(t), interfaceDetector: &DefaultNetworkInterfaceDetector{}} node := &structs.Node{ Attributes: make(map[string]string), } @@ -236,7 +236,7 @@ func TestNetworkFingerprint_basic(t *testing.T) { } func TestNetworkFingerprint_default_device_absent(t *testing.T) { - f := &NetworkFingerprint{logger: testlog.Logger(t), interfaceDetector: &NetworkInterfaceDetectorOnlyLo{}} + f := &NetworkFingerprint{logger: testlog.HCLogger(t), interfaceDetector: &NetworkInterfaceDetectorOnlyLo{}} node := &structs.Node{ Attributes: make(map[string]string), } @@ -259,7 +259,7 @@ func TestNetworkFingerprint_default_device_absent(t *testing.T) { } func TestNetworkFingerPrint_default_device(t *testing.T) { - f := &NetworkFingerprint{logger: testlog.Logger(t), interfaceDetector: &NetworkInterfaceDetectorOnlyLo{}} + f := &NetworkFingerprint{logger: testlog.HCLogger(t), interfaceDetector: &NetworkInterfaceDetectorOnlyLo{}} node := &structs.Node{ Attributes: make(map[string]string), } @@ -310,7 +310,7 @@ func TestNetworkFingerPrint_default_device(t *testing.T) { } func TestNetworkFingerPrint_LinkLocal_Allowed(t *testing.T) { - f := &NetworkFingerprint{logger: testlog.Logger(t), interfaceDetector: &NetworkInterfaceDetectorMultipleInterfaces{}} + f := &NetworkFingerprint{logger: testlog.HCLogger(t), interfaceDetector: &NetworkInterfaceDetectorMultipleInterfaces{}} node := &structs.Node{ Attributes: make(map[string]string), } @@ -357,7 +357,7 @@ func TestNetworkFingerPrint_LinkLocal_Allowed(t *testing.T) { } func TestNetworkFingerPrint_LinkLocal_Allowed_MixedIntf(t *testing.T) { - f := &NetworkFingerprint{logger: testlog.Logger(t), interfaceDetector: &NetworkInterfaceDetectorMultipleInterfaces{}} + f := &NetworkFingerprint{logger: testlog.HCLogger(t), interfaceDetector: &NetworkInterfaceDetectorMultipleInterfaces{}} node := &structs.Node{ Attributes: make(map[string]string), } @@ -411,7 +411,7 @@ func TestNetworkFingerPrint_LinkLocal_Allowed_MixedIntf(t *testing.T) { } func TestNetworkFingerPrint_LinkLocal_Disallowed(t *testing.T) { - f := &NetworkFingerprint{logger: testlog.Logger(t), interfaceDetector: &NetworkInterfaceDetectorMultipleInterfaces{}} + f := &NetworkFingerprint{logger: testlog.HCLogger(t), interfaceDetector: &NetworkInterfaceDetectorMultipleInterfaces{}} node := &structs.Node{ Attributes: make(map[string]string), } diff --git a/client/fingerprint/nomad.go b/client/fingerprint/nomad.go index a8415f69d..59d4c5b39 100644 --- a/client/fingerprint/nomad.go +++ b/client/fingerprint/nomad.go @@ -1,20 +1,19 @@ package fingerprint import ( - "log" - + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" ) // NomadFingerprint is used to fingerprint the Nomad version type NomadFingerprint struct { StaticFingerprinter - logger *log.Logger + logger log.Logger } // NewNomadFingerprint is used to create a Nomad fingerprint -func NewNomadFingerprint(logger *log.Logger) Fingerprint { - f := &NomadFingerprint{logger: logger} +func NewNomadFingerprint(logger log.Logger) Fingerprint { + f := &NomadFingerprint{logger: logger.Named("nomad")} return f } diff --git a/client/fingerprint/nomad_test.go b/client/fingerprint/nomad_test.go index f86eea4ba..a57b322ef 100644 --- a/client/fingerprint/nomad_test.go +++ b/client/fingerprint/nomad_test.go @@ -11,7 +11,7 @@ import ( ) func TestNomadFingerprint(t *testing.T) { - f := NewNomadFingerprint(testlog.Logger(t)) + f := NewNomadFingerprint(testlog.HCLogger(t)) v := "foo" r := "123" diff --git a/client/fingerprint/signal.go b/client/fingerprint/signal.go index 9aac819e1..1d328d1a2 100644 --- a/client/fingerprint/signal.go +++ b/client/fingerprint/signal.go @@ -1,22 +1,22 @@ package fingerprint import ( - "log" "strings" "github.com/hashicorp/consul-template/signals" + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" ) // SignalFingerprint is used to fingerprint the available signals type SignalFingerprint struct { StaticFingerprinter - logger *log.Logger + logger log.Logger } // NewSignalFingerprint is used to create a Signal fingerprint -func NewSignalFingerprint(logger *log.Logger) Fingerprint { - f := &SignalFingerprint{logger: logger} +func NewSignalFingerprint(logger log.Logger) Fingerprint { + f := &SignalFingerprint{logger: logger.Named("signals")} return f } diff --git a/client/fingerprint/signal_test.go b/client/fingerprint/signal_test.go index ff0577125..d88c4a85a 100644 --- a/client/fingerprint/signal_test.go +++ b/client/fingerprint/signal_test.go @@ -8,7 +8,7 @@ import ( ) func TestSignalFingerprint(t *testing.T) { - fp := NewSignalFingerprint(testlog.Logger(t)) + fp := NewSignalFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } diff --git a/client/fingerprint/storage.go b/client/fingerprint/storage.go index edf14e247..501b85a52 100644 --- a/client/fingerprint/storage.go +++ b/client/fingerprint/storage.go @@ -2,10 +2,10 @@ package fingerprint import ( "fmt" - "log" "os" "strconv" + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" ) @@ -16,11 +16,11 @@ const bytesPerMegabyte = 1024 * 1024 // applications that the Nomad agent will run on this machine. type StorageFingerprint struct { StaticFingerprinter - logger *log.Logger + logger log.Logger } -func NewStorageFingerprint(logger *log.Logger) Fingerprint { - fp := &StorageFingerprint{logger: logger} +func NewStorageFingerprint(logger log.Logger) Fingerprint { + fp := &StorageFingerprint{logger: logger.Named("storage")} return fp } diff --git a/client/fingerprint/storage_test.go b/client/fingerprint/storage_test.go index 0b7a47d60..3227d4d2b 100644 --- a/client/fingerprint/storage_test.go +++ b/client/fingerprint/storage_test.go @@ -9,7 +9,7 @@ import ( ) func TestStorageFingerprint(t *testing.T) { - fp := NewStorageFingerprint(testlog.Logger(t)) + fp := NewStorageFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } diff --git a/client/fingerprint/vault.go b/client/fingerprint/vault.go index 0613e98a0..8ad2c3cab 100644 --- a/client/fingerprint/vault.go +++ b/client/fingerprint/vault.go @@ -2,11 +2,11 @@ package fingerprint import ( "fmt" - "log" "strconv" "strings" "time" + log "github.com/hashicorp/go-hclog" cstructs "github.com/hashicorp/nomad/client/structs" vapi "github.com/hashicorp/vault/api" ) @@ -18,14 +18,14 @@ const ( // VaultFingerprint is used to fingerprint for Vault type VaultFingerprint struct { - logger *log.Logger + logger log.Logger client *vapi.Client lastState string } // NewVaultFingerprint is used to create a Vault fingerprint -func NewVaultFingerprint(logger *log.Logger) Fingerprint { - return &VaultFingerprint{logger: logger, lastState: vaultUnavailable} +func NewVaultFingerprint(logger log.Logger) Fingerprint { + return &VaultFingerprint{logger: logger.Named("vault"), lastState: vaultUnavailable} } func (f *VaultFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp *cstructs.FingerprintResponse) error { @@ -55,7 +55,7 @@ func (f *VaultFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp *c f.clearVaultAttributes(resp) // Print a message indicating that Vault is not available anymore if f.lastState == vaultAvailable { - f.logger.Printf("[INFO] fingerprint.vault: Vault is unavailable") + f.logger.Info("Vault is unavailable") } f.lastState = vaultUnavailable return nil @@ -71,7 +71,7 @@ func (f *VaultFingerprint) Fingerprint(req *cstructs.FingerprintRequest, resp *c // If Vault was previously unavailable print a message to indicate the Agent // is available now if f.lastState == vaultUnavailable { - f.logger.Printf("[INFO] fingerprint.vault: Vault is available") + f.logger.Info("Vault is available") } f.lastState = vaultAvailable resp.Detected = true diff --git a/client/fingerprint/vault_test.go b/client/fingerprint/vault_test.go index 3ca1fa448..f25d7b76a 100644 --- a/client/fingerprint/vault_test.go +++ b/client/fingerprint/vault_test.go @@ -14,7 +14,7 @@ func TestVaultFingerprint(t *testing.T) { tv := testutil.NewTestVault(t) defer tv.Stop() - fp := NewVaultFingerprint(testlog.Logger(t)) + fp := NewVaultFingerprint(testlog.HCLogger(t)) node := &structs.Node{ Attributes: make(map[string]string), } diff --git a/client/fingerprint_manager.go b/client/fingerprint_manager.go index 340bac876..8e007fb05 100644 --- a/client/fingerprint_manager.go +++ b/client/fingerprint_manager.go @@ -2,12 +2,10 @@ package client import ( "fmt" - "log" - "os" "sync" "time" - hclog "github.com/hashicorp/go-hclog" + log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/fingerprint" @@ -30,7 +28,7 @@ type FingerprintManager struct { // updateNodeFromDriver is a callback to the client to update the state of a // specific driver for the node updateNodeFromDriver func(string, *structs.DriverInfo, *structs.DriverInfo) *structs.Node - logger hclog.Logger + logger log.Logger } // NewFingerprintManager is a constructor that creates and returns an instance @@ -40,16 +38,15 @@ func NewFingerprintManager(getConfig func() *config.Config, shutdownCh chan struct{}, updateNodeAttributes func(*cstructs.FingerprintResponse) *structs.Node, updateNodeFromDriver func(string, *structs.DriverInfo, *structs.DriverInfo) *structs.Node, - logger hclog.Logger) *FingerprintManager { + logger log.Logger) *FingerprintManager { - logger = logger.Named("fingerprint_mgr") return &FingerprintManager{ getConfig: getConfig, updateNodeAttributes: updateNodeAttributes, updateNodeFromDriver: updateNodeFromDriver, node: node, shutdownCh: shutdownCh, - logger: logger, + logger: logger.Named("fingerprint_mgr"), } } @@ -144,13 +141,10 @@ func (fp *FingerprintManager) Run() error { // setupFingerprints is used to fingerprint the node to see if these attributes are // supported func (fm *FingerprintManager) setupFingerprinters(fingerprints []string) error { - //FIXME Update fingerprinters to hclog - tempLogger := log.New(os.Stdout, "", log.LstdFlags|log.Lmicroseconds) - var appliedFingerprints []string for _, name := range fingerprints { - f, err := fingerprint.NewFingerprint(name, tempLogger) + f, err := fingerprint.NewFingerprint(name, fm.logger) if err != nil { fm.logger.Error("error fingerprinting", "error", err, "fingerprinter", name) @@ -180,11 +174,9 @@ func (fm *FingerprintManager) setupFingerprinters(fingerprints []string) error { // setupDrivers is used to fingerprint the node to see if these drivers are // supported func (fm *FingerprintManager) setupDrivers(drivers []string) error { - //FIXME Update fingerprinters to hclog - tempLogger := log.New(os.Stdout, "", log.LstdFlags|log.Lmicroseconds) - + //TODO(alex,hclog) Update fingerprinters to hclog var availDrivers []string - driverCtx := driver.NewDriverContext("", "", "", "", fm.getConfig(), fm.getNode(), tempLogger, nil) + driverCtx := driver.NewDriverContext("", "", "", "", fm.getConfig(), fm.getNode(), fm.logger.StandardLogger(&log.StandardLoggerOptions{InferLevels: true}), nil) for _, name := range drivers { d, err := driver.NewDriver(name, driverCtx) diff --git a/client/fingerprint_manager_test.go b/client/fingerprint_manager_test.go index 11c8d49ae..00e139689 100644 --- a/client/fingerprint_manager_test.go +++ b/client/fingerprint_manager_test.go @@ -15,7 +15,7 @@ func TestFingerprintManager_Run_MockDriver(t *testing.T) { require := require.New(t) testClient := TestClient(t, nil) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -24,7 +24,7 @@ func TestFingerprintManager_Run_MockDriver(t *testing.T) { testClient.shutdownCh, testClient.updateNodeFromFingerprint, testClient.updateNodeFromDriver, - testlog.Logger(t), + testlog.HCLogger(t), ) err := fm.Run() @@ -42,7 +42,7 @@ func TestFingerprintManager_Run_ResourcesFingerprint(t *testing.T) { require := require.New(t) testClient := TestClient(t, nil) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -69,7 +69,7 @@ func TestFingerprintManager_Fingerprint_Run(t *testing.T) { require := require.New(t) testClient := TestClient(t, nil) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -101,7 +101,7 @@ func TestFingerprintManager_Fingerprint_Periodic(t *testing.T) { } }) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -163,7 +163,7 @@ func TestFingerprintManager_HealthCheck_Driver(t *testing.T) { } }) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -263,7 +263,7 @@ func TestFingerprintManager_HealthCheck_Periodic(t *testing.T) { } }) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -359,7 +359,7 @@ func TestFimgerprintManager_Run_InWhitelist(t *testing.T) { } }) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -389,7 +389,7 @@ func TestFingerprintManager_Run_InBlacklist(t *testing.T) { } }) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -421,7 +421,7 @@ func TestFingerprintManager_Run_Combination(t *testing.T) { } }) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -454,7 +454,7 @@ func TestFingerprintManager_Run_WhitelistDrivers(t *testing.T) { } }) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -484,7 +484,7 @@ func TestFingerprintManager_Run_AllDriversBlacklisted(t *testing.T) { } }) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -518,7 +518,7 @@ func TestFingerprintManager_Run_DriversWhiteListBlacklistCombination(t *testing. } }) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( @@ -551,7 +551,7 @@ func TestFingerprintManager_Run_DriversInBlacklist(t *testing.T) { } }) - testClient.logger = testlog.Logger(t) + testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( diff --git a/client/fs_endpoint_test.go b/client/fs_endpoint_test.go index 49d5d7ec5..ec555fb6c 100644 --- a/client/fs_endpoint_test.go +++ b/client/fs_endpoint_test.go @@ -41,7 +41,7 @@ func tempAllocDir(t testing.TB) *allocdir.AllocDir { t.Fatalf("failed to chmod dir: %v", err) } - return allocdir.NewAllocDir(testlog.Logger(t), dir) + return allocdir.NewAllocDir(testlog.HCLogger(t), dir) } type nopWriteCloser struct { From 03422aa5297bf9911ab341ad270d43d3ca72551c Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Mon, 17 Sep 2018 13:15:48 -0400 Subject: [PATCH 065/152] fifo: add new fifo package for named pipes (#4665) * fifo: add new fifo package for named pipes --- client/lib/fifo/doc.go | 8 + client/lib/fifo/fifo_test.go | 117 +++++++++ client/lib/fifo/fifo_unix.go | 36 +++ client/lib/fifo/fifo_windows.go | 105 ++++++++ .../github.com/Microsoft/go-winio/backup.go | 14 +- vendor/github.com/Microsoft/go-winio/ea.go | 137 ++++++++++ vendor/github.com/Microsoft/go-winio/file.go | 32 ++- .../github.com/Microsoft/go-winio/fileinfo.go | 3 +- vendor/github.com/Microsoft/go-winio/pipe.go | 121 +++++---- .../Microsoft/go-winio/zsyscall_windows.go | 8 - vendor/github.com/containerd/fifo/LICENSE | 201 +++++++++++++++ vendor/github.com/containerd/fifo/Makefile | 27 ++ vendor/github.com/containerd/fifo/fifo.go | 236 ++++++++++++++++++ .../containerd/fifo/handle_linux.go | 97 +++++++ .../containerd/fifo/handle_nolinux.go | 65 +++++ .../containerd/fifo/mkfifo_nosolaris.go | 25 ++ .../containerd/fifo/mkfifo_solaris.go | 27 ++ vendor/github.com/containerd/fifo/readme.md | 32 +++ vendor/vendor.json | 3 +- 19 files changed, 1220 insertions(+), 74 deletions(-) create mode 100644 client/lib/fifo/doc.go create mode 100644 client/lib/fifo/fifo_test.go create mode 100644 client/lib/fifo/fifo_unix.go create mode 100644 client/lib/fifo/fifo_windows.go create mode 100644 vendor/github.com/Microsoft/go-winio/ea.go create mode 100644 vendor/github.com/containerd/fifo/LICENSE create mode 100644 vendor/github.com/containerd/fifo/Makefile create mode 100644 vendor/github.com/containerd/fifo/fifo.go create mode 100644 vendor/github.com/containerd/fifo/handle_linux.go create mode 100644 vendor/github.com/containerd/fifo/handle_nolinux.go create mode 100644 vendor/github.com/containerd/fifo/mkfifo_nosolaris.go create mode 100644 vendor/github.com/containerd/fifo/mkfifo_solaris.go create mode 100644 vendor/github.com/containerd/fifo/readme.md diff --git a/client/lib/fifo/doc.go b/client/lib/fifo/doc.go new file mode 100644 index 000000000..d1d2779f8 --- /dev/null +++ b/client/lib/fifo/doc.go @@ -0,0 +1,8 @@ +/* +Package fifo implements functions to create and open a fifo for inter-process +communication in an OS agnostic way. A few assumptions should be made when +using this package. First, New() must always be called before Open(). Second +Open() returns an io.ReadWriteCloser that is only connected with the +io.ReadWriteCloser returned from New(). +*/ +package fifo diff --git a/client/lib/fifo/fifo_test.go b/client/lib/fifo/fifo_test.go new file mode 100644 index 000000000..f1814f1ba --- /dev/null +++ b/client/lib/fifo/fifo_test.go @@ -0,0 +1,117 @@ +package fifo + +import ( + "bytes" + "io" + "io/ioutil" + "os" + "path/filepath" + "runtime" + "sync" + "testing" + "time" + + "github.com/hashicorp/nomad/helper/uuid" + "github.com/stretchr/testify/require" +) + +func TestFIFO(t *testing.T) { + require := require.New(t) + var path string + + if runtime.GOOS == "windows" { + path = "//./pipe/fifo" + } else { + dir, err := ioutil.TempDir("", "") + require.NoError(err) + defer os.RemoveAll(dir) + + path = filepath.Join(dir, "fifo") + } + + reader, err := New(path) + require.NoError(err) + + toWrite := [][]byte{ + []byte("abc\n"), + []byte(""), + []byte("def\n"), + []byte("nomad"), + []byte("\n"), + } + + var readBuf bytes.Buffer + var wait sync.WaitGroup + wait.Add(1) + go func() { + defer wait.Done() + io.Copy(&readBuf, reader) + }() + + writer, err := Open(path) + require.NoError(err) + for _, b := range toWrite { + n, err := writer.Write(b) + require.NoError(err) + require.Equal(n, len(b)) + } + require.NoError(writer.Close()) + time.Sleep(500 * time.Millisecond) + require.NoError(reader.Close()) + + wait.Wait() + + expected := "abc\ndef\nnomad\n" + require.Equal(expected, readBuf.String()) + + require.NoError(Remove(path)) +} + +func TestWriteClose(t *testing.T) { + require := require.New(t) + var path string + + if runtime.GOOS == "windows" { + path = "//./pipe/" + uuid.Generate()[:4] + } else { + dir, err := ioutil.TempDir("", "") + require.NoError(err) + defer os.RemoveAll(dir) + + path = filepath.Join(dir, "fifo") + } + + reader, err := New(path) + require.NoError(err) + + var readBuf bytes.Buffer + var wait sync.WaitGroup + wait.Add(1) + go func() { + defer wait.Done() + io.Copy(&readBuf, reader) + }() + + writer, err := Open(path) + require.NoError(err) + + var count int + wait.Add(1) + go func() { + defer wait.Done() + for count = 0; count < int(^uint16(0)); count++ { + _, err := writer.Write([]byte(",")) + if err != nil && IsClosedErr(err) { + break + } + require.NoError(err) + time.Sleep(5 * time.Millisecond) + } + }() + + time.Sleep(500 * time.Millisecond) + require.NoError(writer.Close()) + wait.Wait() + + require.Equal(count, len(readBuf.String())) +} diff --git a/client/lib/fifo/fifo_unix.go b/client/lib/fifo/fifo_unix.go new file mode 100644 index 000000000..bc98a4618 --- /dev/null +++ b/client/lib/fifo/fifo_unix.go @@ -0,0 +1,36 @@ +// +build !windows + +package fifo + +import ( + "context" + "io" + "os" + "syscall" + + cfifo "github.com/containerd/fifo" +) + +// New creates a fifo at the given path and returns an io.ReadWriteCloser for it +// The fifo must not already exist +func New(path string) (io.ReadWriteCloser, error) { + return cfifo.OpenFifo(context.Background(), path, syscall.O_RDONLY|syscall.O_CREAT|syscall.O_NONBLOCK, 0600) +} + +// Open opens a fifo that already exists and returns an io.ReadWriteCloser for it +func Open(path string) (io.ReadWriteCloser, error) { + return cfifo.OpenFifo(context.Background(), path, syscall.O_WRONLY|syscall.O_NONBLOCK, 0600) +} + +// Remove a fifo that already exists at a given path +func Remove(path string) error { + return os.Remove(path) +} + +func IsClosedErr(err error) bool { + err2, ok := err.(*os.PathError) + if ok { + return err2.Err == os.ErrClosed + } + return false +} diff --git a/client/lib/fifo/fifo_windows.go b/client/lib/fifo/fifo_windows.go new file mode 100644 index 000000000..adae1c7e1 --- /dev/null +++ b/client/lib/fifo/fifo_windows.go @@ -0,0 +1,105 @@ +package fifo + +import ( + "io" + "net" + "os" + "sync" + "time" + + winio "github.com/Microsoft/go-winio" +) + +// PipeBufferSize is the size of the input and output buffers for the windows +// named pipe +const PipeBufferSize = int32(^uint16(0)) + +type winFIFO struct { + listener net.Listener + conn net.Conn + connLock sync.Mutex +} + +func (f *winFIFO) Read(p []byte) (n int, err error) { + f.connLock.Lock() + defer f.connLock.Unlock() + if f.conn == nil { + c, err := f.listener.Accept() + if err != nil { + return 0, err + } + + f.conn = c + } + + // If the connection is closed then we need to close the listener + // to emulate unix fifo behavior + n, err = f.conn.Read(p) + if err == io.EOF { + f.listener.Close() + } + return n, err +} + +func (f *winFIFO) Write(p []byte) (n int, err error) { + f.connLock.Lock() + defer f.connLock.Unlock() + if f.conn == nil { + c, err := f.listener.Accept() + if err != nil { + return 0, err + } + + f.conn = c + } + + // If the connection is closed then we need to close the listener + // to emulate unix fifo behavior + n, err = f.conn.Write(p) + if err == io.EOF { + f.listener.Close() + } + return n, err + +} + +func (f *winFIFO) Close() error { + return f.listener.Close() +} + +// New creates a fifo at the given path and returns an io.ReadWriteCloser for it. The fifo +// must not already exist +func New(path string) (io.ReadWriteCloser, error) { + l, err := winio.ListenPipe(path, &winio.PipeConfig{ + InputBufferSize: PipeBufferSize, + OutputBufferSize: PipeBufferSize, + }) + if err != nil { + return nil, err + } + + return &winFIFO{ + listener: l, + }, nil +} + +// OpenWriter opens a fifo that already exists and returns an io.ReadWriteCloser for it +func Open(path string) (io.ReadWriteCloser, error) { + return winio.DialPipe(path, nil) +} + +// Remove a fifo that already exists at a given path +func Remove(path string) error { + dur := 500 * time.Millisecond + conn, err := winio.DialPipe(path, &dur) + if err == nil { + return conn.Close() + } + + os.Remove(path) + return nil +} + +func IsClosedErr(err error) bool { + return err == winio.ErrFileClosed +} diff --git a/vendor/github.com/Microsoft/go-winio/backup.go b/vendor/github.com/Microsoft/go-winio/backup.go index 27d6ace0c..2be34af43 100644 --- a/vendor/github.com/Microsoft/go-winio/backup.go +++ b/vendor/github.com/Microsoft/go-winio/backup.go @@ -68,10 +68,20 @@ func NewBackupStreamReader(r io.Reader) *BackupStreamReader { return &BackupStreamReader{r, 0} } -// Next returns the next backup stream and prepares for calls to Write(). It skips the remainder of the current stream if +// Next returns the next backup stream and prepares for calls to Read(). It skips the remainder of the current stream if // it was not completely read. func (r *BackupStreamReader) Next() (*BackupHeader, error) { if r.bytesLeft > 0 { + if s, ok := r.r.(io.Seeker); ok { + // Make sure Seek on io.SeekCurrent sometimes succeeds + // before trying the actual seek. + if _, err := s.Seek(0, io.SeekCurrent); err == nil { + if _, err = s.Seek(r.bytesLeft, io.SeekCurrent); err != nil { + return nil, err + } + r.bytesLeft = 0 + } + } if _, err := io.Copy(ioutil.Discard, r); err != nil { return nil, err } @@ -220,7 +230,7 @@ type BackupFileWriter struct { ctx uintptr } -// NewBackupFileWrtier returns a new BackupFileWriter from a file handle. If includeSecurity is true, +// NewBackupFileWriter returns a new BackupFileWriter from a file handle. If includeSecurity is true, // Write() will attempt to restore the security descriptor from the stream. func NewBackupFileWriter(f *os.File, includeSecurity bool) *BackupFileWriter { w := &BackupFileWriter{f, includeSecurity, 0} diff --git a/vendor/github.com/Microsoft/go-winio/ea.go b/vendor/github.com/Microsoft/go-winio/ea.go new file mode 100644 index 000000000..4051c1b33 --- /dev/null +++ b/vendor/github.com/Microsoft/go-winio/ea.go @@ -0,0 +1,137 @@ +package winio + +import ( + "bytes" + "encoding/binary" + "errors" +) + +type fileFullEaInformation struct { + NextEntryOffset uint32 + Flags uint8 + NameLength uint8 + ValueLength uint16 +} + +var ( + fileFullEaInformationSize = binary.Size(&fileFullEaInformation{}) + + errInvalidEaBuffer = errors.New("invalid extended attribute buffer") + errEaNameTooLarge = errors.New("extended attribute name too large") + errEaValueTooLarge = errors.New("extended attribute value too large") +) + +// ExtendedAttribute represents a single Windows EA. +type ExtendedAttribute struct { + Name string + Value []byte + Flags uint8 +} + +func parseEa(b []byte) (ea ExtendedAttribute, nb []byte, err error) { + var info fileFullEaInformation + err = binary.Read(bytes.NewReader(b), binary.LittleEndian, &info) + if err != nil { + err = errInvalidEaBuffer + return + } + + nameOffset := fileFullEaInformationSize + nameLen := int(info.NameLength) + valueOffset := nameOffset + int(info.NameLength) + 1 + valueLen := int(info.ValueLength) + nextOffset := int(info.NextEntryOffset) + if valueLen+valueOffset > len(b) || nextOffset < 0 || nextOffset > len(b) { + err = errInvalidEaBuffer + return + } + + ea.Name = string(b[nameOffset : nameOffset+nameLen]) + ea.Value = b[valueOffset : valueOffset+valueLen] + ea.Flags = info.Flags + if info.NextEntryOffset != 0 { + nb = b[info.NextEntryOffset:] + } + return +} + +// DecodeExtendedAttributes decodes a list of EAs from a FILE_FULL_EA_INFORMATION +// buffer retrieved from BackupRead, ZwQueryEaFile, etc. +func DecodeExtendedAttributes(b []byte) (eas []ExtendedAttribute, err error) { + for len(b) != 0 { + ea, nb, err := parseEa(b) + if err != nil { + return nil, err + } + + eas = append(eas, ea) + b = nb + } + return +} + +func writeEa(buf *bytes.Buffer, ea *ExtendedAttribute, last bool) error { + if int(uint8(len(ea.Name))) != len(ea.Name) { + return errEaNameTooLarge + } + if int(uint16(len(ea.Value))) != len(ea.Value) { + return errEaValueTooLarge + } + entrySize := uint32(fileFullEaInformationSize + len(ea.Name) + 1 + len(ea.Value)) + withPadding := (entrySize + 3) &^ 3 + nextOffset := uint32(0) + if !last { + nextOffset = withPadding + } + info := fileFullEaInformation{ + NextEntryOffset: nextOffset, + Flags: ea.Flags, + NameLength: uint8(len(ea.Name)), + ValueLength: uint16(len(ea.Value)), + } + + err := binary.Write(buf, binary.LittleEndian, &info) + if err != nil { + return err + } + + _, err = buf.Write([]byte(ea.Name)) + if err != nil { + return err + } + + err = buf.WriteByte(0) + if err != nil { + return err + } + + _, err = buf.Write(ea.Value) + if err != nil { + return err + } + + _, err = buf.Write([]byte{0, 0, 0}[0 : withPadding-entrySize]) + if err != nil { + return err + } + + return nil +} + +// EncodeExtendedAttributes encodes a list of EAs into a FILE_FULL_EA_INFORMATION +// buffer for use with BackupWrite, ZwSetEaFile, etc. +func EncodeExtendedAttributes(eas []ExtendedAttribute) ([]byte, error) { + var buf bytes.Buffer + for i := range eas { + last := false + if i == len(eas)-1 { + last = true + } + + err := writeEa(&buf, &eas[i], last) + if err != nil { + return nil, err + } + } + return buf.Bytes(), nil +} diff --git a/vendor/github.com/Microsoft/go-winio/file.go b/vendor/github.com/Microsoft/go-winio/file.go index 613f31b52..4334ff1cb 100644 --- a/vendor/github.com/Microsoft/go-winio/file.go +++ b/vendor/github.com/Microsoft/go-winio/file.go @@ -16,13 +16,19 @@ import ( //sys createIoCompletionPort(file syscall.Handle, port syscall.Handle, key uintptr, threadCount uint32) (newport syscall.Handle, err error) = CreateIoCompletionPort //sys getQueuedCompletionStatus(port syscall.Handle, bytes *uint32, key *uintptr, o **ioOperation, timeout uint32) (err error) = GetQueuedCompletionStatus //sys setFileCompletionNotificationModes(h syscall.Handle, flags uint8) (err error) = SetFileCompletionNotificationModes -//sys timeBeginPeriod(period uint32) (n int32) = winmm.timeBeginPeriod type atomicBool int32 func (b *atomicBool) isSet() bool { return atomic.LoadInt32((*int32)(b)) != 0 } func (b *atomicBool) setFalse() { atomic.StoreInt32((*int32)(b), 0) } func (b *atomicBool) setTrue() { atomic.StoreInt32((*int32)(b), 1) } +func (b *atomicBool) swap(new bool) bool { + var newInt int32 + if new { + newInt = 1 + } + return atomic.SwapInt32((*int32)(b), newInt) == 1 +} const ( cFILE_SKIP_COMPLETION_PORT_ON_SUCCESS = 1 @@ -71,7 +77,8 @@ func initIo() { type win32File struct { handle syscall.Handle wg sync.WaitGroup - closing bool + wgLock sync.RWMutex + closing atomicBool readDeadline deadlineHandler writeDeadline deadlineHandler } @@ -107,14 +114,18 @@ func MakeOpenFile(h syscall.Handle) (io.ReadWriteCloser, error) { // closeHandle closes the resources associated with a Win32 handle func (f *win32File) closeHandle() { - if !f.closing { + f.wgLock.Lock() + // Atomically set that we are closing, releasing the resources only once. + if !f.closing.swap(true) { + f.wgLock.Unlock() // cancel all IO and wait for it to complete - f.closing = true cancelIoEx(f.handle, nil) f.wg.Wait() // at this point, no new IO can start syscall.Close(f.handle) f.handle = 0 + } else { + f.wgLock.Unlock() } } @@ -127,10 +138,13 @@ func (f *win32File) Close() error { // prepareIo prepares for a new IO operation. // The caller must call f.wg.Done() when the IO is finished, prior to Close() returning. func (f *win32File) prepareIo() (*ioOperation, error) { - f.wg.Add(1) - if f.closing { + f.wgLock.RLock() + if f.closing.isSet() { + f.wgLock.RUnlock() return nil, ErrFileClosed } + f.wg.Add(1) + f.wgLock.RUnlock() c := &ioOperation{} c.ch = make(chan ioResult) return c, nil @@ -138,8 +152,6 @@ func (f *win32File) prepareIo() (*ioOperation, error) { // ioCompletionProcessor processes completed async IOs forever func ioCompletionProcessor(h syscall.Handle) { - // Set the timer resolution to 1. This fixes a performance regression in golang 1.6. - timeBeginPeriod(1) for { var bytes uint32 var key uintptr @@ -159,7 +171,7 @@ func (f *win32File) asyncIo(c *ioOperation, d *deadlineHandler, bytes uint32, er return int(bytes), err } - if f.closing { + if f.closing.isSet() { cancelIoEx(f.handle, &c.o) } @@ -175,7 +187,7 @@ func (f *win32File) asyncIo(c *ioOperation, d *deadlineHandler, bytes uint32, er case r = <-c.ch: err = r.err if err == syscall.ERROR_OPERATION_ABORTED { - if f.closing { + if f.closing.isSet() { err = ErrFileClosed } } diff --git a/vendor/github.com/Microsoft/go-winio/fileinfo.go b/vendor/github.com/Microsoft/go-winio/fileinfo.go index b1d60abb8..ada2fbab6 100644 --- a/vendor/github.com/Microsoft/go-winio/fileinfo.go +++ b/vendor/github.com/Microsoft/go-winio/fileinfo.go @@ -20,7 +20,8 @@ const ( // FileBasicInfo contains file access time and file attributes information. type FileBasicInfo struct { CreationTime, LastAccessTime, LastWriteTime, ChangeTime syscall.Filetime - FileAttributes uintptr // includes padding + FileAttributes uint32 + pad uint32 // padding } // GetFileBasicInfo retrieves times and attributes for a file. diff --git a/vendor/github.com/Microsoft/go-winio/pipe.go b/vendor/github.com/Microsoft/go-winio/pipe.go index da706cc8a..d99eedb64 100644 --- a/vendor/github.com/Microsoft/go-winio/pipe.go +++ b/vendor/github.com/Microsoft/go-winio/pipe.go @@ -15,13 +15,13 @@ import ( //sys connectNamedPipe(pipe syscall.Handle, o *syscall.Overlapped) (err error) = ConnectNamedPipe //sys createNamedPipe(name string, flags uint32, pipeMode uint32, maxInstances uint32, outSize uint32, inSize uint32, defaultTimeout uint32, sa *syscall.SecurityAttributes) (handle syscall.Handle, err error) [failretval==syscall.InvalidHandle] = CreateNamedPipeW //sys createFile(name string, access uint32, mode uint32, sa *syscall.SecurityAttributes, createmode uint32, attrs uint32, templatefile syscall.Handle) (handle syscall.Handle, err error) [failretval==syscall.InvalidHandle] = CreateFileW -//sys waitNamedPipe(name string, timeout uint32) (err error) = WaitNamedPipeW //sys getNamedPipeInfo(pipe syscall.Handle, flags *uint32, outSize *uint32, inSize *uint32, maxInstances *uint32) (err error) = GetNamedPipeInfo //sys getNamedPipeHandleState(pipe syscall.Handle, state *uint32, curInstances *uint32, maxCollectionCount *uint32, collectDataTimeout *uint32, userName *uint16, maxUserNameSize uint32) (err error) = GetNamedPipeHandleStateW //sys localAlloc(uFlags uint32, length uint32) (ptr uintptr) = LocalAlloc const ( cERROR_PIPE_BUSY = syscall.Errno(231) + cERROR_NO_DATA = syscall.Errno(232) cERROR_PIPE_CONNECTED = syscall.Errno(535) cERROR_SEM_TIMEOUT = syscall.Errno(121) @@ -120,6 +120,11 @@ func (f *win32MessageBytePipe) Read(b []byte) (int, error) { // zero-byte message, ensure that all future Read() calls // also return EOF. f.readEOF = true + } else if err == syscall.ERROR_MORE_DATA { + // ERROR_MORE_DATA indicates that the pipe's read mode is message mode + // and the message still has more bytes. Treat this as a success, since + // this package presents all named pipes as byte streams. + err = nil } return n, err } @@ -133,12 +138,14 @@ func (s pipeAddress) String() string { } // DialPipe connects to a named pipe by path, timing out if the connection -// takes longer than the specified duration. If timeout is nil, then the timeout -// is the default timeout established by the pipe server. +// takes longer than the specified duration. If timeout is nil, then we use +// a default timeout of 5 seconds. (We do not use WaitNamedPipe.) func DialPipe(path string, timeout *time.Duration) (net.Conn, error) { var absTimeout time.Time if timeout != nil { absTimeout = time.Now().Add(*timeout) + } else { + absTimeout = time.Now().Add(time.Second * 2) } var err error var h syscall.Handle @@ -147,22 +154,13 @@ func DialPipe(path string, timeout *time.Duration) (net.Conn, error) { if err != cERROR_PIPE_BUSY { break } - now := time.Now() - var ms uint32 - if absTimeout.IsZero() { - ms = cNMPWAIT_USE_DEFAULT_WAIT - } else if now.After(absTimeout) { - ms = cNMPWAIT_NOWAIT - } else { - ms = uint32(absTimeout.Sub(now).Nanoseconds() / 1000 / 1000) - } - err = waitNamedPipe(path, ms) - if err != nil { - if err == cERROR_SEM_TIMEOUT { - return nil, ErrTimeout - } - break + if time.Now().After(absTimeout) { + return nil, ErrTimeout } + + // Wait 10 msec and try again. This is a rather simplistic + // view, as we always try each 10 milliseconds. + time.Sleep(time.Millisecond * 10) } if err != nil { return nil, &os.PathError{Op: "open", Path: path, Err: err} @@ -174,16 +172,6 @@ func DialPipe(path string, timeout *time.Duration) (net.Conn, error) { return nil, err } - var state uint32 - err = getNamedPipeHandleState(h, &state, nil, nil, nil, nil, 0) - if err != nil { - return nil, err - } - - if state&cPIPE_READMODE_MESSAGE != 0 { - return nil, &os.PathError{Op: "open", Path: path, Err: errors.New("message readmode pipes not supported")} - } - f, err := makeWin32File(h) if err != nil { syscall.Close(h) @@ -254,6 +242,36 @@ func (l *win32PipeListener) makeServerPipe() (*win32File, error) { return f, nil } +func (l *win32PipeListener) makeConnectedServerPipe() (*win32File, error) { + p, err := l.makeServerPipe() + if err != nil { + return nil, err + } + + // Wait for the client to connect. + ch := make(chan error) + go func(p *win32File) { + ch <- connectPipe(p) + }(p) + + select { + case err = <-ch: + if err != nil { + p.Close() + p = nil + } + case <-l.closeCh: + // Abort the connect request by closing the handle. + p.Close() + p = nil + err = <-ch + if err == nil || err == ErrFileClosed { + err = ErrPipeListenerClosed + } + } + return p, err +} + func (l *win32PipeListener) listenerRoutine() { closed := false for !closed { @@ -261,31 +279,20 @@ func (l *win32PipeListener) listenerRoutine() { case <-l.closeCh: closed = true case responseCh := <-l.acceptCh: - p, err := l.makeServerPipe() - if err == nil { - // Wait for the client to connect. - ch := make(chan error) - go func() { - ch <- connectPipe(p) - }() - select { - case err = <-ch: - if err != nil { - p.Close() - p = nil - } - case <-l.closeCh: - // Abort the connect request by closing the handle. - p.Close() - p = nil - err = <-ch - if err == nil || err == ErrFileClosed { - err = ErrPipeListenerClosed - } - closed = true + var ( + p *win32File + err error + ) + for { + p, err = l.makeConnectedServerPipe() + // If the connection was immediately closed by the client, try + // again. + if err != cERROR_NO_DATA { + break } } responseCh <- acceptResponse{p, err} + closed = err == ErrPipeListenerClosed } } syscall.Close(l.firstHandle) @@ -334,13 +341,23 @@ func ListenPipe(path string, c *PipeConfig) (net.Listener, error) { if err != nil { return nil, err } - // Immediately open and then close a client handle so that the named pipe is - // created but not currently accepting connections. + // Create a client handle and connect it. This results in the pipe + // instance always existing, so that clients see ERROR_PIPE_BUSY + // rather than ERROR_FILE_NOT_FOUND. This ties the first instance + // up so that no other instances can be used. This would have been + // cleaner if the Win32 API matched CreateFile with ConnectNamedPipe + // instead of CreateNamedPipe. (Apparently created named pipes are + // considered to be in listening state regardless of whether any + // active calls to ConnectNamedPipe are outstanding.) h2, err := createFile(path, 0, 0, nil, syscall.OPEN_EXISTING, cSECURITY_SQOS_PRESENT|cSECURITY_ANONYMOUS, 0) if err != nil { syscall.Close(h) return nil, err } + // Close the client handle. The server side of the instance will + // still be busy, leading to ERROR_PIPE_BUSY instead of + // ERROR_NOT_FOUND, as long as we don't close the server handle, + // or disconnect the client with DisconnectNamedPipe. syscall.Close(h2) l := &win32PipeListener{ firstHandle: h, diff --git a/vendor/github.com/Microsoft/go-winio/zsyscall_windows.go b/vendor/github.com/Microsoft/go-winio/zsyscall_windows.go index 4f7a52eeb..3f527639a 100644 --- a/vendor/github.com/Microsoft/go-winio/zsyscall_windows.go +++ b/vendor/github.com/Microsoft/go-winio/zsyscall_windows.go @@ -38,14 +38,12 @@ func errnoErr(e syscall.Errno) error { var ( modkernel32 = windows.NewLazySystemDLL("kernel32.dll") - modwinmm = windows.NewLazySystemDLL("winmm.dll") modadvapi32 = windows.NewLazySystemDLL("advapi32.dll") procCancelIoEx = modkernel32.NewProc("CancelIoEx") procCreateIoCompletionPort = modkernel32.NewProc("CreateIoCompletionPort") procGetQueuedCompletionStatus = modkernel32.NewProc("GetQueuedCompletionStatus") procSetFileCompletionNotificationModes = modkernel32.NewProc("SetFileCompletionNotificationModes") - proctimeBeginPeriod = modwinmm.NewProc("timeBeginPeriod") procConnectNamedPipe = modkernel32.NewProc("ConnectNamedPipe") procCreateNamedPipeW = modkernel32.NewProc("CreateNamedPipeW") procCreateFileW = modkernel32.NewProc("CreateFileW") @@ -122,12 +120,6 @@ func setFileCompletionNotificationModes(h syscall.Handle, flags uint8) (err erro return } -func timeBeginPeriod(period uint32) (n int32) { - r0, _, _ := syscall.Syscall(proctimeBeginPeriod.Addr(), 1, uintptr(period), 0, 0) - n = int32(r0) - return -} - func connectNamedPipe(pipe syscall.Handle, o *syscall.Overlapped) (err error) { r1, _, e1 := syscall.Syscall(procConnectNamedPipe.Addr(), 2, uintptr(pipe), uintptr(unsafe.Pointer(o)), 0) if r1 == 0 { diff --git a/vendor/github.com/containerd/fifo/LICENSE b/vendor/github.com/containerd/fifo/LICENSE new file mode 100644 index 000000000..261eeb9e9 --- /dev/null +++ b/vendor/github.com/containerd/fifo/LICENSE @@ -0,0 +1,201 @@ + Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "[]" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright [yyyy] [name of copyright owner] + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/vendor/github.com/containerd/fifo/Makefile b/vendor/github.com/containerd/fifo/Makefile new file mode 100644 index 000000000..c10cb3c95 --- /dev/null +++ b/vendor/github.com/containerd/fifo/Makefile @@ -0,0 +1,27 @@ +# Copyright The containerd Authors. + +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at + +# http://www.apache.org/licenses/LICENSE-2.0 + +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +.PHONY: fmt vet test deps + +test: deps + go test -v ./... + +deps: + go get -d -t ./... + +fmt: + gofmt -s -l . + +vet: + go vet ./... diff --git a/vendor/github.com/containerd/fifo/fifo.go b/vendor/github.com/containerd/fifo/fifo.go new file mode 100644 index 000000000..e79813da7 --- /dev/null +++ b/vendor/github.com/containerd/fifo/fifo.go @@ -0,0 +1,236 @@ +/* + Copyright The containerd Authors. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +package fifo + +import ( + "io" + "os" + "runtime" + "sync" + "syscall" + + "github.com/pkg/errors" + "golang.org/x/net/context" +) + +type fifo struct { + flag int + opened chan struct{} + closed chan struct{} + closing chan struct{} + err error + file *os.File + closingOnce sync.Once // close has been called + closedOnce sync.Once // fifo is closed + handle *handle +} + +var leakCheckWg *sync.WaitGroup + +// OpenFifo opens a fifo. Returns io.ReadWriteCloser. +// Context can be used to cancel this function until open(2) has not returned. +// Accepted flags: +// - syscall.O_CREAT - create new fifo if one doesn't exist +// - syscall.O_RDONLY - open fifo only from reader side +// - syscall.O_WRONLY - open fifo only from writer side +// - syscall.O_RDWR - open fifo from both sides, never block on syscall level +// - syscall.O_NONBLOCK - return io.ReadWriteCloser even if other side of the +// fifo isn't open. read/write will be connected after the actual fifo is +// open or after fifo is closed. +func OpenFifo(ctx context.Context, fn string, flag int, perm os.FileMode) (io.ReadWriteCloser, error) { + if _, err := os.Stat(fn); err != nil { + if os.IsNotExist(err) && flag&syscall.O_CREAT != 0 { + if err := mkfifo(fn, uint32(perm&os.ModePerm)); err != nil && !os.IsExist(err) { + return nil, errors.Wrapf(err, "error creating fifo %v", fn) + } + } else { + return nil, err + } + } + + block := flag&syscall.O_NONBLOCK == 0 || flag&syscall.O_RDWR != 0 + + flag &= ^syscall.O_CREAT + flag &= ^syscall.O_NONBLOCK + + h, err := getHandle(fn) + if err != nil { + return nil, err + } + + f := &fifo{ + handle: h, + flag: flag, + opened: make(chan struct{}), + closed: make(chan struct{}), + closing: make(chan struct{}), + } + + wg := leakCheckWg + if wg != nil { + wg.Add(2) + } + + go func() { + if wg != nil { + defer wg.Done() + } + select { + case <-ctx.Done(): + select { + case <-f.opened: + default: + f.Close() + } + case <-f.opened: + case <-f.closed: + } + }() + go func() { + if wg != nil { + defer wg.Done() + } + var file *os.File + fn, err := h.Path() + if err == nil { + file, err = os.OpenFile(fn, flag, 0) + } + select { + case <-f.closing: + if err == nil { + select { + case <-ctx.Done(): + err = ctx.Err() + default: + err = errors.Errorf("fifo %v was closed before opening", h.Name()) + } + if file != nil { + file.Close() + } + } + default: + } + if err != nil { + f.closedOnce.Do(func() { + f.err = err + close(f.closed) + }) + return + } + f.file = file + close(f.opened) + }() + if block { + select { + case <-f.opened: + case <-f.closed: + return nil, f.err + } + } + return f, nil +} + +// Read from a fifo to a byte array. +func (f *fifo) Read(b []byte) (int, error) { + if f.flag&syscall.O_WRONLY > 0 { + return 0, errors.New("reading from write-only fifo") + } + select { + case <-f.opened: + return f.file.Read(b) + default: + } + select { + case <-f.opened: + return f.file.Read(b) + case <-f.closed: + return 0, errors.New("reading from a closed fifo") + } +} + +// Write from byte array to a fifo. +func (f *fifo) Write(b []byte) (int, error) { + if f.flag&(syscall.O_WRONLY|syscall.O_RDWR) == 0 { + return 0, errors.New("writing to read-only fifo") + } + select { + case <-f.opened: + return f.file.Write(b) + default: + } + select { + case <-f.opened: + return f.file.Write(b) + case <-f.closed: + return 0, errors.New("writing to a closed fifo") + } +} + +// Close the fifo. Next reads/writes will error. This method can also be used +// before open(2) has returned and fifo was never opened. +func (f *fifo) Close() (retErr error) { + for { + select { + case <-f.closed: + f.handle.Close() + return + default: + select { + case <-f.opened: + f.closedOnce.Do(func() { + retErr = f.file.Close() + f.err = retErr + close(f.closed) + }) + default: + if f.flag&syscall.O_RDWR != 0 { + runtime.Gosched() + break + } + f.closingOnce.Do(func() { + close(f.closing) + }) + reverseMode := syscall.O_WRONLY + if f.flag&syscall.O_WRONLY > 0 { + reverseMode = syscall.O_RDONLY + } + fn, err := f.handle.Path() + // if Close() is called concurrently(shouldn't) it may cause error + // because handle is closed + select { + case <-f.closed: + default: + if err != nil { + // Path has become invalid. We will leak a goroutine. + // This case should not happen in linux. + f.closedOnce.Do(func() { + f.err = err + close(f.closed) + }) + <-f.closed + break + } + f, err := os.OpenFile(fn, reverseMode|syscall.O_NONBLOCK, 0) + if err == nil { + f.Close() + } + runtime.Gosched() + } + } + } + } +} diff --git a/vendor/github.com/containerd/fifo/handle_linux.go b/vendor/github.com/containerd/fifo/handle_linux.go new file mode 100644 index 000000000..6ac89b6a4 --- /dev/null +++ b/vendor/github.com/containerd/fifo/handle_linux.go @@ -0,0 +1,97 @@ +// +build linux + +/* + Copyright The containerd Authors. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +package fifo + +import ( + "fmt" + "os" + "sync" + "syscall" + + "github.com/pkg/errors" +) + +const O_PATH = 010000000 + +type handle struct { + f *os.File + fd uintptr + dev uint64 + ino uint64 + closeOnce sync.Once + name string +} + +func getHandle(fn string) (*handle, error) { + f, err := os.OpenFile(fn, O_PATH, 0) + if err != nil { + return nil, errors.Wrapf(err, "failed to open %v with O_PATH", fn) + } + + var ( + stat syscall.Stat_t + fd = f.Fd() + ) + if err := syscall.Fstat(int(fd), &stat); err != nil { + f.Close() + return nil, errors.Wrapf(err, "failed to stat handle %v", fd) + } + + h := &handle{ + f: f, + name: fn, + dev: uint64(stat.Dev), + ino: stat.Ino, + fd: fd, + } + + // check /proc just in case + if _, err := os.Stat(h.procPath()); err != nil { + f.Close() + return nil, errors.Wrapf(err, "couldn't stat %v", h.procPath()) + } + + return h, nil +} + +func (h *handle) procPath() string { + return fmt.Sprintf("/proc/self/fd/%d", h.fd) +} + +func (h *handle) Name() string { + return h.name +} + +func (h *handle) Path() (string, error) { + var stat syscall.Stat_t + if err := syscall.Stat(h.procPath(), &stat); err != nil { + return "", errors.Wrapf(err, "path %v could not be statted", h.procPath()) + } + if uint64(stat.Dev) != h.dev || stat.Ino != h.ino { + return "", errors.Errorf("failed to verify handle %v/%v %v/%v", stat.Dev, h.dev, stat.Ino, h.ino) + } + return h.procPath(), nil +} + +func (h *handle) Close() error { + h.closeOnce.Do(func() { + h.f.Close() + }) + return nil +} diff --git a/vendor/github.com/containerd/fifo/handle_nolinux.go b/vendor/github.com/containerd/fifo/handle_nolinux.go new file mode 100644 index 000000000..4f2a282b2 --- /dev/null +++ b/vendor/github.com/containerd/fifo/handle_nolinux.go @@ -0,0 +1,65 @@ +// +build !linux + +/* + Copyright The containerd Authors. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +package fifo + +import ( + "syscall" + + "github.com/pkg/errors" +) + +type handle struct { + fn string + dev uint64 + ino uint64 +} + +func getHandle(fn string) (*handle, error) { + var stat syscall.Stat_t + if err := syscall.Stat(fn, &stat); err != nil { + return nil, errors.Wrapf(err, "failed to stat %v", fn) + } + + h := &handle{ + fn: fn, + dev: uint64(stat.Dev), + ino: uint64(stat.Ino), + } + + return h, nil +} + +func (h *handle) Path() (string, error) { + var stat syscall.Stat_t + if err := syscall.Stat(h.fn, &stat); err != nil { + return "", errors.Wrapf(err, "path %v could not be statted", h.fn) + } + if uint64(stat.Dev) != h.dev || uint64(stat.Ino) != h.ino { + return "", errors.Errorf("failed to verify handle %v/%v %v/%v for %v", stat.Dev, h.dev, stat.Ino, h.ino, h.fn) + } + return h.fn, nil +} + +func (h *handle) Name() string { + return h.fn +} + +func (h *handle) Close() error { + return nil +} diff --git a/vendor/github.com/containerd/fifo/mkfifo_nosolaris.go b/vendor/github.com/containerd/fifo/mkfifo_nosolaris.go new file mode 100644 index 000000000..2799a06d1 --- /dev/null +++ b/vendor/github.com/containerd/fifo/mkfifo_nosolaris.go @@ -0,0 +1,25 @@ +// +build !solaris + +/* + Copyright The containerd Authors. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +package fifo + +import "syscall" + +func mkfifo(path string, mode uint32) (err error) { + return syscall.Mkfifo(path, mode) +} diff --git a/vendor/github.com/containerd/fifo/mkfifo_solaris.go b/vendor/github.com/containerd/fifo/mkfifo_solaris.go new file mode 100644 index 000000000..1ecd722ae --- /dev/null +++ b/vendor/github.com/containerd/fifo/mkfifo_solaris.go @@ -0,0 +1,27 @@ +// +build solaris + +/* + Copyright The containerd Authors. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +package fifo + +import ( + "golang.org/x/sys/unix" +) + +func mkfifo(path string, mode uint32) (err error) { + return unix.Mkfifo(path, mode) +} diff --git a/vendor/github.com/containerd/fifo/readme.md b/vendor/github.com/containerd/fifo/readme.md new file mode 100644 index 000000000..393830f92 --- /dev/null +++ b/vendor/github.com/containerd/fifo/readme.md @@ -0,0 +1,32 @@ +### fifo + +[![Build Status](https://travis-ci.org/containerd/fifo.svg?branch=master)](https://travis-ci.org/containerd/fifo) + +Go package for handling fifos in a sane way. + +``` +// OpenFifo opens a fifo. Returns io.ReadWriteCloser. +// Context can be used to cancel this function until open(2) has not returned. +// Accepted flags: +// - syscall.O_CREAT - create new fifo if one doesn't exist +// - syscall.O_RDONLY - open fifo only from reader side +// - syscall.O_WRONLY - open fifo only from writer side +// - syscall.O_RDWR - open fifo from both sides, never block on syscall level +// - syscall.O_NONBLOCK - return io.ReadWriteCloser even if other side of the +// fifo isn't open. read/write will be connected after the actual fifo is +// open or after fifo is closed. +func OpenFifo(ctx context.Context, fn string, flag int, perm os.FileMode) (io.ReadWriteCloser, error) + + +// Read from a fifo to a byte array. +func (f *fifo) Read(b []byte) (int, error) + + +// Write from byte array to a fifo. +func (f *fifo) Write(b []byte) (int, error) + + +// Close the fifo. Next reads/writes will error. This method can also be used +// before open(2) has returned and fifo was never opened. +func (f *fifo) Close() error +``` diff --git a/vendor/vendor.json b/vendor/vendor.json index 059c9cd56..2d9147b03 100644 --- a/vendor/vendor.json +++ b/vendor/vendor.json @@ -7,7 +7,7 @@ {"path":"github.com/Azure/go-ansiterm","checksumSHA1":"9NFR6RG8H2fNyKHscGmuGLQhRm4=","revision":"d6e3b3328b783f23731bc4d058875b0371ff8109","revisionTime":"2017-09-29T23:40:23Z","version":"master","versionExact":"master"}, {"path":"github.com/Azure/go-ansiterm/winterm","checksumSHA1":"3/UphB+6Hbx5otA4PjFjvObT+L4=","revision":"d6e3b3328b783f23731bc4d058875b0371ff8109","revisionTime":"2017-09-29T23:40:23Z","version":"master","versionExact":"master"}, {"path":"github.com/DataDog/datadog-go/statsd","checksumSHA1":"WvApwvvSe3i/3KO8300dyeFmkbI=","revision":"b10af4b12965a1ad08d164f57d14195b4140d8de","revisionTime":"2017-08-09T10:47:06Z"}, - {"path":"github.com/Microsoft/go-winio","checksumSHA1":"AzjRkOQtVBTwIw4RJLTygFhJs3s=","revision":"f533f7a102197536779ea3a8cb881d639e21ec5a","revisionTime":"2017-05-24T00:36:31Z"}, + {"path":"github.com/Microsoft/go-winio","checksumSHA1":"PbR6ZKoLeSZl8aXxDQqXih0wSgE=","revision":"97e4973ce50b2ff5f09635a57e2b88a037aae829","revisionTime":"2018-08-23T22:24:21Z"}, {"path":"github.com/NVIDIA/gpu-monitoring-tools","checksumSHA1":"kF1vk+8Xvb3nGBiw9+qbUc0SZ4M=","revision":"86f2a9fac6c5b597dc494420005144b8ef7ec9fb","revisionTime":"2018-08-29T22:20:09Z"}, {"path":"github.com/NVIDIA/gpu-monitoring-tools/bindings/go/nvml","checksumSHA1":"P8FATSSgpe5A17FyPrGpsX95Xw8=","revision":"86f2a9fac6c5b597dc494420005144b8ef7ec9fb","revisionTime":"2018-08-29T22:20:09Z"}, {"path":"github.com/NYTimes/gziphandler","checksumSHA1":"jktW57+vJsziNVPeXMCoujTzdW4=","revision":"97ae7fbaf81620fe97840685304a78a306a39c64","revisionTime":"2017-09-16T00:36:49Z"}, @@ -60,6 +60,7 @@ {"path":"github.com/circonus-labs/circonusllhist","checksumSHA1":"VbfeVqeOM+dTNxCmpvmYS0LwQn0=","revision":"7d649b46cdc2cd2ed102d350688a75a4fd7778c6","revisionTime":"2016-11-21T13:51:53Z"}, {"path":"github.com/containerd/console","checksumSHA1":"IGtuR58l2zmYRcNf8sPDlCSgovE=","origin":"github.com/opencontainers/runc/vendor/github.com/containerd/console","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, {"path":"github.com/containerd/continuity/pathdriver","checksumSHA1":"GqIrOttKaO7k6HIaHQLPr3cY7rY=","origin":"github.com/docker/docker/vendor/github.com/containerd/continuity/pathdriver","revision":"320063a2ad06a1d8ada61c94c29dbe44e2d87473","revisionTime":"2018-08-16T08:14:46Z"}, + {"path":"github.com/containerd/fifo","checksumSHA1":"Ur3lVmFp+HTGUzQU+/ZBolKe8FU=","revision":"3d5202aec260678c48179c56f40e6f38a095738c","revisionTime":"2018-03-07T16:51:37Z"}, {"path":"github.com/containernetworking/cni/pkg/types","checksumSHA1":"NeAp/3+Hedu9tnMai+LihERPj84=","revision":"5c3c17164270150467498a32c71436c7cd5501be","revisionTime":"2016-06-02T16:00:07Z"}, {"path":"github.com/coreos/go-semver/semver","checksumSHA1":"97BsbXOiZ8+Kr+LIuZkQFtSj7H4=","revision":"1817cd4bea52af76542157eeabd74b057d1a199e","revisionTime":"2017-06-13T09:22:38Z"}, {"path":"github.com/coreos/go-systemd/dbus","checksumSHA1":"/zxxFPYjUB7Wowz33r5AhTDvoz0=","origin":"github.com/opencontainers/runc/vendor/github.com/coreos/go-systemd/dbus","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, From cebfead6bca255339ecbaa9c71bbb78cc11470bb Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Mon, 17 Sep 2018 15:44:37 -0700 Subject: [PATCH 066/152] add logger back --- client/client.go | 1 + 1 file changed, 1 insertion(+) diff --git a/client/client.go b/client/client.go index eef3f4e16..09f6d063e 100644 --- a/client/client.go +++ b/client/client.go @@ -240,6 +240,7 @@ func NewClient(cfg *config.Config, consulCatalog consul.CatalogAPI, consulServic connPool: pool.NewPool(logger, clientRPCCache, clientMaxStreams, tlsWrap), tlsWrap: tlsWrap, streamingRpcs: structs.NewStreamingRpcRegistry(), + logger: logger, rpcLogger: logger.Named("rpc"), allocs: make(map[string]AllocRunner), allocUpdates: make(chan *structs.Allocation, 64), From 6f92b04226820cfbc87d5d9a7e024a26cde60112 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 13 Sep 2018 17:25:01 -0700 Subject: [PATCH 067/152] health_hook: simplify locking; test thoroughly Use doneCh like @dadgar suggested in the original PR. Thoroughly test hook as concurrent Update calls make for a tricky concurrency problem. --- client/allocrunnerv2/health_hook.go | 52 ++--- client/allocrunnerv2/health_hook_test.go | 272 +++++++++++++++++------ 2 files changed, 227 insertions(+), 97 deletions(-) diff --git a/client/allocrunnerv2/health_hook.go b/client/allocrunnerv2/health_hook.go index 1b51fdbb3..9aa5e2754 100644 --- a/client/allocrunnerv2/health_hook.go +++ b/client/allocrunnerv2/health_hook.go @@ -14,31 +14,6 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) -// watchSema acts as a semaphore to allow one watcher at a time. -type watchSema struct { - ch chan int -} - -func newWatchSema() *watchSema { - return &watchSema{ch: make(chan int, 1)} -} - -// acquire the semaphore. -func (w *watchSema) acquire() { - w.ch <- 1 -} - -// release the semaphore. -func (w *watchSema) release() { - <-w.ch -} - -// wait blocks until the semaphore is free to be acquired. -func (w *watchSema) wait() { - w.acquire() - w.release() -} - // healthMutator is able to set/clear alloc health. type healthSetter interface { // Set health via the mutator @@ -65,9 +40,10 @@ type allocHealthWatcherHook struct { // Update and synchronous hooks. hookLock sync.Mutex - // watchLock is acquired by the health watching/setting goroutine to - // ensure only one health watching goroutine is running at a time. - watchLock *watchSema + // watchDone is created before calling watchHealth and is closed when + // watchHealth exits. Must be passed into watchHealth to avoid races. + // Initialized already closed as Update may be called before Prerun. + watchDone chan struct{} // ranOnce is set once Prerun or Update have run at least once. This // prevents Prerun from running if an Update has already been @@ -97,13 +73,17 @@ func newAllocHealthWatcherHook(logger log.Logger, alloc *structs.Allocation, hs return noopAllocHealthWatcherHook{} } + // Initialize watchDone with a closed chan in case Update runs before Prerun + closedDone := make(chan struct{}) + close(closedDone) + h := &allocHealthWatcherHook{ alloc: alloc, cancelFn: func() {}, // initialize to prevent nil func panics + watchDone: closedDone, consul: consul, healthSetter: hs, listener: listener, - watchLock: newWatchSema(), } h.logger = logger.Named(h.Name()) @@ -151,7 +131,10 @@ func (h *allocHealthWatcherHook) init() error { tracker := allochealth.NewTracker(ctx, h.logger, h.alloc, h.listener, h.consul, minHealthyTime, useChecks) tracker.Start() - go h.watchHealth(ctx, tracker) + + // Create a new done chan and start watching for health updates + h.watchDone = make(chan struct{}) + go h.watchHealth(ctx, tracker, h.watchDone) return nil } @@ -179,7 +162,7 @@ func (h *allocHealthWatcherHook) Update(req *interfaces.RunnerUpdateRequest) err h.cancelFn() // Wait until the watcher exits - h.watchLock.wait() + <-h.watchDone // Deployment has changed, reset status if req.Alloc.DeploymentID != h.alloc.DeploymentID { @@ -200,7 +183,7 @@ func (h *allocHealthWatcherHook) Destroy() error { h.listener.Close() // Wait until the watcher exits - h.watchLock.wait() + <-h.watchDone return nil } @@ -208,9 +191,8 @@ func (h *allocHealthWatcherHook) Destroy() error { // watchHealth watches alloc health until it is set, the alloc is stopped, or // the context is canceled. watchHealth will be canceled and restarted on // Updates so calls are serialized with a lock. -func (h *allocHealthWatcherHook) watchHealth(ctx context.Context, tracker *allochealth.Tracker) { - h.watchLock.acquire() - defer h.watchLock.release() +func (h *allocHealthWatcherHook) watchHealth(ctx context.Context, tracker *allochealth.Tracker, done chan<- struct{}) { + defer close(done) select { case <-ctx.Done(): diff --git a/client/allocrunnerv2/health_hook_test.go b/client/allocrunnerv2/health_hook_test.go index 42b2d4ac5..db03a0815 100644 --- a/client/allocrunnerv2/health_hook_test.go +++ b/client/allocrunnerv2/health_hook_test.go @@ -6,12 +6,16 @@ import ( "testing" "time" + consulapi "github.com/hashicorp/consul/api" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" "github.com/hashicorp/nomad/client/consul" cstructs "github.com/hashicorp/nomad/client/structs" + agentconsul "github.com/hashicorp/nomad/command/agent/consul" "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/helper/uuid" "github.com/hashicorp/nomad/nomad/mock" "github.com/hashicorp/nomad/nomad/structs" + "github.com/stretchr/testify/assert" "github.com/stretchr/testify/require" ) @@ -20,6 +24,12 @@ var _ interfaces.RunnerPrerunHook = (*allocHealthWatcherHook)(nil) var _ interfaces.RunnerUpdateHook = (*allocHealthWatcherHook)(nil) var _ interfaces.RunnerDestroyHook = (*allocHealthWatcherHook)(nil) +// allocHealth is emitted to a chan whenever SetHealth is called +type allocHealth struct { + healthy bool + taskEvents map[string]*structs.TaskEvent +} + // mockHealthSetter implements healthSetter that stores health internally type mockHealthSetter struct { setCalls int @@ -28,6 +38,17 @@ type mockHealthSetter struct { isDeploy *bool taskEvents map[string]*structs.TaskEvent mu sync.Mutex + + healthCh chan allocHealth +} + +// newMockHealthSetter returns a mock HealthSetter that emits all SetHealth +// calls on a buffered chan. Callers who do need need notifications of health +// changes may just create the struct directly. +func newMockHealthSetter() *mockHealthSetter { + return &mockHealthSetter{ + healthCh: make(chan allocHealth, 1), + } } func (m *mockHealthSetter) SetHealth(healthy, isDeploy bool, taskEvents map[string]*structs.TaskEvent) { @@ -38,6 +59,10 @@ func (m *mockHealthSetter) SetHealth(healthy, isDeploy bool, taskEvents map[stri m.healthy = &healthy m.isDeploy = &isDeploy m.taskEvents = taskEvents + + if m.healthCh != nil { + m.healthCh <- allocHealth{healthy, taskEvents} + } } func (m *mockHealthSetter) ClearHealth() { @@ -50,6 +75,7 @@ func (m *mockHealthSetter) ClearHealth() { m.taskEvents = nil } +// TestHealthHook_PrerunDestroy asserts a health hook does not error if it is run and destroyed. func TestHealthHook_PrerunDestroy(t *testing.T) { t.Parallel() require := require.New(t) @@ -73,10 +99,194 @@ func TestHealthHook_PrerunDestroy(t *testing.T) { // Prerun require.NoError(prerunh.Prerun(context.Background())) + // Assert isDeploy is false (other tests peek at isDeploy to determine + // if an Update applied) + ahw := h.(*allocHealthWatcherHook) + ahw.hookLock.Lock() + assert.False(t, ahw.isDeploy) + ahw.hookLock.Unlock() + // Destroy require.NoError(destroyh.Destroy()) } +// TestHealthHook_PrerunUpdateDestroy asserts Updates may be applied concurrently. +func TestHealthHook_PrerunUpdateDestroy(t *testing.T) { + t.Parallel() + require := require.New(t) + + alloc := mock.Alloc() + + b := cstructs.NewAllocBroadcaster() + defer b.Close() + + consul := consul.NewMockConsulServiceClient(t) + hs := &mockHealthSetter{} + + h := newAllocHealthWatcherHook(testlog.HCLogger(t), alloc.Copy(), hs, b.Listen(), consul).(*allocHealthWatcherHook) + + // Prerun + require.NoError(h.Prerun(context.Background())) + + // Update multiple times in a goroutine to mimic Client behavior + // (Updates are concurrent with alloc runner but are applied serially). + errs := make(chan error, 2) + go func() { + defer close(errs) + for i := 0; i < cap(errs); i++ { + alloc.AllocModifyIndex++ + errs <- h.Update(&interfaces.RunnerUpdateRequest{Alloc: alloc.Copy()}) + } + }() + + for err := range errs { + assert.NoError(t, err) + } + + // Destroy + require.NoError(h.Destroy()) +} + +// TestHealthHook_UpdatePrerunDestroy asserts that a hook may have Update +// called before Prerun. +func TestHealthHook_UpdatePrerunDestroy(t *testing.T) { + t.Parallel() + require := require.New(t) + + alloc := mock.Alloc() + + b := cstructs.NewAllocBroadcaster() + defer b.Close() + + consul := consul.NewMockConsulServiceClient(t) + hs := &mockHealthSetter{} + + h := newAllocHealthWatcherHook(testlog.HCLogger(t), alloc.Copy(), hs, b.Listen(), consul).(*allocHealthWatcherHook) + + // Set a DeploymentID to cause ClearHealth to be called + alloc.DeploymentID = uuid.Generate() + + // Update in a goroutine to mimic Client behavior (Updates are + // concurrent with alloc runner). + errs := make(chan error, 1) + go func(alloc *structs.Allocation) { + errs <- h.Update(&interfaces.RunnerUpdateRequest{Alloc: alloc}) + close(errs) + }(alloc.Copy()) + + for err := range errs { + assert.NoError(t, err) + } + + // Prerun should be a noop + require.NoError(h.Prerun(context.Background())) + + // Assert that the Update took affect by isDeploy being true + h.hookLock.Lock() + assert.True(t, h.isDeploy) + h.hookLock.Unlock() + + // Destroy + require.NoError(h.Destroy()) +} + +// TestHealthHook_Destroy asserts that a hook may have only Destroy called. +func TestHealthHook_Destroy(t *testing.T) { + t.Parallel() + require := require.New(t) + + b := cstructs.NewAllocBroadcaster() + defer b.Close() + + consul := consul.NewMockConsulServiceClient(t) + hs := &mockHealthSetter{} + + h := newAllocHealthWatcherHook(testlog.HCLogger(t), mock.Alloc(), hs, b.Listen(), consul).(*allocHealthWatcherHook) + + // Destroy + require.NoError(h.Destroy()) +} + +// TestHealthHook_SetHealth asserts SetHealth is called when health status is +// set. Uses task state and health checks. +func TestHealthHook_SetHealth(t *testing.T) { + t.Parallel() + require := require.New(t) + + alloc := mock.Alloc() + alloc.Job.TaskGroups[0].Migrate.MinHealthyTime = 1 // let's speed things up + task := alloc.Job.TaskGroups[0].Tasks[0] + + // Synthesize running alloc and tasks + alloc.ClientStatus = structs.AllocClientStatusRunning + alloc.TaskStates = map[string]*structs.TaskState{ + task.Name: { + State: structs.TaskStateRunning, + StartedAt: time.Now(), + }, + } + + // Make Consul response + check := &consulapi.AgentCheck{ + Name: task.Services[0].Checks[0].Name, + Status: consulapi.HealthPassing, + } + taskRegs := map[string]*agentconsul.TaskRegistration{ + task.Name: { + Services: map[string]*agentconsul.ServiceRegistration{ + task.Services[0].Name: { + Service: &consulapi.AgentService{ + ID: "foo", + Service: task.Services[0].Name, + }, + Checks: []*consulapi.AgentCheck{check}, + }, + }, + }, + } + + b := cstructs.NewAllocBroadcaster() + defer b.Close() + + // Don't reply on the first call + called := false + consul := consul.NewMockConsulServiceClient(t) + consul.AllocRegistrationsFn = func(string) (*agentconsul.AllocRegistration, error) { + if !called { + called = true + return nil, nil + } + + reg := &agentconsul.AllocRegistration{ + Tasks: taskRegs, + } + + return reg, nil + } + + hs := newMockHealthSetter() + + h := newAllocHealthWatcherHook(testlog.HCLogger(t), alloc.Copy(), hs, b.Listen(), consul).(*allocHealthWatcherHook) + + // Prerun + require.NoError(h.Prerun(context.Background())) + + // Wait for health to be set (healthy) + select { + case <-time.After(5 * time.Second): + t.Fatalf("timeout waiting for health to be set") + case health := <-hs.healthCh: + require.True(health.healthy) + + // Healthy allocs shouldn't emit task events + ev := health.taskEvents[task.Name] + require.Nilf(ev, "%#v", health.taskEvents) + } + + // Destroy + require.NoError(h.Destroy()) +} + // TestHealthHook_SystemNoop asserts that system jobs return the noop tracker. func TestHealthHook_SystemNoop(t *testing.T) { t.Parallel() @@ -106,65 +316,3 @@ func TestHealthHook_BatchNoop(t *testing.T) { _, ok := h.(noopAllocHealthWatcherHook) require.True(t, ok) } - -// TestHealthHook_WatchSema asserts only one caller can acquire the watchSema -// lock at once and all other callers are blocked. -func TestHealthHook_WatchSema(t *testing.T) { - t.Parallel() - - s := newWatchSema() - - s.acquire() - - // Second acquire should block - acq2 := make(chan struct{}) - go func() { - s.acquire() - close(acq2) - }() - - select { - case <-acq2: - t.Fatalf("should not have been able to acquire lock") - case <-time.After(33 * time.Millisecond): - // Ok! It's blocking! - } - - // Release and assert second acquire is now active - s.release() - - select { - case <-acq2: - // Ok! Second acquire unblocked! - case <-time.After(33 * time.Millisecond): - t.Fatalf("second acquire should not have blocked") - } - - // wait() should block until lock is released - waitCh := make(chan struct{}) - go func() { - s.wait() - close(waitCh) - }() - - select { - case <-waitCh: - t.Fatalf("should not have been able to acquire lock") - case <-time.After(33 * time.Millisecond): - // Ok! It's blocking! - } - - // Release and assert wait() finished - s.release() - - select { - case <-waitCh: - // Ok! wait() unblocked! - case <-time.After(33 * time.Millisecond): - t.Fatalf("wait should not have blocked") - } - - // Assert wait() did not acquire the lock - s.acquire() - s.release() -} From a977e2202883a4d0a899b07117f463cb0860515b Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 13 Sep 2018 17:27:14 -0700 Subject: [PATCH 068/152] test: cleanup mock consul service client Updated to hclog. It exposed fields that required an unexported lock to access. Created a getter methodn instead. Only old allocrunner currently used this feature. --- client/consul/consul_testing.go | 8 +++++++- 1 file changed, 7 insertions(+), 1 deletion(-) diff --git a/client/consul/consul_testing.go b/client/consul/consul_testing.go index a7997a3eb..f161b6f08 100644 --- a/client/consul/consul_testing.go +++ b/client/consul/consul_testing.go @@ -31,7 +31,7 @@ func NewMockConsulOp(op, allocID, task string) MockConsulOp { // MockConsulServiceClient implements the ConsulServiceAPI interface to record // and log task registration/deregistration. type MockConsulServiceClient struct { - Ops []MockConsulOp + ops []MockConsulOp mu sync.Mutex logger log.Logger @@ -85,3 +85,9 @@ func (m *MockConsulServiceClient) AllocRegistrations(allocID string) (*consul.Al return nil, nil } + +func (m *MockConsulServiceClient) GetOps() []MockConsulOp { + m.mu.Lock() + defer m.mu.Unlock() + return m.ops +} From 86bd32953961b83b823be20aa24f0ff6bbcb4983 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 17 Sep 2018 17:44:25 -0700 Subject: [PATCH 069/152] fix build errors post merges --- client/allocrunnerv2/health_hook_test.go | 27 +++++++++++++++--------- client/consul/consul_testing.go | 10 ++++----- client/fingerprint/network_linux.go | 10 ++++----- 3 files changed, 27 insertions(+), 20 deletions(-) diff --git a/client/allocrunnerv2/health_hook_test.go b/client/allocrunnerv2/health_hook_test.go index db03a0815..c555297a6 100644 --- a/client/allocrunnerv2/health_hook_test.go +++ b/client/allocrunnerv2/health_hook_test.go @@ -83,10 +83,12 @@ func TestHealthHook_PrerunDestroy(t *testing.T) { b := cstructs.NewAllocBroadcaster() defer b.Close() - consul := consul.NewMockConsulServiceClient(t) + logger := testlog.HCLogger(t) + + consul := consul.NewMockConsulServiceClient(t, logger) hs := &mockHealthSetter{} - h := newAllocHealthWatcherHook(testlog.HCLogger(t), mock.Alloc(), hs, b.Listen(), consul) + h := newAllocHealthWatcherHook(logger, mock.Alloc(), hs, b.Listen(), consul) // Assert we implemented the right interfaces prerunh, ok := h.(interfaces.RunnerPrerunHook) @@ -120,10 +122,11 @@ func TestHealthHook_PrerunUpdateDestroy(t *testing.T) { b := cstructs.NewAllocBroadcaster() defer b.Close() - consul := consul.NewMockConsulServiceClient(t) + logger := testlog.HCLogger(t) + consul := consul.NewMockConsulServiceClient(t, logger) hs := &mockHealthSetter{} - h := newAllocHealthWatcherHook(testlog.HCLogger(t), alloc.Copy(), hs, b.Listen(), consul).(*allocHealthWatcherHook) + h := newAllocHealthWatcherHook(logger, alloc.Copy(), hs, b.Listen(), consul).(*allocHealthWatcherHook) // Prerun require.NoError(h.Prerun(context.Background())) @@ -158,10 +161,11 @@ func TestHealthHook_UpdatePrerunDestroy(t *testing.T) { b := cstructs.NewAllocBroadcaster() defer b.Close() - consul := consul.NewMockConsulServiceClient(t) + logger := testlog.HCLogger(t) + consul := consul.NewMockConsulServiceClient(t, logger) hs := &mockHealthSetter{} - h := newAllocHealthWatcherHook(testlog.HCLogger(t), alloc.Copy(), hs, b.Listen(), consul).(*allocHealthWatcherHook) + h := newAllocHealthWatcherHook(logger, alloc.Copy(), hs, b.Listen(), consul).(*allocHealthWatcherHook) // Set a DeploymentID to cause ClearHealth to be called alloc.DeploymentID = uuid.Generate() @@ -198,10 +202,11 @@ func TestHealthHook_Destroy(t *testing.T) { b := cstructs.NewAllocBroadcaster() defer b.Close() - consul := consul.NewMockConsulServiceClient(t) + logger := testlog.HCLogger(t) + consul := consul.NewMockConsulServiceClient(t, logger) hs := &mockHealthSetter{} - h := newAllocHealthWatcherHook(testlog.HCLogger(t), mock.Alloc(), hs, b.Listen(), consul).(*allocHealthWatcherHook) + h := newAllocHealthWatcherHook(logger, mock.Alloc(), hs, b.Listen(), consul).(*allocHealthWatcherHook) // Destroy require.NoError(h.Destroy()) @@ -248,9 +253,11 @@ func TestHealthHook_SetHealth(t *testing.T) { b := cstructs.NewAllocBroadcaster() defer b.Close() + logger := testlog.HCLogger(t) + // Don't reply on the first call called := false - consul := consul.NewMockConsulServiceClient(t) + consul := consul.NewMockConsulServiceClient(t, logger) consul.AllocRegistrationsFn = func(string) (*agentconsul.AllocRegistration, error) { if !called { called = true @@ -266,7 +273,7 @@ func TestHealthHook_SetHealth(t *testing.T) { hs := newMockHealthSetter() - h := newAllocHealthWatcherHook(testlog.HCLogger(t), alloc.Copy(), hs, b.Listen(), consul).(*allocHealthWatcherHook) + h := newAllocHealthWatcherHook(logger, alloc.Copy(), hs, b.Listen(), consul).(*allocHealthWatcherHook) // Prerun require.NoError(h.Prerun(context.Background())) diff --git a/client/consul/consul_testing.go b/client/consul/consul_testing.go index f161b6f08..6b266652f 100644 --- a/client/consul/consul_testing.go +++ b/client/consul/consul_testing.go @@ -44,7 +44,7 @@ type MockConsulServiceClient struct { func NewMockConsulServiceClient(t testing.T, logger log.Logger) *MockConsulServiceClient { logger = logger.Named("mock_consul") m := MockConsulServiceClient{ - Ops: make([]MockConsulOp, 0, 20), + ops: make([]MockConsulOp, 0, 20), logger: logger, } return &m @@ -54,7 +54,7 @@ func (m *MockConsulServiceClient) UpdateTask(old, new *consul.TaskServices) erro m.mu.Lock() defer m.mu.Unlock() m.logger.Trace("UpdateTask", "alloc_id", new.AllocID, "task", new.Name) - m.Ops = append(m.Ops, NewMockConsulOp("update", new.AllocID, new.Name)) + m.ops = append(m.ops, NewMockConsulOp("update", new.AllocID, new.Name)) return nil } @@ -62,7 +62,7 @@ func (m *MockConsulServiceClient) RegisterTask(task *consul.TaskServices) error m.mu.Lock() defer m.mu.Unlock() m.logger.Trace("RegisterTask", "alloc_id", task.AllocID, "task", task.Name) - m.Ops = append(m.Ops, NewMockConsulOp("add", task.AllocID, task.Name)) + m.ops = append(m.ops, NewMockConsulOp("add", task.AllocID, task.Name)) return nil } @@ -70,14 +70,14 @@ func (m *MockConsulServiceClient) RemoveTask(task *consul.TaskServices) { m.mu.Lock() defer m.mu.Unlock() m.logger.Trace("RemoveTask", "alloc_id", task.AllocID, "task", task.Name) - m.Ops = append(m.Ops, NewMockConsulOp("remove", task.AllocID, task.Name)) + m.ops = append(m.ops, NewMockConsulOp("remove", task.AllocID, task.Name)) } func (m *MockConsulServiceClient) AllocRegistrations(allocID string) (*consul.AllocRegistration, error) { m.mu.Lock() defer m.mu.Unlock() m.logger.Trace("AllocRegistrations", "alloc_id", allocID) - m.Ops = append(m.Ops, NewMockConsulOp("alloc_registrations", allocID, "")) + m.ops = append(m.ops, NewMockConsulOp("alloc_registrations", allocID, "")) if m.AllocRegistrationsFn != nil { return m.AllocRegistrationsFn(allocID) diff --git a/client/fingerprint/network_linux.go b/client/fingerprint/network_linux.go index 5a1ba001f..ff0ccff70 100644 --- a/client/fingerprint/network_linux.go +++ b/client/fingerprint/network_linux.go @@ -16,14 +16,14 @@ func (f *NetworkFingerprint) linkSpeedSys(device string) int { // Read contents of the device/speed file content, err := ioutil.ReadFile(path) if err != nil { - f.logger.Printf("[DEBUG] fingerprint.network: Unable to read link speed from %s", path) + f.logger.Debug("unable to read link speed", "path", path) return 0 } lines := strings.Split(string(content), "\n") mbs, err := strconv.Atoi(lines[0]) if err != nil || mbs <= 0 { - f.logger.Printf("[DEBUG] fingerprint.network: Unable to parse link speed from %s", path) + f.logger.Debug("unable to parse link speed", "path", path) return 0 } @@ -50,7 +50,7 @@ func (f *NetworkFingerprint) linkSpeed(device string) int { func (f *NetworkFingerprint) linkSpeedEthtool(path, device string) int { outBytes, err := exec.Command(path, device).Output() if err != nil { - f.logger.Printf("[WARN] fingerprint.network: Error calling ethtool (%s %s): %v", path, device, err) + f.logger.Warn("error calling ethtool", "error", err, "path", path, "device", device) return 0 } @@ -59,7 +59,7 @@ func (f *NetworkFingerprint) linkSpeedEthtool(path, device string) int { m := re.FindString(output) if m == "" { // no matches found, output may be in a different format - f.logger.Printf("[WARN] fingerprint.network: Unable to parse Speed in output of '%s %s'", path, device) + f.logger.Warn("unable to parse speed", "path", path, "device", device) return 0 } @@ -70,7 +70,7 @@ func (f *NetworkFingerprint) linkSpeedEthtool(path, device string) int { // convert to Mb/s mbs, err := strconv.Atoi(raw) if err != nil || mbs <= 0 { - f.logger.Printf("[WARN] fingerprint.network: Unable to parse Mb/s in output of '%s %s'", path, device) + f.logger.Warn("unable to parse Mb/s", "path", path, "device", device) return 0 } From e6e2930a00a90d2bdc31a3bd442f01ffa512e709 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Fri, 14 Sep 2018 17:08:26 -0700 Subject: [PATCH 070/152] tr: implement stats collection hook Tested except for the net/rpc specific error case which may need changing in the gRPC world. --- client/allocrunnerv2/alloc_runner.go | 39 +++- .../interfaces/task_lifecycle.go | 8 + .../allocrunnerv2/taskrunner/service_hook.go | 2 +- client/allocrunnerv2/taskrunner/stats_hook.go | 117 ++++++++++++ .../taskrunner/stats_hook_test.go | 180 ++++++++++++++++++ .../allocrunnerv2/taskrunner/task_runner.go | 98 ++++++++++ .../taskrunner/task_runner_hooks.go | 2 + client/client.go | 12 +- 8 files changed, 445 insertions(+), 13 deletions(-) create mode 100644 client/allocrunnerv2/taskrunner/stats_hook.go create mode 100644 client/allocrunnerv2/taskrunner/stats_hook_test.go diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index dc65e180b..73e5773cd 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -416,6 +416,7 @@ func (ar *allocRunner) Listener() *cstructs.AllocListener { // exit (thus closing WaitCh). func (ar *allocRunner) Destroy() { // Stop tasks + ar.tasksLock.RLock() for name, tr := range ar.tasks { err := tr.Kill(context.TODO(), structs.NewTaskEvent(structs.TaskKilled)) if err != nil { @@ -426,6 +427,7 @@ func (ar *allocRunner) Destroy() { } } } + ar.tasksLock.RUnlock() // Wait for tasks to exit and postrun hooks to finish <-ar.waitCh @@ -474,15 +476,38 @@ func (ar *allocRunner) IsMigrating() bool { return ar.prevAllocWatcher.IsMigrating() } -// StatsReporter needs implementing -//XXX func (ar *allocRunner) StatsReporter() allocrunner.AllocStatsReporter { - return noopStatsReporter{} + return ar } -//FIXME implement -type noopStatsReporter struct{} +// LatestAllocStats returns the latest stats for an allocation. If taskFilter +// is set, only stats for that task -- if it exists -- are returned. +func (ar *allocRunner) LatestAllocStats(taskFilter string) (*cstructs.AllocResourceUsage, error) { + ar.tasksLock.RLock() + defer ar.tasksLock.RUnlock() -func (noopStatsReporter) LatestAllocStats(taskFilter string) (*cstructs.AllocResourceUsage, error) { - return nil, fmt.Errorf("not implemented") + astat := &cstructs.AllocResourceUsage{ + Tasks: make(map[string]*cstructs.TaskResourceUsage, len(ar.tasks)), + ResourceUsage: &cstructs.ResourceUsage{ + MemoryStats: &cstructs.MemoryStats{}, + CpuStats: &cstructs.CpuStats{}, + }, + } + + for name, tr := range ar.tasks { + if taskFilter != "" && taskFilter != name { + // Getting stats for a particular task and its not this one! + continue + } + + if usage := tr.LatestResourceUsage(); usage != nil { + astat.Tasks[name] = usage + astat.ResourceUsage.Add(usage.ResourceUsage) + if usage.Timestamp > astat.Timestamp { + astat.Timestamp = usage.Timestamp + } + } + } + + return astat, nil } diff --git a/client/allocrunnerv2/interfaces/task_lifecycle.go b/client/allocrunnerv2/interfaces/task_lifecycle.go index 4de78c37a..8be2bfd6e 100644 --- a/client/allocrunnerv2/interfaces/task_lifecycle.go +++ b/client/allocrunnerv2/interfaces/task_lifecycle.go @@ -75,6 +75,11 @@ type TaskPrestartHook interface { Prestart(context.Context, *TaskPrestartRequest, *TaskPrestartResponse) error } +// DriverStats is the interface implemented by DriverHandles to return task stats. +type DriverStats interface { + Stats() (*cstructs.TaskResourceUsage, error) +} + type TaskPoststartRequest struct { // Exec hook (may be nil) DriverExec driver.ScriptExecutor @@ -84,6 +89,9 @@ type TaskPoststartRequest struct { // TaskEnv is the task's environment TaskEnv *env.TaskEnv + + // Stats collector + DriverStats DriverStats } type TaskPoststartResponse struct{} diff --git a/client/allocrunnerv2/taskrunner/service_hook.go b/client/allocrunnerv2/taskrunner/service_hook.go index c1097e36e..28ed0c8aa 100644 --- a/client/allocrunnerv2/taskrunner/service_hook.go +++ b/client/allocrunnerv2/taskrunner/service_hook.go @@ -122,7 +122,7 @@ func (h *serviceHook) Update(ctx context.Context, req *interfaces.TaskUpdateRequ return h.consul.UpdateTask(oldTaskServices, newTaskServices) } -func (h *serviceHook) Exited(ctx context.Context, req *interfaces.TaskExitedRequest, _ *interfaces.TaskExitedResponse) error { +func (h *serviceHook) Exited(context.Context, *interfaces.TaskExitedRequest, *interfaces.TaskExitedResponse) error { h.mu.Lock() defer h.mu.Unlock() diff --git a/client/allocrunnerv2/taskrunner/stats_hook.go b/client/allocrunnerv2/taskrunner/stats_hook.go new file mode 100644 index 000000000..a309e7c34 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/stats_hook.go @@ -0,0 +1,117 @@ +package taskrunner + +import ( + "context" + "strings" + "sync" + "time" + + hclog "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/driver" + cstructs "github.com/hashicorp/nomad/client/structs" +) + +type StatsUpdater interface { + UpdateStats(*cstructs.TaskResourceUsage) +} + +type statsHook struct { + updater StatsUpdater + interval time.Duration + + // stopCh is closed by Exited + stopCh chan struct{} + + mu sync.Mutex + + logger hclog.Logger +} + +func newStatsHook(su StatsUpdater, interval time.Duration, logger hclog.Logger) *statsHook { + h := &statsHook{ + updater: su, + interval: interval, + } + h.logger = logger.Named(h.Name()) + return h +} + +func (*statsHook) Name() string { + return "stats_hook" +} + +func (h *statsHook) Poststart(ctx context.Context, req *interfaces.TaskPoststartRequest, _ *interfaces.TaskPoststartResponse) error { + h.mu.Lock() + defer h.mu.Unlock() + + // This shouldn't happen, but better safe than risk leaking a goroutine + if h.stopCh != nil { + h.logger.Debug("poststart called twice without exiting between") + close(h.stopCh) + } + + h.stopCh = make(chan struct{}) + go h.collectResourceUsageStats(h.logger, req.DriverStats, h.stopCh) + + return nil +} + +func (h *statsHook) Exited(context.Context, *interfaces.TaskExitedRequest, *interfaces.TaskExitedResponse) error { + h.mu.Lock() + defer h.mu.Unlock() + + if h.stopCh == nil { + // No stats running + return nil + } + + // Close chan to stop stats collection + close(h.stopCh) + + // Clear chan so we don't double close for any reason + h.stopCh = nil + + return nil +} + +// collectResourceUsageStats starts collecting resource usage stats of a Task. +// Collection ends when the passed channel is closed +func (h *statsHook) collectResourceUsageStats(logger hclog.Logger, handle interfaces.DriverStats, stopCh <-chan struct{}) { + // start collecting the stats right away and then start collecting every + // collection interval + next := time.NewTimer(0) + defer next.Stop() + for { + select { + case <-next.C: + // Reset the timer + next.Reset(h.interval) + + // Collect stats from driver + ru, err := handle.Stats() + if err != nil { + // Check if the driver doesn't implement stats + if err.Error() == driver.DriverStatsNotImplemented.Error() { + h.logger.Debug("driver does not support stats") + return + } + + //XXX This is a net/rpc specific error + // We do not log when the plugin is shutdown as this is simply a + // race between the stopCollection channel being closed and calling + // Stats on the handle. + if !strings.Contains(err.Error(), "connection is shut down") { + h.logger.Debug("error fetching stats of task", "error", err) + } + + continue + } + + // Update stats on TaskRunner and emit them + h.updater.UpdateStats(ru) + case <-stopCh: + return + } + } +} diff --git a/client/allocrunnerv2/taskrunner/stats_hook_test.go b/client/allocrunnerv2/taskrunner/stats_hook_test.go new file mode 100644 index 000000000..98c9854c3 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/stats_hook_test.go @@ -0,0 +1,180 @@ +package taskrunner + +import ( + "context" + "testing" + "time" + + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/driver" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/helper/testlog" + "github.com/stretchr/testify/require" +) + +// Statically assert the stats hook implements the expected interfaces +var _ interfaces.TaskPoststartHook = (*statsHook)(nil) +var _ interfaces.TaskExitedHook = (*statsHook)(nil) + +type mockStatsUpdater struct { + // Ch is sent task resource usage updates if not nil + Ch chan *cstructs.TaskResourceUsage +} + +// newMockStatsUpdater returns a mockStatsUpdater that blocks on Ch for every +// call to UpdateStats +func newMockStatsUpdater() *mockStatsUpdater { + return &mockStatsUpdater{ + Ch: make(chan *cstructs.TaskResourceUsage), + } +} + +func (m *mockStatsUpdater) UpdateStats(ru *cstructs.TaskResourceUsage) { + if m.Ch != nil { + m.Ch <- ru + } +} + +type mockDriverStats struct { + // err is returned by Stats if it is non-nil + err error +} + +func (m *mockDriverStats) Stats() (*cstructs.TaskResourceUsage, error) { + if m.err != nil { + return nil, m.err + } + ru := &cstructs.TaskResourceUsage{ + ResourceUsage: &cstructs.ResourceUsage{ + MemoryStats: &cstructs.MemoryStats{ + RSS: 1, + Measured: []string{"RSS"}, + }, + CpuStats: &cstructs.CpuStats{ + SystemMode: 1, + Measured: []string{"System Mode"}, + }, + }, + Timestamp: time.Now().UnixNano(), + Pids: map[string]*cstructs.ResourceUsage{}, + } + ru.Pids["task"] = ru.ResourceUsage + return ru, nil +} + +// TestTaskRunner_StatsHook_PoststartExited asserts the stats hook starts and +// stops. +func TestTaskRunner_StatsHook_PoststartExited(t *testing.T) { + t.Parallel() + + require := require.New(t) + logger := testlog.HCLogger(t) + su := newMockStatsUpdater() + ds := new(mockDriverStats) + + poststartReq := &interfaces.TaskPoststartRequest{DriverStats: ds} + + // Create hook + h := newStatsHook(su, time.Minute, logger) + + // Always call Exited to cleanup goroutines + defer h.Exited(context.Background(), nil, nil) + + // Run prestart + require.NoError(h.Poststart(context.Background(), poststartReq, nil)) + + // An initial stats collection should run and call the updater + select { + case ru := <-su.Ch: + require.Equal(uint64(1), ru.ResourceUsage.MemoryStats.RSS) + case <-time.After(10 * time.Second): + t.Fatalf("timeout waiting for initial stats collection") + } + + require.NoError(h.Exited(context.Background(), nil, nil)) +} + +// TestTaskRunner_StatsHook_Periodic asserts the stats hook collects stats on +// an interval. +func TestTaskRunner_StatsHook_Periodic(t *testing.T) { + t.Parallel() + + require := require.New(t) + logger := testlog.HCLogger(t) + su := newMockStatsUpdater() + + ds := new(mockDriverStats) + poststartReq := &interfaces.TaskPoststartRequest{DriverStats: ds} + + // interval needs to be high enough that even on a slow/busy VM + // Exited() can complete within the interval. + const interval = 500 * time.Millisecond + + h := newStatsHook(su, interval, logger) + defer h.Exited(context.Background(), nil, nil) + + // Run prestart + require.NoError(h.Poststart(context.Background(), poststartReq, nil)) + + // An initial stats collection should run and call the updater + var firstrun int64 + select { + case ru := <-su.Ch: + if ru.Timestamp <= 0 { + t.Fatalf("expected nonzero timestamp (%v)", ru.Timestamp) + } + firstrun = ru.Timestamp + case <-time.After(10 * time.Second): + t.Fatalf("timeout waiting for initial stats collection") + } + + // Should get another update in ~500ms (see interval above) + select { + case ru := <-su.Ch: + if ru.Timestamp <= firstrun { + t.Fatalf("expected timestamp (%v) after first run (%v)", ru.Timestamp, firstrun) + } + case <-time.After(10 * time.Second): + t.Fatalf("timeout waiting for second stats collection") + } + + // Exiting should prevent further updates + require.NoError(h.Exited(context.Background(), nil, nil)) + + // Should *not* get another update in ~500ms (see interval above) + select { + case ru := <-su.Ch: + t.Fatalf("unexpected update after exit (firstrun=%v; update=%v", firstrun, ru.Timestamp) + case <-time.After(2 * interval): + // Ok! No update after exit as expected. + } +} + +// TestTaskRunner_StatsHook_NotImplemented asserts the stats hook stops if the +// driver returns NotImplemented. +func TestTaskRunner_StatsHook_NotImplemented(t *testing.T) { + t.Parallel() + + require := require.New(t) + logger := testlog.HCLogger(t) + su := newMockStatsUpdater() + ds := &mockDriverStats{ + err: driver.DriverStatsNotImplemented, + } + + poststartReq := &interfaces.TaskPoststartRequest{DriverStats: ds} + + h := newStatsHook(su, 1, logger) + defer h.Exited(context.Background(), nil, nil) + + // Run prestart + require.NoError(h.Poststart(context.Background(), poststartReq, nil)) + + // An initial stats collection should run and *not* call the updater + select { + case ru := <-su.Ch: + t.Fatalf("unexpected resource update (timestamp=%v)", ru.Timestamp) + case <-time.After(500 * time.Millisecond): + // Ok! No update received because error was returned + } +} diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 3118626ef..11da80637 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -17,6 +17,7 @@ import ( "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" cstate "github.com/hashicorp/nomad/client/state" + cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" ) @@ -129,6 +130,11 @@ type TaskRunner struct { // baseLabels are used when emitting tagged metrics. All task runner metrics // will have these tags, and optionally more. baseLabels []metrics.Label + + // resourceUsage is written via UpdateStats and read via + // LatestResourceUsage. May be nil at all times. + resourceUsage *cstructs.TaskResourceUsage + resourceUsageLock sync.Mutex } type Config struct { @@ -659,6 +665,98 @@ func (tr *TaskRunner) triggerUpdateHooks() { } } +// LatestResourceUsage returns the last resource utilization datapoint +// collected. May return nil if the task is not running or no resource +// utilization has been collected yet. +func (tr *TaskRunner) LatestResourceUsage() *cstructs.TaskResourceUsage { + tr.resourceUsageLock.Lock() + ru := tr.resourceUsage + tr.resourceUsageLock.Unlock() + return ru +} + +// UpdateStats updates and emits the latest stats from the driver. +func (tr *TaskRunner) UpdateStats(ru *cstructs.TaskResourceUsage) { + tr.resourceUsageLock.Lock() + tr.resourceUsage = ru + tr.resourceUsageLock.Unlock() + if ru != nil { + tr.emitStats(ru) + } +} + +//TODO Remove Backwardscompat or use tr.Alloc()? +func (tr *TaskRunner) setGaugeForMemory(ru *cstructs.TaskResourceUsage) { + if !tr.clientConfig.DisableTaggedMetrics { + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "rss"}, + float32(ru.ResourceUsage.MemoryStats.RSS), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "rss"}, + float32(ru.ResourceUsage.MemoryStats.RSS), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "cache"}, + float32(ru.ResourceUsage.MemoryStats.Cache), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "swap"}, + float32(ru.ResourceUsage.MemoryStats.Swap), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "max_usage"}, + float32(ru.ResourceUsage.MemoryStats.MaxUsage), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "kernel_usage"}, + float32(ru.ResourceUsage.MemoryStats.KernelUsage), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "kernel_max_usage"}, + float32(ru.ResourceUsage.MemoryStats.KernelMaxUsage), tr.baseLabels) + } + + if tr.clientConfig.BackwardsCompatibleMetrics { + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "rss"}, float32(ru.ResourceUsage.MemoryStats.RSS)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "cache"}, float32(ru.ResourceUsage.MemoryStats.Cache)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "swap"}, float32(ru.ResourceUsage.MemoryStats.Swap)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "max_usage"}, float32(ru.ResourceUsage.MemoryStats.MaxUsage)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "kernel_usage"}, float32(ru.ResourceUsage.MemoryStats.KernelUsage)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "kernel_max_usage"}, float32(ru.ResourceUsage.MemoryStats.KernelMaxUsage)) + } +} + +//TODO Remove Backwardscompat or use tr.Alloc()? +func (tr *TaskRunner) setGaugeForCPU(ru *cstructs.TaskResourceUsage) { + if !tr.clientConfig.DisableTaggedMetrics { + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "total_percent"}, + float32(ru.ResourceUsage.CpuStats.Percent), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "system"}, + float32(ru.ResourceUsage.CpuStats.SystemMode), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "user"}, + float32(ru.ResourceUsage.CpuStats.UserMode), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "throttled_time"}, + float32(ru.ResourceUsage.CpuStats.ThrottledTime), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "throttled_periods"}, + float32(ru.ResourceUsage.CpuStats.ThrottledPeriods), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "total_ticks"}, + float32(ru.ResourceUsage.CpuStats.TotalTicks), tr.baseLabels) + } + + if tr.clientConfig.BackwardsCompatibleMetrics { + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "total_percent"}, float32(ru.ResourceUsage.CpuStats.Percent)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "system"}, float32(ru.ResourceUsage.CpuStats.SystemMode)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "user"}, float32(ru.ResourceUsage.CpuStats.UserMode)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "throttled_time"}, float32(ru.ResourceUsage.CpuStats.ThrottledTime)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "throttled_periods"}, float32(ru.ResourceUsage.CpuStats.ThrottledPeriods)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "total_ticks"}, float32(ru.ResourceUsage.CpuStats.TotalTicks)) + } +} + +// emitStats emits resource usage stats of tasks to remote metrics collector +// sinks +func (tr *TaskRunner) emitStats(ru *cstructs.TaskResourceUsage) { + if !tr.clientConfig.PublishAllocationMetrics { + return + } + + if ru.ResourceUsage.MemoryStats != nil { + tr.setGaugeForMemory(ru) + } + + if ru.ResourceUsage.CpuStats != nil { + tr.setGaugeForCPU(ru) + } +} + // appendTaskEvent updates the task status by appending the new event. func appendTaskEvent(state *structs.TaskState, event *structs.TaskEvent) { const capacity = 10 diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index bad2c84a3..b377fd7d3 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -23,6 +23,7 @@ func (tr *TaskRunner) initHooks() { newTaskDirHook(tr, hookLogger), newArtifactHook(tr, hookLogger), newShutdownDelayHook(task.ShutdownDelay, hookLogger), + newStatsHook(tr, tr.clientConfig.StatsCollectionInterval, hookLogger), } // If Vault is enabled, add the hook @@ -186,6 +187,7 @@ func (tr *TaskRunner) poststart() error { req := interfaces.TaskPoststartRequest{ DriverExec: handle, DriverNetwork: net, + DriverStats: handle, TaskEnv: tr.envBuilder.Build(), } var resp interfaces.TaskPoststartResponse diff --git a/client/client.go b/client/client.go index 09f6d063e..83508769f 100644 --- a/client/client.go +++ b/client/client.go @@ -99,19 +99,21 @@ type ClientStatsReporter interface { LatestHostStats() *stats.HostStats } +// AllocRunner is the interface implemented by the core alloc runner. +//TODO Create via factory to allow testing Client with mock AllocRunners. type AllocRunner interface { - StatsReporter() allocrunner.AllocStatsReporter + Alloc() *structs.Allocation Destroy() GetAllocDir() *allocdir.AllocDir IsDestroyed() bool - IsWaiting() bool IsMigrating() bool + IsWaiting() bool Listener() *cstructs.AllocListener - WaitCh() <-chan struct{} - Update(*structs.Allocation) - Alloc() *structs.Allocation Restore() error Run() + StatsReporter() allocrunner.AllocStatsReporter + Update(*structs.Allocation) + WaitCh() <-chan struct{} } // Client is used to implement the client interaction with Nomad. Clients From 5dee1141d1920d68c352ff3cddfb62d7162439e8 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Mon, 24 Sep 2018 14:37:45 -0400 Subject: [PATCH 071/152] executor v2 (#4656) * client/executor: refactor client to remove interpolation * executor: POC libcontainer based executor * vendor: use hashicorp libcontainer fork * vendor: add libcontainer/nsenter dep * executor: updated executor interface to simplify operations * executor: implement logging pipe * logmon: new logmon plugin to manage task logs * driver/executor: use logmon for log management * executor: fix tests and windows build * executor: fix logging key names * executor: fix test failures * executor: add config field to toggle between using libcontainer and standard executors * logmon: use discover utility to discover nomad executable * executor: only call libcontainer-shim on main in linux * logmon: use seperate path configs for stdout/stderr fifos * executor: windows fixes * executor: created reusable pid stats collection utility that can be used in an executor * executor: update fifo.Open calls * executor: fix build * remove executor from docker driver * executor: Shutdown func to kill and cleanup executor and its children * executor: move linux specific universal executor funcs to seperate file * move logmon initialization to a task runner hook * client: doc fixes and renaming from code review * taskrunner: use shared config struct for logmon fifo fields * taskrunner: logmon only needs to be started once per task --- .../allocrunnerv2/taskrunner/logmon_hook.go | 109 ++ .../taskrunner/logmon_hook_test.go | 7 + .../allocrunnerv2/taskrunner/task_runner.go | 7 + .../taskrunner/task_runner_hooks.go | 3 + client/driver/docker.go | 85 +- client/driver/docker_linux_test.go | 2 +- client/driver/docker_test.go | 44 +- client/driver/docker_unix_test.go | 2 +- client/driver/driver.go | 6 + client/driver/driver_test.go | 66 +- client/driver/exec.go | 158 ++- client/driver/exec_test.go | 17 +- client/driver/exec_unix_test.go | 4 +- client/driver/executor/executor.go | 814 ++++----------- client/driver/executor/executor_basic.go | 47 +- client/driver/executor/executor_linux.go | 799 ++++++++------ client/driver/executor/executor_linux_test.go | 173 ++-- client/driver/executor/executor_test.go | 352 +++---- .../executor/executor_universal_linux.go | 156 +++ client/driver/executor/executor_unix.go | 23 +- client/driver/executor/executor_windows.go | 6 +- client/driver/executor/pid_collector.go | 219 ++++ client/driver/executor/pid_collector_test.go | 45 + .../executor/resource_container_default.go | 12 - .../executor/resource_container_linux.go | 23 +- client/driver/executor_plugin.go | 109 +- client/driver/executor_plugin_pre_0_9_0.go | 220 ++++ client/driver/java.go | 150 ++- client/driver/java_test.go | 18 +- client/driver/logging/rotator.go | 22 +- client/driver/logging/syslog_parser.go | 6 +- client/driver/logging/syslog_server.go | 10 +- client/driver/logging/universal_collector.go | 15 +- client/driver/lxc_test.go | 10 +- client/driver/mock_driver.go | 3 +- client/driver/plugins.go | 18 +- client/driver/qemu.go | 42 +- client/driver/qemu_test.go | 12 +- client/driver/raw_exec.go | 146 ++- client/driver/raw_exec_test.go | 14 +- client/driver/raw_exec_unix_test.go | 27 +- client/driver/rkt.go | 50 +- client/driver/rkt_test.go | 22 +- client/driver/structs/structs.go | 4 + client/driver/structs/structs_linux.go | 10 - client/driver/utils.go | 10 +- client/logmon/client.go | 31 + client/logmon/logmon.go | 209 ++++ client/logmon/plugin.go | 69 ++ client/logmon/proto/logmon.pb.go | 338 ++++++ client/logmon/proto/logmon.proto | 25 + client/logmon/server.go | 36 + command/commands.go | 5 + command/executor_plugin.go | 7 +- command/logmon_plugin.go | 36 + main.go | 1 + nsenter_linux.go | 25 + .../runc/libcontainer/container_linux.go | 3 +- .../runc/libcontainer/factory_linux.go | 28 +- .../runc/libcontainer/network_linux.go | 157 --- .../runc/libcontainer/nsenter/README.md | 41 + .../runc/libcontainer/nsenter/namespace.h | 32 + .../runc/libcontainer/nsenter/nsenter.go | 12 + .../libcontainer/nsenter/nsenter_gccgo.go | 25 + .../nsenter/nsenter_unsupported.go | 5 + .../runc/libcontainer/nsenter/nsexec.c | 980 ++++++++++++++++++ .../runc/libcontainer/rootfs_linux.go | 39 +- vendor/vendor.json | 33 +- 68 files changed, 4186 insertions(+), 2048 deletions(-) create mode 100644 client/allocrunnerv2/taskrunner/logmon_hook.go create mode 100644 client/allocrunnerv2/taskrunner/logmon_hook_test.go create mode 100644 client/driver/executor/executor_universal_linux.go create mode 100644 client/driver/executor/pid_collector.go create mode 100644 client/driver/executor/pid_collector_test.go create mode 100644 client/driver/executor_plugin_pre_0_9_0.go delete mode 100644 client/driver/structs/structs_linux.go create mode 100644 client/logmon/client.go create mode 100644 client/logmon/logmon.go create mode 100644 client/logmon/plugin.go create mode 100644 client/logmon/proto/logmon.pb.go create mode 100644 client/logmon/proto/logmon.proto create mode 100644 client/logmon/server.go create mode 100644 command/logmon_plugin.go create mode 100644 nsenter_linux.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/nsenter/README.md create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/nsenter/namespace.h create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter_gccgo.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter_unsupported.go create mode 100644 vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsexec.c diff --git a/client/allocrunnerv2/taskrunner/logmon_hook.go b/client/allocrunnerv2/taskrunner/logmon_hook.go new file mode 100644 index 000000000..2c47142ca --- /dev/null +++ b/client/allocrunnerv2/taskrunner/logmon_hook.go @@ -0,0 +1,109 @@ +package taskrunner + +import ( + "context" + "fmt" + "path/filepath" + "runtime" + + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/logmon" + "github.com/hashicorp/nomad/helper/uuid" +) + +// logmonHook launches logmon and manages task logging +type logmonHook struct { + // logmon is the handle to the log monitor process for the task. + logmon logmon.LogMon + logmonPluginClient *plugin.Client + + config *logmonHookConfig + + logger hclog.Logger +} + +type logmonHookConfig struct { + logDir string + stdoutFifo string + stderrFifo string +} + +func newLogMonHook(cfg *logmonHookConfig, logger hclog.Logger) *logmonHook { + hook := &logmonHook{ + config: cfg, + logger: logger, + } + + return hook +} + +func newLogMonHookConfig(taskName, logDir string) *logmonHookConfig { + cfg := &logmonHookConfig{ + logDir: logDir, + } + if runtime.GOOS == "windows" { + id := uuid.Generate()[:8] + cfg.stdoutFifo = fmt.Sprintf("//./pipe/%s-%s.stdout", taskName, id) + cfg.stderrFifo = fmt.Sprintf("//./pipe/%s-%s.stderr", taskName, id) + } else { + cfg.stdoutFifo = filepath.Join(logDir, fmt.Sprintf(".%s.stdout.fifo", taskName)) + cfg.stderrFifo = filepath.Join(logDir, fmt.Sprintf(".%s.stderr.fifo", taskName)) + } + return cfg +} + +func (*logmonHook) Name() string { + return "logmon" +} + +func (h *logmonHook) launchLogMon() error { + l, c, err := logmon.LaunchLogMon(h.logger) + if err != nil { + return err + } + + h.logmon = l + h.logmonPluginClient = c + return nil +} + +func (h *logmonHook) Prestart(ctx context.Context, + req *interfaces.TaskPrestartRequest, resp *interfaces.TaskPrestartResponse) error { + + // Launch logmon instance for the task. + if err := h.launchLogMon(); err != nil { + h.logger.Error("failed to launch logmon process", "error", err) + return err + } + + err := h.logmon.Start(&logmon.LogConfig{ + LogDir: h.config.logDir, + StdoutLogFile: fmt.Sprintf("%s.stdout", req.Task.Name), + StderrLogFile: fmt.Sprintf("%s.stderr", req.Task.Name), + StdoutFifo: h.config.stdoutFifo, + StderrFifo: h.config.stderrFifo, + MaxFiles: req.Task.LogConfig.MaxFiles, + MaxFileSizeMB: req.Task.LogConfig.MaxFileSizeMB, + }) + if err != nil { + h.logger.Error("failed to start logmon", "error", err) + return err + } + + resp.Done = true + return nil +} + +func (h *logmonHook) Stop(context.Context, *interfaces.TaskStopRequest, *interfaces.TaskStopResponse) error { + + if h.logmon != nil { + h.logmon.Stop() + } + if h.logmonPluginClient != nil { + h.logmonPluginClient.Kill() + } + + return nil +} diff --git a/client/allocrunnerv2/taskrunner/logmon_hook_test.go b/client/allocrunnerv2/taskrunner/logmon_hook_test.go new file mode 100644 index 000000000..b7df170e9 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/logmon_hook_test.go @@ -0,0 +1,7 @@ +package taskrunner + +import "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + +// Statically assert the logmon hook implements the expected interfaces +var _ interfaces.TaskPrestartHook = (*logmonHook)(nil) +var _ interfaces.TaskStopHook = (*logmonHook)(nil) diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 11da80637..4b5bed7be 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -131,6 +131,10 @@ type TaskRunner struct { // will have these tags, and optionally more. baseLabels []metrics.Label + // logmonHookConfig is used to get the paths to the stdout and stderr fifos + // to be passed to the driver for task logging + logmonHookConfig *logmonHookConfig + // resourceUsage is written via UpdateStats and read via // LatestResourceUsage. May be nil at all times. resourceUsage *cstructs.TaskResourceUsage @@ -394,6 +398,9 @@ func (tr *TaskRunner) runDriver() error { // Create a new context for Start since the environment may have been updated. ctx = driver.NewExecContext(tr.taskDir, tr.envBuilder.Build()) + ctx.StdoutFifo = tr.logmonHookConfig.stdoutFifo + ctx.StderrFifo = tr.logmonHookConfig.stderrFifo + // Start the job sresp, err := tr.driver.Start(ctx, tr.task) if err != nil { diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index b377fd7d3..8baf4a6c5 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -16,11 +16,14 @@ func (tr *TaskRunner) initHooks() { hookLogger := tr.logger.Named("task_hook") task := tr.Task() + tr.logmonHookConfig = newLogMonHookConfig(task.Name, tr.taskDir.LogDir) + // Create the task directory hook. This is run first to ensure the // directoy path exists for other hooks. tr.runnerHooks = []interfaces.TaskHook{ newValidateHook(tr.clientConfig, hookLogger), newTaskDirHook(tr, hookLogger), + newLogMonHook(tr.logmonHookConfig, hookLogger), newArtifactHook(tr, hookLogger), newShutdownDelayHook(task.ShutdownDelay, hookLogger), newStatsHook(tr, tr.clientConfig.StatsCollectionInterval, hookLogger), diff --git a/client/driver/docker.go b/client/driver/docker.go index 4680dbdc7..1ec795e2b 100644 --- a/client/driver/docker.go +++ b/client/driver/docker.go @@ -25,10 +25,8 @@ import ( "github.com/docker/docker/registry" "github.com/hashicorp/go-multierror" - plugin "github.com/hashicorp/go-plugin" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/driver/env" - "github.com/hashicorp/nomad/client/driver/executor" dstructs "github.com/hashicorp/nomad/client/driver/structs" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/helper" @@ -477,12 +475,9 @@ type dockerPID struct { ContainerID string KillTimeout time.Duration MaxKillTimeout time.Duration - PluginConfig *PluginReattachConfig } type DockerHandle struct { - pluginClient *plugin.Client - executor executor.Executor client *docker.Client waitClient *docker.Client logger *log.Logger @@ -824,50 +819,10 @@ func (d *DockerDriver) Prestart(ctx *ExecContext, task *structs.Task) (*Prestart } func (d *DockerDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse, error) { - pluginLogFile := filepath.Join(ctx.TaskDir.Dir, "executor.out") - executorConfig := &dstructs.ExecutorConfig{ - LogFile: pluginLogFile, - LogLevel: d.config.LogLevel, - } - - exec, pluginClient, err := createExecutor(d.config.LogOutput, d.config, executorConfig) - if err != nil { - return nil, err - } - executorCtx := &executor.ExecutorContext{ - TaskEnv: ctx.TaskEnv, - Task: task, - Driver: "docker", - LogDir: ctx.TaskDir.LogDir, - TaskDir: ctx.TaskDir.Dir, - PortLowerBound: d.config.ClientMinPort, - PortUpperBound: d.config.ClientMaxPort, - } - if err := exec.SetContext(executorCtx); err != nil { - pluginClient.Kill() - return nil, fmt.Errorf("failed to set executor context: %v", err) - } - - // The user hasn't specified any logging options so launch our own syslog - // server if possible. - syslogAddr := "" - if len(d.driverConfig.Logging) == 0 { - if runtime.GOOS == "darwin" { - d.logger.Printf("[DEBUG] driver.docker: disabling syslog driver as Docker for Mac workaround") - } else { - ss, err := exec.LaunchSyslogServer() - if err != nil { - pluginClient.Kill() - return nil, fmt.Errorf("failed to start syslog collector: %v", err) - } - syslogAddr = ss.Addr - } - } - - config, err := d.createContainerConfig(ctx, task, d.driverConfig, syslogAddr) + // TODO: implement alternative to launching a syslog server in the executor + config, err := d.createContainerConfig(ctx, task, d.driverConfig, "") if err != nil { d.logger.Printf("[ERR] driver.docker: failed to create container configuration for image %q (%q): %v", d.driverConfig.ImageName, d.imageID, err) - pluginClient.Kill() return nil, fmt.Errorf("Failed to create container configuration for image %q (%q): %v", d.driverConfig.ImageName, d.imageID, err) } @@ -875,7 +830,6 @@ func (d *DockerDriver) Start(ctx *ExecContext, task *structs.Task) (*StartRespon if err != nil { wrapped := fmt.Sprintf("Failed to create container: %v", err) d.logger.Printf("[ERR] driver.docker: %s", wrapped) - pluginClient.Kill() return nil, structs.WrapRecoverable(wrapped, err) } @@ -888,7 +842,6 @@ func (d *DockerDriver) Start(ctx *ExecContext, task *structs.Task) (*StartRespon // Start the container if err := d.startContainer(container); err != nil { d.logger.Printf("[ERR] driver.docker: failed to start container %s: %s", container.ID, err) - pluginClient.Kill() return nil, structs.NewRecoverableError(fmt.Errorf("Failed to start container %s: %s", container.ID, err), structs.IsRecoverable(err)) } @@ -899,7 +852,6 @@ func (d *DockerDriver) Start(ctx *ExecContext, task *structs.Task) (*StartRespon if err != nil { err = fmt.Errorf("failed to inspect started container %s: %s", container.ID, err) d.logger.Printf("[ERR] driver.docker: %v", err) - pluginClient.Kill() return nil, structs.NewRecoverableError(err, true) } container = runningContainer @@ -923,8 +875,6 @@ func (d *DockerDriver) Start(ctx *ExecContext, task *structs.Task) (*StartRespon h := &DockerHandle{ client: client, waitClient: waitClient, - executor: exec, - pluginClient: pluginClient, logger: d.logger, jobName: d.DriverContext.jobName, taskGroupName: d.DriverContext.taskGroupName, @@ -938,7 +888,7 @@ func (d *DockerDriver) Start(ctx *ExecContext, task *structs.Task) (*StartRespon doneCh: make(chan bool), waitCh: make(chan *dstructs.WaitResult, 1), removeContainerOnExit: d.config.ReadBoolDefault(dockerCleanupContainerConfigOption, dockerCleanupContainerConfigDefault), - net: net, + net: net, } go h.collectStats() go h.run() @@ -1752,9 +1702,6 @@ func (d *DockerDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, er } d.logger.Printf("[INFO] driver.docker: re-attaching to docker process: %s", pid.ContainerID) d.logger.Printf("[DEBUG] driver.docker: re-attached to handle: %s", handleID) - pluginConfig := &plugin.ClientConfig{ - Reattach: pid.PluginConfig.PluginConfig(), - } client, waitClient, err := d.dockerClients() if err != nil { @@ -1780,18 +1727,6 @@ func (d *DockerDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, er if !found { return nil, fmt.Errorf("Failed to find container %s", pid.ContainerID) } - exec, pluginClient, err := createExecutorWithConfig(pluginConfig, d.config.LogOutput) - if err != nil { - d.logger.Printf("[INFO] driver.docker: couldn't re-attach to the plugin process: %v", err) - d.logger.Printf("[DEBUG] driver.docker: stopping container %q", pid.ContainerID) - if e := client.StopContainer(pid.ContainerID, uint(pid.KillTimeout.Seconds())); e != nil { - d.logger.Printf("[DEBUG] driver.docker: couldn't stop container: %v", e) - } - return nil, err - } - - ver, _ := exec.Version() - d.logger.Printf("[DEBUG] driver.docker: version of executor: %v", ver.Version) // Increment the reference count since we successfully attached to this // container @@ -1802,8 +1737,6 @@ func (d *DockerDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, er h := &DockerHandle{ client: client, waitClient: waitClient, - executor: exec, - pluginClient: pluginClient, logger: d.logger, jobName: d.DriverContext.jobName, taskGroupName: d.DriverContext.taskGroupName, @@ -1832,7 +1765,6 @@ func (h *DockerHandle) ID() string { ImageID: h.ImageID, KillTimeout: h.killTimeout, MaxKillTimeout: h.maxKillTimeout, - PluginConfig: NewPluginReattachConfig(h.pluginClient.ReattachConfig()), } data, err := json.Marshal(pid) if err != nil { @@ -1856,9 +1788,6 @@ func (h *DockerHandle) Network() *cstructs.DriverNetwork { func (h *DockerHandle) Update(task *structs.Task) error { // Store the updated kill timeout. h.killTimeout = GetKillTimeout(task.KillTimeout, h.maxKillTimeout) - if err := h.executor.UpdateTask(task); err != nil { - h.logger.Printf("[DEBUG] driver.docker: failed to update log config: %v", err) - } // Update is not possible return nil @@ -1923,10 +1852,8 @@ func (h *DockerHandle) Signal(s os.Signal) error { // Kill is used to terminate the task. This uses `docker stop -t killTimeout` func (h *DockerHandle) Kill() error { // Stop the container - err := h.waitClient.StopContainer(h.containerID, uint(h.killTimeout.Seconds())) + err := h.client.StopContainer(h.containerID, uint(h.killTimeout.Seconds())) if err != nil { - h.executor.Exit() - h.pluginClient.Kill() // Container has already been removed. if strings.Contains(err.Error(), NoSuchContainerError) { @@ -1986,10 +1913,6 @@ func (h *DockerHandle) run() { close(h.doneCh) // Shutdown the syslog collector - if err := h.executor.Exit(); err != nil { - h.logger.Printf("[ERR] driver.docker: failed to kill the syslog collector: %v", err) - } - h.pluginClient.Kill() // Stop the container just incase the docker daemon's wait returned // incorrectly diff --git a/client/driver/docker_linux_test.go b/client/driver/docker_linux_test.go index 562a7de38..d761bfe73 100644 --- a/client/driver/docker_linux_test.go +++ b/client/driver/docker_linux_test.go @@ -58,7 +58,7 @@ func TestDockerDriver_PidsLimit(t *testing.T) { task.Config["args"] = []string{"-c", "sleep 2 & sleep 2"} ctx := testDockerDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewDockerDriver(ctx.DriverCtx) // Copy the image into the task's directory diff --git a/client/driver/docker_test.go b/client/driver/docker_test.go index 2b1d7c46f..7b9b3b097 100644 --- a/client/driver/docker_test.go +++ b/client/driver/docker_test.go @@ -117,7 +117,7 @@ func dockerSetupWithClient(t *testing.T, task *structs.Task, client *docker.Clie if presp != nil && presp.CreatedResources != nil { driver.Cleanup(tctx.ExecCtx, presp.CreatedResources) } - tctx.AllocDir.Destroy() + tctx.Destroy() t.Fatalf("error in prestart: %v", err) } // Update the exec ctx with the driver network env vars @@ -126,13 +126,13 @@ func dockerSetupWithClient(t *testing.T, task *structs.Task, client *docker.Clie sresp, err := driver.Start(tctx.ExecCtx, task) if err != nil { driver.Cleanup(tctx.ExecCtx, presp.CreatedResources) - tctx.AllocDir.Destroy() + tctx.Destroy() t.Fatalf("Failed to start driver: %s\nStack\n%s", err, debug.Stack()) } if sresp.Handle == nil { driver.Cleanup(tctx.ExecCtx, presp.CreatedResources) - tctx.AllocDir.Destroy() + tctx.Destroy() t.Fatalf("handle is nil\nStack\n%s", debug.Stack()) } @@ -142,7 +142,7 @@ func dockerSetupWithClient(t *testing.T, task *structs.Task, client *docker.Clie cleanup := func() { driver.Cleanup(tctx.ExecCtx, presp.CreatedResources) sresp.Handle.Kill() - tctx.AllocDir.Destroy() + tctx.Destroy() } return client, sresp.Handle.(*DockerHandle), cleanup @@ -169,7 +169,7 @@ func TestDockerDriver_Fingerprint(t *testing.T) { ctx := testDockerDriverContexts(t, &structs.Task{Name: "foo", Driver: "docker", Resources: basicResources}) //ctx.DriverCtx.config.Options = map[string]string{"docker.cleanup.image": "false"} - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewDockerDriver(ctx.DriverCtx) node := &structs.Node{ Attributes: make(map[string]string), @@ -319,7 +319,7 @@ func TestDockerDriver_StartOpen_Wait(t *testing.T) { ctx := testDockerDriverContexts(t, task) //ctx.DriverCtx.config.Options = map[string]string{"docker.cleanup.image": "false"} - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewDockerDriver(ctx.DriverCtx) copyImage(t, ctx.ExecCtx.TaskDir, "busybox.tar") @@ -423,7 +423,7 @@ func TestDockerDriver_Start_StoppedContainer(t *testing.T) { } tctx := testDockerDriverContexts(t, task) - defer tctx.AllocDir.Destroy() + defer tctx.Destroy() copyImage(t, tctx.ExecCtx.TaskDir, "busybox.tar") client := newTestDockerClient(t) @@ -495,7 +495,7 @@ func TestDockerDriver_Start_LoadImage(t *testing.T) { ctx := testDockerDriverContexts(t, task) //ctx.DriverCtx.config.Options = map[string]string{"docker.cleanup.image": "false"} - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewDockerDriver(ctx.DriverCtx) // Copy the image into the task's directory @@ -563,7 +563,7 @@ func TestDockerDriver_Start_BadPull_Recoverable(t *testing.T) { ctx := testDockerDriverContexts(t, task) //ctx.DriverCtx.config.Options = map[string]string{"docker.cleanup.image": "false"} - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewDockerDriver(ctx.DriverCtx) _, err := d.Prestart(ctx.ExecCtx, task) @@ -616,7 +616,7 @@ func TestDockerDriver_Start_Wait_AllocDir(t *testing.T) { ctx := testDockerDriverContexts(t, task) //ctx.DriverCtx.config.Options = map[string]string{"docker.cleanup.image": "false"} - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewDockerDriver(ctx.DriverCtx) copyImage(t, ctx.ExecCtx.TaskDir, "busybox.tar") @@ -769,7 +769,7 @@ func TestDockerDriver_StartN(t *testing.T) { for idx, task := range taskList { ctx := testDockerDriverContexts(t, task) //ctx.DriverCtx.config.Options = map[string]string{"docker.cleanup.image": "false"} - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewDockerDriver(ctx.DriverCtx) copyImage(t, ctx.ExecCtx.TaskDir, "busybox.tar") @@ -834,7 +834,7 @@ func TestDockerDriver_StartNVersions(t *testing.T) { for idx, task := range taskList { ctx := testDockerDriverContexts(t, task) //ctx.DriverCtx.config.Options = map[string]string{"docker.cleanup.image": "false"} - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewDockerDriver(ctx.DriverCtx) copyImage(t, ctx.ExecCtx.TaskDir, "busybox.tar") copyImage(t, ctx.ExecCtx.TaskDir, "busybox_musl.tar") @@ -1063,7 +1063,7 @@ func TestDockerDriver_Sysctl_Ulimit_Errors(t *testing.T) { ctx := testDockerDriverContexts(t, task) driver := NewDockerDriver(ctx.DriverCtx) copyImage(t, ctx.ExecCtx.TaskDir, "busybox.tar") - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() _, err := driver.Prestart(ctx.ExecCtx, task) assert.NotNil(t, err, "Expected non nil error") @@ -1118,7 +1118,7 @@ func TestDockerDriver_ForcePull_IsInvalidConfig(t *testing.T) { task.Config["force_pull"] = "nothing" ctx := testDockerDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() //ctx.DriverCtx.config.Options = map[string]string{"docker.cleanup.image": "false"} driver := NewDockerDriver(ctx.DriverCtx) @@ -1279,7 +1279,7 @@ func TestDockerDriver_Capabilities(t *testing.T) { driver := NewDockerDriver(tctx.DriverCtx) copyImage(t, tctx.ExecCtx.TaskDir, "busybox.tar") - defer tctx.AllocDir.Destroy() + defer tctx.Destroy() presp, err := driver.Prestart(tctx.ExecCtx, task) defer driver.Cleanup(tctx.ExecCtx, presp.CreatedResources) @@ -1571,7 +1571,7 @@ func TestDockerDriver_User(t *testing.T) { ctx := testDockerDriverContexts(t, task) //ctx.DriverCtx.config.Options = map[string]string{"docker.cleanup.image": "false"} driver := NewDockerDriver(ctx.DriverCtx) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() copyImage(t, ctx.ExecCtx.TaskDir, "busybox.tar") _, err := driver.Prestart(ctx.ExecCtx, task) @@ -1731,7 +1731,7 @@ func setupDockerVolumes(t *testing.T, cfg *config.Config, hostpath string) (*str } // Build alloc and task directory structure - allocDir := allocdir.NewAllocDir(testlog.Logger(t), filepath.Join(cfg.AllocDir, uuid.Generate())) + allocDir := allocdir.NewAllocDir(testlog.HCLogger(t), filepath.Join(cfg.AllocDir, uuid.Generate())) if err := allocDir.Build(); err != nil { t.Fatalf("failed to build alloc dir: %v", err) } @@ -2055,7 +2055,7 @@ func TestDockerDriver_Mounts(t *testing.T) { ctx := testDockerDriverContexts(t, task) driver := NewDockerDriver(ctx.DriverCtx) copyImage(t, ctx.ExecCtx.TaskDir, "busybox.tar") - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() _, err := driver.Prestart(ctx.ExecCtx, task) if err == nil && c.Error != "" { @@ -2089,7 +2089,7 @@ func TestDockerDriver_Cleanup(t *testing.T) { }, } tctx := testDockerDriverContexts(t, task) - defer tctx.AllocDir.Destroy() + defer tctx.Destroy() // Run Prestart driver := NewDockerDriver(tctx.DriverCtx).(*DockerDriver) @@ -2277,7 +2277,7 @@ func TestDockerDriver_Devices_IsInvalidConfig(t *testing.T) { ctx := testDockerDriverContexts(t, task) driver := NewDockerDriver(ctx.DriverCtx) copyImage(t, ctx.ExecCtx.TaskDir, "busybox.tar") - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() if _, err := driver.Prestart(ctx.ExecCtx, task); err == nil || err.Error() != tc.err.Error() { t.Fatalf("error expected in prestart, got %v, expected %v", err, tc.err) @@ -2382,7 +2382,7 @@ func TestDockerDriver_Kill(t *testing.T) { } ctx := testDockerDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewDockerDriver(ctx.DriverCtx) copyImage(t, ctx.ExecCtx.TaskDir, "busybox.tar") @@ -2502,7 +2502,7 @@ func TestDockerDriver_AdvertiseIPv6Address(t *testing.T) { tctx := testDockerDriverContexts(t, task) driver := NewDockerDriver(tctx.DriverCtx) copyImage(t, tctx.ExecCtx.TaskDir, "busybox.tar") - defer tctx.AllocDir.Destroy() + defer tctx.Destroy() presp, err := driver.Prestart(tctx.ExecCtx, task) defer driver.Cleanup(tctx.ExecCtx, presp.CreatedResources) diff --git a/client/driver/docker_unix_test.go b/client/driver/docker_unix_test.go index 6bed49eef..47a01eca0 100644 --- a/client/driver/docker_unix_test.go +++ b/client/driver/docker_unix_test.go @@ -43,7 +43,7 @@ func TestDockerDriver_Signal(t *testing.T) { } ctx := testDockerDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewDockerDriver(ctx.DriverCtx) // Copy the image into the task's directory diff --git a/client/driver/driver.go b/client/driver/driver.go index a667efe85..c65150a21 100644 --- a/client/driver/driver.go +++ b/client/driver/driver.go @@ -339,6 +339,12 @@ type ExecContext struct { // TaskEnv contains the task's environment variables. TaskEnv *env.TaskEnv + + // StdoutFifo is the path to the named pipe to write stdout to + StdoutFifo string + + // StderrFifo is the path to the named pipe to write stderr to + StderrFifo string } // NewExecContext is used to create a new execution context diff --git a/client/driver/driver_test.go b/client/driver/driver_test.go index ca2c37d1d..4c3112b3f 100644 --- a/client/driver/driver_test.go +++ b/client/driver/driver_test.go @@ -1,20 +1,25 @@ package driver import ( + "fmt" "io" "io/ioutil" "math/rand" "os" "path/filepath" "reflect" + "runtime" "testing" "time" + plugin "github.com/hashicorp/go-plugin" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver/env" + "github.com/hashicorp/nomad/client/logmon" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/helper/testtask" + "github.com/hashicorp/nomad/helper/uuid" "github.com/hashicorp/nomad/nomad/mock" "github.com/hashicorp/nomad/nomad/structs" ) @@ -95,26 +100,36 @@ func testConfig(t *testing.T) *config.Config { conf.MaxKillTimeout = 10 * time.Second conf.Region = "global" conf.Node = mock.Node() + conf.LogLevel = "DEBUG" + conf.LogOutput = testlog.NewWriter(t) return conf } type testContext struct { - AllocDir *allocdir.AllocDir - DriverCtx *DriverContext - ExecCtx *ExecContext - EnvBuilder *env.Builder + AllocDir *allocdir.AllocDir + DriverCtx *DriverContext + ExecCtx *ExecContext + EnvBuilder *env.Builder + logmon logmon.LogMon + logmonPlugin *plugin.Client +} + +func (ctx *testContext) Destroy() { + ctx.AllocDir.Destroy() + ctx.logmon.Stop() + ctx.logmonPlugin.Kill() } // testDriverContext sets up an alloc dir, task dir, DriverContext, and ExecContext. // -// It is up to the caller to call AllocDir.Destroy to cleanup. +// It is up to the caller to call Destroy to cleanup. func testDriverContexts(t *testing.T, task *structs.Task) *testContext { cfg := testConfig(t) cfg.Node = mock.Node() alloc := mock.Alloc() alloc.NodeID = cfg.Node.ID - allocDir := allocdir.NewAllocDir(testlog.Logger(t), filepath.Join(cfg.AllocDir, alloc.ID)) + allocDir := allocdir.NewAllocDir(testlog.HCLogger(t), filepath.Join(cfg.AllocDir, alloc.ID)) if err := allocDir.Build(); err != nil { t.Fatalf("AllocDir.Build() failed: %v", err) } @@ -139,12 +154,45 @@ func testDriverContexts(t *testing.T, task *structs.Task) *testContext { execCtx := NewExecContext(td, eb.Build()) logger := testlog.Logger(t) + hcLogger := testlog.HCLogger(t) emitter := func(m string, args ...interface{}) { - logger.Printf("[EVENT] "+m, args...) + hcLogger.Info(fmt.Sprintf("[EVENT] "+m, args...)) } driverCtx := NewDriverContext(alloc.Job.Name, alloc.TaskGroup, task.Name, alloc.ID, cfg, cfg.Node, logger, emitter) + l, c, err := logmon.LaunchLogMon(hcLogger) + if err != nil { + allocDir.Destroy() + t.Fatalf("LaunchLogMon() failed: %v", err) + } - return &testContext{allocDir, driverCtx, execCtx, eb} + var stdoutFifo, stderrFifo string + if runtime.GOOS == "windows" { + id := uuid.Generate()[:8] + stdoutFifo = fmt.Sprintf("//./pipe/%s.stdout.%s", id, task.Name) + stderrFifo = fmt.Sprintf("//./pipe/%s.stderr.%s", id, task.Name) + } else { + stdoutFifo = filepath.Join(td.LogDir, fmt.Sprintf("%s.stdout", task.Name)) + stderrFifo = filepath.Join(td.LogDir, fmt.Sprintf("%s.stderr", task.Name)) + } + + err = l.Start(&logmon.LogConfig{ + LogDir: td.LogDir, + StdoutLogFile: fmt.Sprintf("%s.stdout", task.Name), + StderrLogFile: fmt.Sprintf("%s.stderr", task.Name), + StdoutFifo: stdoutFifo, + StderrFifo: stderrFifo, + MaxFiles: 10, + MaxFileSizeMB: 10, + }) + if err != nil { + allocDir.Destroy() + t.Fatalf("LogMon.Start() failed: %v", err) + } + + execCtx.StdoutFifo = stdoutFifo + execCtx.StderrFifo = stderrFifo + + return &testContext{allocDir, driverCtx, execCtx, eb, l, c} } // setupTaskEnv creates a test env for GetTaskEnv testing. Returns task dir, @@ -179,7 +227,7 @@ func setupTaskEnv(t *testing.T, driver string) (*allocdir.TaskDir, map[string]st alloc.Name = "Bar" alloc.AllocatedResources.Tasks["web"].Networks[0].DynamicPorts[0].Value = 2000 conf := testConfig(t) - allocDir := allocdir.NewAllocDir(testlog.Logger(t), filepath.Join(conf.AllocDir, alloc.ID)) + allocDir := allocdir.NewAllocDir(testlog.HCLogger(t), filepath.Join(conf.AllocDir, alloc.ID)) taskDir := allocDir.NewTaskDir(task.Name) eb := env.NewBuilder(conf.Node, alloc, task, conf.Region) tmpDriver, err := NewDriver(driver, NewEmptyDriverContext()) diff --git a/client/driver/exec.go b/client/driver/exec.go index edab4143f..e7e0afe78 100644 --- a/client/driver/exec.go +++ b/client/driver/exec.go @@ -37,17 +37,17 @@ type ExecDriverConfig struct { // execHandle is returned from Start/Open as a handle to the PID type execHandle struct { - pluginClient *plugin.Client - executor executor.Executor - isolationConfig *dstructs.IsolationConfig - userPid int - taskDir *allocdir.TaskDir - killTimeout time.Duration - maxKillTimeout time.Duration - logger *log.Logger - waitCh chan *dstructs.WaitResult - doneCh chan struct{} - version string + pluginClient *plugin.Client + executor executor.Executor + userPid int + taskShutdownSignal string + taskDir *allocdir.TaskDir + killTimeout time.Duration + maxKillTimeout time.Duration + logger *log.Logger + waitCh chan *dstructs.WaitResult + doneCh chan struct{} + version string } // NewExecDriver is used to create a new exec driver @@ -110,26 +110,16 @@ func (d *ExecDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse pluginLogFile := filepath.Join(ctx.TaskDir.Dir, "executor.out") executorConfig := &dstructs.ExecutorConfig{ - LogFile: pluginLogFile, - LogLevel: d.config.LogLevel, + LogFile: pluginLogFile, + LogLevel: d.config.LogLevel, + FSIsolation: true, } exec, pluginClient, err := createExecutor(d.config.LogOutput, d.config, executorConfig) if err != nil { return nil, err } - executorCtx := &executor.ExecutorContext{ - TaskEnv: ctx.TaskEnv, - Driver: "exec", - LogDir: ctx.TaskDir.LogDir, - TaskDir: ctx.TaskDir.Dir, - Task: task, - } - if err := exec.SetContext(executorCtx); err != nil { - pluginClient.Kill() - return nil, fmt.Errorf("failed to set executor context: %v", err) - } - taskKillSignal, err := getTaskKillSignal(task.KillSignal) + _, err = getTaskKillSignal(task.KillSignal) if err != nil { return nil, err } @@ -137,13 +127,21 @@ func (d *ExecDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse execCmd := &executor.ExecCommand{ Cmd: command, Args: driverConfig.Args, - TaskKillSignal: taskKillSignal, - FSIsolation: true, ResourceLimits: true, User: getExecutorUser(task), + Resources: &executor.Resources{ + CPU: task.Resources.CPU, + MemoryMB: task.Resources.MemoryMB, + IOPS: task.Resources.IOPS, + DiskMB: task.Resources.DiskMB, + }, + Env: ctx.TaskEnv.List(), + TaskDir: ctx.TaskDir.Dir, + StdoutPath: ctx.StdoutFifo, + StderrPath: ctx.StderrFifo, } - ps, err := exec.LaunchCmd(execCmd) + ps, err := exec.Launch(execCmd) if err != nil { pluginClient.Kill() return nil, err @@ -154,17 +152,17 @@ func (d *ExecDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse // Return a driver handle maxKill := d.DriverContext.config.MaxKillTimeout h := &execHandle{ - pluginClient: pluginClient, - userPid: ps.Pid, - executor: exec, - isolationConfig: ps.IsolationConfig, - killTimeout: GetKillTimeout(task.KillTimeout, maxKill), - maxKillTimeout: maxKill, - logger: d.logger, - version: d.config.Version.VersionNumber(), - doneCh: make(chan struct{}), - waitCh: make(chan *dstructs.WaitResult, 1), - taskDir: ctx.TaskDir, + pluginClient: pluginClient, + userPid: ps.Pid, + taskShutdownSignal: task.KillSignal, + executor: exec, + killTimeout: GetKillTimeout(task.KillTimeout, maxKill), + maxKillTimeout: maxKill, + logger: d.logger, + version: d.config.Version.VersionNumber(), + doneCh: make(chan struct{}), + waitCh: make(chan *dstructs.WaitResult, 1), + taskDir: ctx.TaskDir, } go h.run() return &StartResponse{Handle: h}, nil @@ -173,12 +171,11 @@ func (d *ExecDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse func (d *ExecDriver) Cleanup(*ExecContext, *CreatedResources) error { return nil } type execId struct { - Version string - KillTimeout time.Duration - MaxKillTimeout time.Duration - UserPid int - IsolationConfig *dstructs.IsolationConfig - PluginConfig *PluginReattachConfig + Version string + KillTimeout time.Duration + MaxKillTimeout time.Duration + UserPid int + PluginConfig *PluginReattachConfig } func (d *ExecDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, error) { @@ -198,12 +195,6 @@ func (d *ExecDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, erro if e := destroyPlugin(id.PluginConfig.Pid, id.UserPid); e != nil { merrs.Errors = append(merrs.Errors, fmt.Errorf("error destroying plugin and userpid: %v", e)) } - if id.IsolationConfig != nil { - ePid := pluginConfig.Reattach.Pid - if e := executor.ClientCleanup(id.IsolationConfig, ePid); e != nil { - merrs.Errors = append(merrs.Errors, fmt.Errorf("destroying cgroup failed: %v", e)) - } - } return nil, fmt.Errorf("error connecting to plugin: %v", merrs.ErrorOrNil()) } @@ -211,17 +202,16 @@ func (d *ExecDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, erro d.logger.Printf("[DEBUG] driver.exec : version of executor: %v", ver.Version) // Return a driver handle h := &execHandle{ - pluginClient: client, - executor: exec, - userPid: id.UserPid, - isolationConfig: id.IsolationConfig, - logger: d.logger, - version: id.Version, - killTimeout: id.KillTimeout, - maxKillTimeout: id.MaxKillTimeout, - doneCh: make(chan struct{}), - waitCh: make(chan *dstructs.WaitResult, 1), - taskDir: ctx.TaskDir, + pluginClient: client, + executor: exec, + userPid: id.UserPid, + logger: d.logger, + version: id.Version, + killTimeout: id.KillTimeout, + maxKillTimeout: id.MaxKillTimeout, + doneCh: make(chan struct{}), + waitCh: make(chan *dstructs.WaitResult, 1), + taskDir: ctx.TaskDir, } go h.run() return h, nil @@ -229,12 +219,11 @@ func (d *ExecDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, erro func (h *execHandle) ID() string { id := execId{ - Version: h.version, - KillTimeout: h.killTimeout, - MaxKillTimeout: h.maxKillTimeout, - PluginConfig: NewPluginReattachConfig(h.pluginClient.ReattachConfig()), - UserPid: h.userPid, - IsolationConfig: h.isolationConfig, + Version: h.version, + KillTimeout: h.killTimeout, + MaxKillTimeout: h.maxKillTimeout, + PluginConfig: NewPluginReattachConfig(h.pluginClient.ReattachConfig()), + UserPid: h.userPid, } data, err := json.Marshal(id) @@ -251,7 +240,12 @@ func (h *execHandle) WaitCh() chan *dstructs.WaitResult { func (h *execHandle) Update(task *structs.Task) error { // Store the updated kill timeout. h.killTimeout = GetKillTimeout(task.KillTimeout, h.maxKillTimeout) - h.executor.UpdateTask(task) + h.executor.UpdateResources(&executor.Resources{ + CPU: task.Resources.CPU, + MemoryMB: task.Resources.MemoryMB, + IOPS: task.Resources.IOPS, + DiskMB: task.Resources.DiskMB, + }) // Update is not possible return nil @@ -275,11 +269,11 @@ func (d *execHandle) Network() *cstructs.DriverNetwork { } func (h *execHandle) Kill() error { - if err := h.executor.ShutDown(); err != nil { + if err := h.executor.Shutdown(h.taskShutdownSignal, h.killTimeout); err != nil { if h.pluginClient.Exited() { return nil } - return fmt.Errorf("executor Shutdown failed: %v", err) + return fmt.Errorf("executor Kill failed: %v", err) } select { @@ -288,8 +282,8 @@ func (h *execHandle) Kill() error { if h.pluginClient.Exited() { break } - if err := h.executor.Exit(); err != nil { - return fmt.Errorf("executor Exit failed: %v", err) + if err := h.executor.Shutdown(h.taskShutdownSignal, h.killTimeout); err != nil { + return fmt.Errorf("executor Destroy failed: %v", err) } } return nil @@ -303,22 +297,8 @@ func (h *execHandle) run() { ps, werr := h.executor.Wait() close(h.doneCh) - // If the exitcode is 0 and we had an error that means the plugin didn't - // connect and doesn't know the state of the user process so we are killing - // the user process so that when we create a new executor on restarting the - // new user process doesn't have collisions with resources that the older - // user pid might be holding onto. - if ps.ExitCode == 0 && werr != nil { - if h.isolationConfig != nil { - ePid := h.pluginClient.ReattachConfig().Pid - if e := executor.ClientCleanup(h.isolationConfig, ePid); e != nil { - h.logger.Printf("[ERR] driver.exec: destroying resource container failed: %v", e) - } - } - } - - // Exit the executor - if err := h.executor.Exit(); err != nil { + // Destroy the executor + if err := h.executor.Shutdown(h.taskShutdownSignal, 0); err != nil { h.logger.Printf("[ERR] driver.exec: error destroying executor: %v", err) } h.pluginClient.Kill() diff --git a/client/driver/exec_test.go b/client/driver/exec_test.go index 3dcbddd2a..787d008f0 100644 --- a/client/driver/exec_test.go +++ b/client/driver/exec_test.go @@ -56,7 +56,7 @@ func TestExecDriver_Fingerprint(t *testing.T) { Resources: structs.DefaultResources(), } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewExecDriver(ctx.DriverCtx) node := &structs.Node{ Attributes: map[string]string{ @@ -100,7 +100,7 @@ func TestExecDriver_StartOpen_Wait(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -144,7 +144,7 @@ func TestExecDriver_Start_Wait(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -198,7 +198,7 @@ func TestExecDriver_Start_Wait_AllocDir(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -252,7 +252,7 @@ func TestExecDriver_Start_Kill_Wait(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -304,7 +304,7 @@ func TestExecDriverUser(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -343,7 +343,7 @@ func TestExecDriver_HandlerExec(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + //defer ctx.Destroy() d := NewExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -397,7 +397,7 @@ func TestExecDriver_HandlerExec(t *testing.T) { if line == "" { continue } - if !strings.Contains(line, ":/nomad/") { + if !strings.Contains(line, ":/nomad/") && !strings.Contains(line, ":name=") { t.Errorf("Not a member of the alloc's cgroup: expected=...:/nomad/... -- found=%q", line) continue } @@ -420,6 +420,7 @@ func TestExecDriver_HandlerExec(t *testing.T) { } if err := handle.Kill(); err != nil { + t.Logf("Check allocdir: %x", ctx.AllocDir.AllocDir) t.Fatalf("error killing exec handle: %v", err) } } diff --git a/client/driver/exec_unix_test.go b/client/driver/exec_unix_test.go index 18283c680..7a1af5d81 100644 --- a/client/driver/exec_unix_test.go +++ b/client/driver/exec_unix_test.go @@ -38,7 +38,7 @@ func TestExecDriver_KillUserPid_OnPluginReconnectFailure(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -111,7 +111,7 @@ func TestExecDriver_Signal(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewExecDriver(ctx.DriverCtx) testFile := filepath.Join(ctx.ExecCtx.TaskDir.Dir, "test.sh") diff --git a/client/driver/executor/executor.go b/client/driver/executor/executor.go index 76272372f..280f107c5 100644 --- a/client/driver/executor/executor.go +++ b/client/driver/executor/executor.go @@ -5,46 +5,35 @@ import ( "fmt" "io" "io/ioutil" - "log" - "net" "os" "os/exec" "path/filepath" "runtime" - "strconv" "strings" - "sync" "syscall" "time" "github.com/armon/circbuf" - "github.com/hashicorp/go-multierror" - "github.com/mitchellh/go-ps" - "github.com/shirou/gopsutil/process" + hclog "github.com/hashicorp/go-hclog" + multierror "github.com/hashicorp/go-multierror" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/driver/env" - "github.com/hashicorp/nomad/client/driver/logging" + "github.com/hashicorp/nomad/client/lib/fifo" "github.com/hashicorp/nomad/client/stats" shelpers "github.com/hashicorp/nomad/helper/stats" - "github.com/hashicorp/nomad/nomad/structs" - - syslog "github.com/RackSec/srslog" + "github.com/hashicorp/consul-template/signals" dstructs "github.com/hashicorp/nomad/client/driver/structs" cstructs "github.com/hashicorp/nomad/client/structs" ) const ( - // pidScanInterval is the interval at which the executor scans the process - // tree for finding out the pids that the executor and it's child processes - // have forked - pidScanInterval = 5 * time.Second + // ExecutorVersionLatest is the current and latest version of the executor + ExecutorVersionLatest = "2.0.0" - // processOutputCloseTolerance is the length of time we will wait for the - // launched process to close its stdout/stderr before we force close it. If - // data is written after this tolerance, we will not capture it. - processOutputCloseTolerance = 2 * time.Second + // ExecutorVersionPre0_9 is the version of executor use prior to the release + // of 0.9.x + ExecutorVersionPre0_9 = "1.1.0" ) var ( @@ -56,45 +45,46 @@ var ( // Executor is the interface which allows a driver to launch and supervise // a process type Executor interface { - SetContext(ctx *ExecutorContext) error - LaunchCmd(command *ExecCommand) (*ProcessState, error) - LaunchSyslogServer() (*SyslogServerState, error) + // Launch a user process configured by the given ExecCommand + Launch(launchCmd *ExecCommand) (*ProcessState, error) + + // Wait blocks until the process exits or an error occures Wait() (*ProcessState, error) - ShutDown() error - Exit() error - UpdateLogConfig(logConfig *structs.LogConfig) error - UpdateTask(task *structs.Task) error + + // Shutdown will shutdown the executor by stopping the user process, + // cleaning up and resources created by the executor. The shutdown sequence + // will first send the given signal to the process. This defaults to "SIGINT" + // if not specified. The executor will then wait for the process to exit + // before cleaning up other resources. If the executor waits longer than the + // given grace period, the process is forcefully killed. + // + // To force kill the user process, gracePeriod can be set to 0. + Shutdown(signal string, gracePeriod time.Duration) error + + // UpdateResources updates any resource isolation enforcement with new + // constraints if supported. + UpdateResources(*Resources) error + + // Version returns the executor API version Version() (*ExecutorVersion, error) + + // Stats fetchs process usage stats for the executor and each pid if availble Stats() (*cstructs.TaskResourceUsage, error) - Signal(s os.Signal) error + + // Signal sends the given signal to the user process + Signal(os.Signal) error + + // Exec executes the given command and args inside the executor context + // and returns the output and exit code. Exec(deadline time.Time, cmd string, args []string) ([]byte, int, error) } -// ExecutorContext holds context to configure the command user -// wants to run and isolate it -type ExecutorContext struct { - // TaskEnv holds information about the environment of a Task - TaskEnv *env.TaskEnv - - // Task is the task whose executor is being launched - Task *structs.Task - - // TaskDir is the host path to the task's root - TaskDir string - - // LogDir is the host path where logs should be written - LogDir string - - // Driver is the name of the driver that invoked the executor - Driver string - - // PortUpperBound is the upper bound of the ports that we can use to start - // the syslog server - PortUpperBound uint - - // PortLowerBound is the lower bound of the ports that we can use to start - // the syslog server - PortLowerBound uint +// Resources describes the resource isolation required +type Resources struct { + CPU int + MemoryMB int + DiskMB int + IOPS int } // ExecCommand holds the user command, args, and other isolation related @@ -106,15 +96,26 @@ type ExecCommand struct { // Args is the args of the command that the user wants to run. Args []string - // TaskKillSignal is an optional field which signal to kill the process - TaskKillSignal os.Signal + // Resources defined by the task + Resources *Resources - // FSIsolation determines whether the command would be run in a chroot. - FSIsolation bool + // StdoutPath is the path the procoess stdout should be written to + StdoutPath string + stdout io.WriteCloser + + // StderrPath is the path the procoess stderr should be written to + StderrPath string + stderr io.WriteCloser + + // Env is the list of KEY=val pairs of environment variables to be set + Env []string // User is the user which the executor uses to run the command. User string + // TaskDir is the directory path on the host where for the task + TaskDir string + // ResourceLimits determines whether resource limits are enforced by the // executor. ResourceLimits bool @@ -125,28 +126,61 @@ type ExecCommand struct { BasicProcessCgroup bool } +type nopCloser struct { + io.Writer +} + +func (nopCloser) Close() error { return nil } + +// Stdout returns a writer for the configured file descriptor +func (c *ExecCommand) Stdout() (io.WriteCloser, error) { + if c.stdout == nil { + if c.StderrPath != "" { + f, err := fifo.Open(c.StdoutPath) + if err != nil { + return nil, fmt.Errorf("failed to create stdout: %v", err) + } + c.stdout = f + } else { + c.stdout = nopCloser{ioutil.Discard} + } + } + return c.stdout, nil +} + +// Stderr returns a writer for the configured file descriptor +func (c *ExecCommand) Stderr() (io.WriteCloser, error) { + if c.stderr == nil { + if c.StderrPath != "" { + f, err := fifo.Open(c.StderrPath) + if err != nil { + return nil, fmt.Errorf("failed to create stderr: %v", err) + } + c.stderr = f + } else { + c.stderr = nopCloser{ioutil.Discard} + } + } + return c.stderr, nil +} + +func (c *ExecCommand) Close() { + stdout, err := c.Stdout() + if err == nil { + stdout.Close() + } + stderr, err := c.Stderr() + if err == nil { + stderr.Close() + } +} + // ProcessState holds information about the state of a user process. type ProcessState struct { - Pid int - ExitCode int - Signal int - IsolationConfig *dstructs.IsolationConfig - Time time.Time -} - -// nomadPid holds a pid and it's cpu percentage calculator -type nomadPid struct { - pid int - cpuStatsTotal *stats.CpuStats - cpuStatsUser *stats.CpuStats - cpuStatsSys *stats.CpuStats -} - -// SyslogServerState holds the address and isolation information of a launched -// syslog server -type SyslogServerState struct { - IsolationConfig *dstructs.IsolationConfig - Addr string + Pid int + ExitCode int + Signal int + Time time.Time } // ExecutorVersion is the version of the executor @@ -162,111 +196,87 @@ func (v *ExecutorVersion) GoString() string { // supervises processes. In addition to process supervision it provides resource // and file system isolation type UniversalExecutor struct { - cmd exec.Cmd - ctx *ExecutorContext - command *ExecCommand + childCmd exec.Cmd + commandCfg *ExecCommand - pids map[int]*nomadPid - pidLock sync.RWMutex - exitState *ProcessState - processExited chan interface{} - fsIsolationEnforced bool - - lre *logRotatorWrapper - lro *logRotatorWrapper - rotatorLock sync.Mutex - - syslogServer *logging.SyslogServer - syslogChan chan *logging.SyslogMessage + exitState *ProcessState + processExited chan interface{} + // resConCtx is used to track and cleanup additional resources created by + // the executor. Currently this is only used for cgroups. resConCtx resourceContainerContext totalCpuStats *stats.CpuStats userCpuStats *stats.CpuStats systemCpuStats *stats.CpuStats - logger *log.Logger + pidCollector *pidCollector + + logger hclog.Logger } // NewExecutor returns an Executor -func NewExecutor(logger *log.Logger) Executor { +func NewExecutor(logger hclog.Logger) Executor { + logger = logger.Named("executor") if err := shelpers.Init(); err != nil { - logger.Printf("[ERR] executor: unable to initialize stats: %v", err) + logger.Error("unable to initialize stats", "error", err) } - - exec := &UniversalExecutor{ + return &UniversalExecutor{ logger: logger, processExited: make(chan interface{}), totalCpuStats: stats.NewCpuStats(), userCpuStats: stats.NewCpuStats(), systemCpuStats: stats.NewCpuStats(), - pids: make(map[int]*nomadPid), + pidCollector: newPidCollector(logger), } - - return exec } // Version returns the api version of the executor func (e *UniversalExecutor) Version() (*ExecutorVersion, error) { - return &ExecutorVersion{Version: "1.1.0"}, nil + return &ExecutorVersion{Version: ExecutorVersionLatest}, nil } -// SetContext is used to set the executors context and should be the first call -// after launching the executor. -func (e *UniversalExecutor) SetContext(ctx *ExecutorContext) error { - e.ctx = ctx - return nil -} - -// LaunchCmd launches the main process and returns its state. It also +// Launch launches the main process and returns its state. It also // configures an applies isolation on certain platforms. -func (e *UniversalExecutor) LaunchCmd(command *ExecCommand) (*ProcessState, error) { - e.logger.Printf("[INFO] executor: launching command %v %v", command.Cmd, strings.Join(command.Args, " ")) +func (e *UniversalExecutor) Launch(command *ExecCommand) (*ProcessState, error) { + e.logger.Info("launching command", "command", command.Cmd, "args", strings.Join(command.Args, " ")) - // Ensure the context has been set first - if e.ctx == nil { - return nil, fmt.Errorf("SetContext must be called before launching a command") - } - - e.command = command + e.commandCfg = command // setting the user of the process if command.User != "" { - e.logger.Printf("[DEBUG] executor: running command as %s", command.User) + e.logger.Debug("running command as user", "user", command.User) if err := e.runAs(command.User); err != nil { return nil, err } } // set the task dir as the working directory for the command - e.cmd.Dir = e.ctx.TaskDir + e.childCmd.Dir = e.commandCfg.TaskDir // start command in separate process group if err := e.setNewProcessGroup(); err != nil { return nil, err } - // configuring the chroot, resource container, and start the plugin - // process in the chroot. - if err := e.configureIsolation(); err != nil { - return nil, err - } - // Apply ourselves into the resource container. The executor MUST be in - // the resource container before the user task is started, otherwise we - // are subject to a fork attack in which a process escapes isolation by - // immediately forking. - if err := e.applyLimits(os.Getpid()); err != nil { + // Setup cgroups on linux + if err := e.configureResourceContainer(os.Getpid()); err != nil { return nil, err } - // Setup the loggers - if err := e.configureLoggers(); err != nil { + stdout, err := e.commandCfg.Stdout() + if err != nil { return nil, err } - e.cmd.Stdout = e.lro.processOutWriter - e.cmd.Stderr = e.lre.processOutWriter + stderr, err := e.commandCfg.Stderr() + if err != nil { + return nil, err + } + + e.childCmd.Stdout = stdout + e.childCmd.Stderr = stderr // Look up the binary path and make it executable - absPath, err := e.lookupBin(e.ctx.TaskEnv.ReplaceEnv(command.Cmd)) + absPath, err := e.lookupBin(command.Cmd) if err != nil { return nil, err } @@ -277,53 +287,38 @@ func (e *UniversalExecutor) LaunchCmd(command *ExecCommand) (*ProcessState, erro path := absPath - // Determine the path to run as it may have to be relative to the chroot. - if e.fsIsolationEnforced { - rel, err := filepath.Rel(e.ctx.TaskDir, path) - if err != nil { - return nil, fmt.Errorf("failed to determine relative path base=%q target=%q: %v", e.ctx.TaskDir, path, err) - } - path = rel - } - // Set the commands arguments - e.cmd.Path = path - e.cmd.Args = append([]string{e.cmd.Path}, e.ctx.TaskEnv.ParseAndReplace(command.Args)...) - e.cmd.Env = e.ctx.TaskEnv.List() + e.childCmd.Path = path + e.childCmd.Args = append([]string{e.childCmd.Path}, command.Args...) + e.childCmd.Env = e.commandCfg.Env // Start the process - if err := e.cmd.Start(); err != nil { - return nil, fmt.Errorf("failed to start command path=%q --- args=%q: %v", path, e.cmd.Args, err) + if err := e.childCmd.Start(); err != nil { + return nil, fmt.Errorf("failed to start command path=%q --- args=%q: %v", path, e.childCmd.Args, err) } - // Close the files. This is copied from the os/exec package. - e.lro.processOutWriter.Close() - e.lre.processOutWriter.Close() - - go e.collectPids() + go e.pidCollector.collectPids(e.processExited, getAllPids) go e.wait() - ic := e.resConCtx.getIsolationConfig() - return &ProcessState{Pid: e.cmd.Process.Pid, ExitCode: -1, IsolationConfig: ic, Time: time.Now()}, nil + return &ProcessState{Pid: e.childCmd.Process.Pid, ExitCode: -1, Time: time.Now()}, nil } // Exec a command inside a container for exec and java drivers. func (e *UniversalExecutor) Exec(deadline time.Time, name string, args []string) ([]byte, int, error) { ctx, cancel := context.WithDeadline(context.Background(), deadline) defer cancel() - return ExecScript(ctx, e.cmd.Dir, e.ctx.TaskEnv, e.cmd.SysProcAttr, name, args) + return ExecScript(ctx, e.childCmd.Dir, e.commandCfg.Env, e.childCmd.SysProcAttr, name, args) } // ExecScript executes cmd with args and returns the output, exit code, and // error. Output is truncated to client/driver/structs.CheckBufSize -func ExecScript(ctx context.Context, dir string, env *env.TaskEnv, attrs *syscall.SysProcAttr, +func ExecScript(ctx context.Context, dir string, env []string, attrs *syscall.SysProcAttr, name string, args []string) ([]byte, int, error) { - name = env.ReplaceEnv(name) - cmd := exec.CommandContext(ctx, name, env.ParseAndReplace(args)...) + cmd := exec.CommandContext(ctx, name, args...) // Copy runtime environment from the main command cmd.SysProcAttr = attrs cmd.Dir = dir - cmd.Env = env.List() + cmd.Env = env // Capture output buf, _ := circbuf.NewBuffer(int64(dstructs.CheckBufSize)) @@ -351,93 +346,25 @@ func ExecScript(ctx context.Context, dir string, env *env.TaskEnv, attrs *syscal return buf.Bytes(), 0, nil } -// configureLoggers sets up the standard out/error file rotators -func (e *UniversalExecutor) configureLoggers() error { - e.rotatorLock.Lock() - defer e.rotatorLock.Unlock() - - logFileSize := int64(e.ctx.Task.LogConfig.MaxFileSizeMB * 1024 * 1024) - if e.lro == nil { - lro, err := logging.NewFileRotator(e.ctx.LogDir, fmt.Sprintf("%v.stdout", e.ctx.Task.Name), - e.ctx.Task.LogConfig.MaxFiles, logFileSize, e.logger) - if err != nil { - return fmt.Errorf("error creating new stdout log file for %q: %v", e.ctx.Task.Name, err) - } - - r, err := newLogRotatorWrapper(e.logger, lro) - if err != nil { - return err - } - e.lro = r - } - - if e.lre == nil { - lre, err := logging.NewFileRotator(e.ctx.LogDir, fmt.Sprintf("%v.stderr", e.ctx.Task.Name), - e.ctx.Task.LogConfig.MaxFiles, logFileSize, e.logger) - if err != nil { - return fmt.Errorf("error creating new stderr log file for %q: %v", e.ctx.Task.Name, err) - } - - r, err := newLogRotatorWrapper(e.logger, lre) - if err != nil { - return err - } - e.lre = r - } - return nil -} - // Wait waits until a process has exited and returns it's exitcode and errors func (e *UniversalExecutor) Wait() (*ProcessState, error) { <-e.processExited return e.exitState, nil } -// COMPAT: prior to Nomad 0.3.2, UpdateTask didn't exist. -// UpdateLogConfig updates the log configuration -func (e *UniversalExecutor) UpdateLogConfig(logConfig *structs.LogConfig) error { - e.ctx.Task.LogConfig = logConfig - if e.lro == nil { - return fmt.Errorf("log rotator for stdout doesn't exist") - } - e.lro.rotatorWriter.MaxFiles = logConfig.MaxFiles - e.lro.rotatorWriter.FileSize = int64(logConfig.MaxFileSizeMB * 1024 * 1024) - - if e.lre == nil { - return fmt.Errorf("log rotator for stderr doesn't exist") - } - e.lre.rotatorWriter.MaxFiles = logConfig.MaxFiles - e.lre.rotatorWriter.FileSize = int64(logConfig.MaxFileSizeMB * 1024 * 1024) - return nil -} - -func (e *UniversalExecutor) UpdateTask(task *structs.Task) error { - e.ctx.Task = task - - // Updating Log Config - e.rotatorLock.Lock() - if e.lro != nil && e.lre != nil { - fileSize := int64(task.LogConfig.MaxFileSizeMB * 1024 * 1024) - e.lro.rotatorWriter.MaxFiles = task.LogConfig.MaxFiles - e.lro.rotatorWriter.FileSize = fileSize - e.lre.rotatorWriter.MaxFiles = task.LogConfig.MaxFiles - e.lre.rotatorWriter.FileSize = fileSize - } - e.rotatorLock.Unlock() +func (e *UniversalExecutor) UpdateResources(resources *Resources) error { return nil } func (e *UniversalExecutor) wait() { defer close(e.processExited) - err := e.cmd.Wait() - ic := e.resConCtx.getIsolationConfig() + err := e.childCmd.Wait() if err == nil { - e.exitState = &ProcessState{Pid: 0, ExitCode: 0, IsolationConfig: ic, Time: time.Now()} + e.exitState = &ProcessState{Pid: 0, ExitCode: 0, Time: time.Now()} return } - e.lre.Close() - e.lro.Close() + e.commandCfg.Close() exitCode := 1 var signal int @@ -458,10 +385,10 @@ func (e *UniversalExecutor) wait() { } } } else { - e.logger.Printf("[WARN] executor: unexpected Cmd.Wait() error type: %v", err) + e.logger.Warn("unexpected Cmd.Wait() error type", "error", err) } - e.exitState = &ProcessState{Pid: 0, ExitCode: exitCode, Signal: signal, IsolationConfig: ic, Time: time.Now()} + e.exitState = &ProcessState{Pid: 0, ExitCode: exitCode, Signal: signal, Time: time.Now()} } var ( @@ -474,114 +401,80 @@ var ( noSuchProcessErr = "no such process" ) -// ClientCleanup is the cleanup routine that a Nomad Client uses to remove the -// remnants of a child UniversalExecutor. -func ClientCleanup(ic *dstructs.IsolationConfig, pid int) error { - return clientCleanup(ic, pid) -} - // Exit cleans up the alloc directory, destroys resource container and kills the // user process -func (e *UniversalExecutor) Exit() error { +func (e *UniversalExecutor) Shutdown(signal string, grace time.Duration) error { var merr multierror.Error - if e.syslogServer != nil { - e.syslogServer.Shutdown() - } - - if e.lre != nil { - e.lre.Close() - } - - if e.lro != nil { - e.lro.Close() - } // If the executor did not launch a process, return. - if e.command == nil { + if e.commandCfg == nil { return nil } + // If there is no process we can't shutdown + if e.childCmd.Process == nil { + return fmt.Errorf("executor failed to shutdown error: no process found") + } + + proc, err := os.FindProcess(e.childCmd.Process.Pid) + if err != nil { + return fmt.Errorf("executor failed to find process: %v", err) + } + + // If grace is 0 then skip shutdown logic + if grace > 0 { + // Default signal to SIGINT if not set + if signal == "" { + signal = "SIGINT" + } + + sig, ok := signals.SignalLookup[signal] + if !ok { + return fmt.Errorf("error unknown signal given for shutdown: %s", signal) + } + + if err := e.shutdownProcess(sig, proc); err != nil { + return err + } + + select { + case <-e.processExited: + case <-time.After(grace): + proc.Kill() + } + } else { + proc.Kill() + } + // Prefer killing the process via the resource container. - if e.cmd.Process != nil && !(e.command.ResourceLimits || e.command.BasicProcessCgroup) { - proc, err := os.FindProcess(e.cmd.Process.Pid) - if err != nil { - e.logger.Printf("[ERR] executor: can't find process with pid: %v, err: %v", - e.cmd.Process.Pid, err) - } else if err := e.cleanupChildProcesses(proc); err != nil && err.Error() != finishedErr { + if !(e.commandCfg.ResourceLimits || e.commandCfg.BasicProcessCgroup) { + if err := e.cleanupChildProcesses(proc); err != nil && err.Error() != finishedErr { merr.Errors = append(merr.Errors, - fmt.Errorf("can't kill process with pid: %v, err: %v", e.cmd.Process.Pid, err)) + fmt.Errorf("can't kill process with pid %d: %v", e.childCmd.Process.Pid, err)) } } - if e.command.ResourceLimits || e.command.BasicProcessCgroup { + if e.commandCfg.ResourceLimits || e.commandCfg.BasicProcessCgroup { if err := e.resConCtx.executorCleanup(); err != nil { merr.Errors = append(merr.Errors, err) } } + return merr.ErrorOrNil() } -// Shutdown sends an interrupt signal to the user process -func (e *UniversalExecutor) ShutDown() error { - if e.cmd.Process == nil { - return fmt.Errorf("executor.shutdown error: no process found") - } - proc, err := os.FindProcess(e.cmd.Process.Pid) - if err != nil { - return fmt.Errorf("executor.shutdown failed to find process: %v", err) - } - return e.shutdownProcess(proc) -} - -// pidStats returns the resource usage stats per pid -func (e *UniversalExecutor) pidStats() (map[string]*cstructs.ResourceUsage, error) { - stats := make(map[string]*cstructs.ResourceUsage) - e.pidLock.RLock() - pids := make(map[int]*nomadPid, len(e.pids)) - for k, v := range e.pids { - pids[k] = v - } - e.pidLock.RUnlock() - for pid, np := range pids { - p, err := process.NewProcess(int32(pid)) - if err != nil { - e.logger.Printf("[TRACE] executor: unable to create new process with pid: %v", pid) - continue - } - ms := &cstructs.MemoryStats{} - if memInfo, err := p.MemoryInfo(); err == nil { - ms.RSS = memInfo.RSS - ms.Swap = memInfo.Swap - ms.Measured = ExecutorBasicMeasuredMemStats - } - - cs := &cstructs.CpuStats{} - if cpuStats, err := p.Times(); err == nil { - cs.SystemMode = np.cpuStatsSys.Percent(cpuStats.System * float64(time.Second)) - cs.UserMode = np.cpuStatsUser.Percent(cpuStats.User * float64(time.Second)) - cs.Measured = ExecutorBasicMeasuredCpuStats - - // calculate cpu usage percent - cs.Percent = np.cpuStatsTotal.Percent(cpuStats.Total() * float64(time.Second)) - } - stats[strconv.Itoa(pid)] = &cstructs.ResourceUsage{MemoryStats: ms, CpuStats: cs} - } - - return stats, nil -} - // lookupBin looks for path to the binary to run by looking for the binary in // the following locations, in-order: task/local/, task/, based on host $PATH. // The return path is absolute. func (e *UniversalExecutor) lookupBin(bin string) (string, error) { // Check in the local directory - local := filepath.Join(e.ctx.TaskDir, allocdir.TaskLocal, bin) + local := filepath.Join(e.commandCfg.TaskDir, allocdir.TaskLocal, bin) if _, err := os.Stat(local); err == nil { return local, nil } // Check at the root of the task's directory - root := filepath.Join(e.ctx.TaskDir, bin) + root := filepath.Join(e.commandCfg.TaskDir, bin) if _, err := os.Stat(root); err == nil { return root, nil } @@ -619,309 +512,26 @@ func (e *UniversalExecutor) makeExecutable(binPath string) error { return nil } -// getFreePort returns a free port ready to be listened on between upper and -// lower bounds -func (e *UniversalExecutor) getListener(lowerBound uint, upperBound uint) (net.Listener, error) { - if runtime.GOOS == "windows" { - return e.listenerTCP(lowerBound, upperBound) - } - - return e.listenerUnix() -} - -// listenerTCP creates a TCP listener using an unused port between an upper and -// lower bound -func (e *UniversalExecutor) listenerTCP(lowerBound uint, upperBound uint) (net.Listener, error) { - for i := lowerBound; i <= upperBound; i++ { - addr, err := net.ResolveTCPAddr("tcp", fmt.Sprintf("localhost:%v", i)) - if err != nil { - return nil, err - } - l, err := net.ListenTCP("tcp", addr) - if err != nil { - continue - } - return l, nil - } - return nil, fmt.Errorf("No free port found") -} - -// listenerUnix creates a Unix domain socket -func (e *UniversalExecutor) listenerUnix() (net.Listener, error) { - f, err := ioutil.TempFile("", "plugin") - if err != nil { - return nil, err - } - path := f.Name() - - if err := f.Close(); err != nil { - return nil, err - } - if err := os.Remove(path); err != nil { - return nil, err - } - - return net.Listen("unix", path) -} - -// collectPids collects the pids of the child processes that the executor is -// running every 5 seconds -func (e *UniversalExecutor) collectPids() { - // Fire the timer right away when the executor starts from there on the pids - // are collected every scan interval - timer := time.NewTimer(0) - defer timer.Stop() - for { - select { - case <-timer.C: - pids, err := e.getAllPids() - if err != nil { - e.logger.Printf("[DEBUG] executor: error collecting pids: %v", err) - } - e.pidLock.Lock() - - // Adding pids which are not being tracked - for pid, np := range pids { - if _, ok := e.pids[pid]; !ok { - e.pids[pid] = np - } - } - // Removing pids which are no longer present - for pid := range e.pids { - if _, ok := pids[pid]; !ok { - delete(e.pids, pid) - } - } - e.pidLock.Unlock() - timer.Reset(pidScanInterval) - case <-e.processExited: - return - } - } -} - -// scanPids scans all the pids on the machine running the current executor and -// returns the child processes of the executor. -func (e *UniversalExecutor) scanPids(parentPid int, allPids []ps.Process) (map[int]*nomadPid, error) { - processFamily := make(map[int]struct{}) - processFamily[parentPid] = struct{}{} - - // A mapping of pids to their parent pids. It is used to build the process - // tree of the executing task - pidsRemaining := make(map[int]int, len(allPids)) - for _, pid := range allPids { - pidsRemaining[pid.Pid()] = pid.PPid() - } - - for { - // flag to indicate if we have found a match - foundNewPid := false - - for pid, ppid := range pidsRemaining { - _, childPid := processFamily[ppid] - - // checking if the pid is a child of any of the parents - if childPid { - processFamily[pid] = struct{}{} - delete(pidsRemaining, pid) - foundNewPid = true - } - } - - // not scanning anymore if we couldn't find a single match - if !foundNewPid { - break - } - } - - res := make(map[int]*nomadPid) - for pid := range processFamily { - np := nomadPid{ - pid: pid, - cpuStatsTotal: stats.NewCpuStats(), - cpuStatsUser: stats.NewCpuStats(), - cpuStatsSys: stats.NewCpuStats(), - } - res[pid] = &np - } - return res, nil -} - -// aggregatedResourceUsage aggregates the resource usage of all the pids and -// returns a TaskResourceUsage data point -func (e *UniversalExecutor) aggregatedResourceUsage(pidStats map[string]*cstructs.ResourceUsage) *cstructs.TaskResourceUsage { - ts := time.Now().UTC().UnixNano() - var ( - systemModeCPU, userModeCPU, percent float64 - totalRSS, totalSwap uint64 - ) - - for _, pidStat := range pidStats { - systemModeCPU += pidStat.CpuStats.SystemMode - userModeCPU += pidStat.CpuStats.UserMode - percent += pidStat.CpuStats.Percent - - totalRSS += pidStat.MemoryStats.RSS - totalSwap += pidStat.MemoryStats.Swap - } - - totalCPU := &cstructs.CpuStats{ - SystemMode: systemModeCPU, - UserMode: userModeCPU, - Percent: percent, - Measured: ExecutorBasicMeasuredCpuStats, - TotalTicks: e.systemCpuStats.TicksConsumed(percent), - } - - totalMemory := &cstructs.MemoryStats{ - RSS: totalRSS, - Swap: totalSwap, - Measured: ExecutorBasicMeasuredMemStats, - } - - resourceUsage := cstructs.ResourceUsage{ - MemoryStats: totalMemory, - CpuStats: totalCPU, - } - return &cstructs.TaskResourceUsage{ - ResourceUsage: &resourceUsage, - Timestamp: ts, - Pids: pidStats, - } -} - // Signal sends the passed signal to the task func (e *UniversalExecutor) Signal(s os.Signal) error { - if e.cmd.Process == nil { + if e.childCmd.Process == nil { return fmt.Errorf("Task not yet run") } - e.logger.Printf("[DEBUG] executor: sending signal %s to PID %d", s, e.cmd.Process.Pid) - err := e.cmd.Process.Signal(s) + e.logger.Debug("sending signal to PID", "signal", s, "pid", e.childCmd.Process.Pid) + err := e.childCmd.Process.Signal(s) if err != nil { - e.logger.Printf("[ERR] executor: sending signal %v failed: %v", s, err) + e.logger.Error("sending signal failed", "signal", s, "error", err) return err } return nil } -func (e *UniversalExecutor) LaunchSyslogServer() (*SyslogServerState, error) { - // Ensure the context has been set first - if e.ctx == nil { - return nil, fmt.Errorf("SetContext must be called before launching the Syslog Server") - } - - e.syslogChan = make(chan *logging.SyslogMessage, 2048) - l, err := e.getListener(e.ctx.PortLowerBound, e.ctx.PortUpperBound) +func (e *UniversalExecutor) Stats() (*cstructs.TaskResourceUsage, error) { + pidStats, err := e.pidCollector.pidStats() if err != nil { return nil, err } - e.logger.Printf("[DEBUG] syslog-server: launching syslog server on addr: %v", l.Addr().String()) - if err := e.configureLoggers(); err != nil { - return nil, err - } - - e.syslogServer = logging.NewSyslogServer(l, e.syslogChan, e.logger) - go e.syslogServer.Start() - go e.collectLogs(e.lre.rotatorWriter, e.lro.rotatorWriter) - syslogAddr := fmt.Sprintf("%s://%s", l.Addr().Network(), l.Addr().String()) - return &SyslogServerState{Addr: syslogAddr}, nil -} - -func (e *UniversalExecutor) collectLogs(we io.Writer, wo io.Writer) { - for logParts := range e.syslogChan { - // If the severity of the log line is err then we write to stderr - // otherwise all messages go to stdout - if logParts.Severity == syslog.LOG_ERR { - we.Write(logParts.Message) - we.Write([]byte{'\n'}) - } else { - wo.Write(logParts.Message) - wo.Write([]byte{'\n'}) - } - } -} - -// logRotatorWrapper wraps our log rotator and exposes a pipe that can feed the -// log rotator data. The processOutWriter should be attached to the process and -// data will be copied from the reader to the rotator. -type logRotatorWrapper struct { - processOutWriter *os.File - processOutReader *os.File - rotatorWriter *logging.FileRotator - hasFinishedCopied chan struct{} - logger *log.Logger -} - -// newLogRotatorWrapper takes a rotator and returns a wrapper that has the -// processOutWriter to attach to the processes stdout or stderr. -func newLogRotatorWrapper(logger *log.Logger, rotator *logging.FileRotator) (*logRotatorWrapper, error) { - r, w, err := os.Pipe() - if err != nil { - return nil, fmt.Errorf("failed to create os.Pipe for extracting logs: %v", err) - } - - wrap := &logRotatorWrapper{ - processOutWriter: w, - processOutReader: r, - rotatorWriter: rotator, - hasFinishedCopied: make(chan struct{}), - logger: logger, - } - wrap.start() - return wrap, nil -} - -// start starts a go-routine that copies from the pipe into the rotator. This is -// called by the constructor and not the user of the wrapper. -func (l *logRotatorWrapper) start() { - go func() { - defer close(l.hasFinishedCopied) - _, err := io.Copy(l.rotatorWriter, l.processOutReader) - if err != nil { - // Close reader to propagate io error across pipe. - // Note that this may block until the process exits on - // Windows due to - // https://github.com/PowerShell/PowerShell/issues/4254 - // or similar issues. Since this is already running in - // a goroutine its safe to block until the process is - // force-killed. - l.processOutReader.Close() - } - }() - return -} - -// Close closes the rotator and the process writer to ensure that the Wait -// command exits. -func (l *logRotatorWrapper) Close() { - // Wait up to the close tolerance before we force close - select { - case <-l.hasFinishedCopied: - case <-time.After(processOutputCloseTolerance): - } - - // Closing the read side of a pipe may block on Windows if the process - // is being debugged as in: - // https://github.com/PowerShell/PowerShell/issues/4254 - // The pipe will be closed and cleaned up when the process exits. - closeDone := make(chan struct{}) - go func() { - defer close(closeDone) - err := l.processOutReader.Close() - if err != nil && !strings.Contains(err.Error(), "file already closed") { - l.logger.Printf("[WARN] executor: error closing read-side of process output pipe: %v", err) - } - - }() - - select { - case <-closeDone: - case <-time.After(processOutputCloseTolerance): - l.logger.Printf("[WARN] executor: timed out waiting for read-side of process output pipe to close") - } - - l.rotatorWriter.Close() - return + return aggregatedResourceUsage(e.systemCpuStats, pidStats), nil } diff --git a/client/driver/executor/executor_basic.go b/client/driver/executor/executor_basic.go index 123ed4703..1951ae584 100644 --- a/client/driver/executor/executor_basic.go +++ b/client/driver/executor/executor_basic.go @@ -1,46 +1,15 @@ -// +build darwin dragonfly freebsd netbsd openbsd solaris windows +// +build !linux package executor -import ( - "os" +import hclog "github.com/hashicorp/go-hclog" - cstructs "github.com/hashicorp/nomad/client/structs" - "github.com/mitchellh/go-ps" -) - -func (e *UniversalExecutor) configureChroot() error { - return nil +func NewExecutorWithIsolation(logger hclog.Logger) Executor { + logger = logger.Named("executor") + logger.Error("isolation executor is not supported on this platform, using default") + return NewExecutor(logger) } -func (e *UniversalExecutor) removeChrootMounts() error { - return nil -} +func (e *UniversalExecutor) configureResourceContainer(_ int) error { return nil } -func (e *UniversalExecutor) runAs(userid string) error { - return nil -} - -func (e *UniversalExecutor) applyLimits(pid int) error { - return nil -} - -func (e *UniversalExecutor) configureIsolation() error { - return nil -} - -func (e *UniversalExecutor) Stats() (*cstructs.TaskResourceUsage, error) { - pidStats, err := e.pidStats() - if err != nil { - return nil, err - } - return e.aggregatedResourceUsage(pidStats), nil -} - -func (e *UniversalExecutor) getAllPids() (map[int]*nomadPid, error) { - allProcesses, err := ps.Processes() - if err != nil { - return nil, err - } - return e.scanPids(os.Getpid(), allProcesses) -} +func (e *UniversalExecutor) runAs(_ string) error { return nil } diff --git a/client/driver/executor/executor_linux.go b/client/driver/executor/executor_linux.go index 7a94d4b69..dfc2db010 100644 --- a/client/driver/executor/executor_linux.go +++ b/client/driver/executor/executor_linux.go @@ -1,189 +1,333 @@ +// +build linux + package executor import ( + "context" "fmt" "os" - "os/user" + "os/exec" + "path" "path/filepath" - "strconv" "strings" "syscall" "time" - "github.com/hashicorp/go-multierror" - "github.com/mitchellh/go-ps" - cgroupFs "github.com/opencontainers/runc/libcontainer/cgroups/fs" - cgroupConfig "github.com/opencontainers/runc/libcontainer/configs" - + "github.com/armon/circbuf" + "github.com/hashicorp/consul-template/signals" + hclog "github.com/hashicorp/go-hclog" + multierror "github.com/hashicorp/go-multierror" + dstructs "github.com/hashicorp/nomad/client/driver/structs" "github.com/hashicorp/nomad/client/stats" cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/helper/discover" + shelpers "github.com/hashicorp/nomad/helper/stats" "github.com/hashicorp/nomad/helper/uuid" - "github.com/hashicorp/nomad/nomad/structs" + "github.com/opencontainers/runc/libcontainer" + "github.com/opencontainers/runc/libcontainer/cgroups" + cgroupFs "github.com/opencontainers/runc/libcontainer/cgroups/fs" + lconfigs "github.com/opencontainers/runc/libcontainer/configs" + + "github.com/syndtr/gocapability/capability" +) + +const ( + defaultCgroupParent = "nomad" ) var ( // The statistics the executor exposes when using cgroups ExecutorCgroupMeasuredMemStats = []string{"RSS", "Cache", "Swap", "Max Usage", "Kernel Usage", "Kernel Max Usage"} ExecutorCgroupMeasuredCpuStats = []string{"System Mode", "User Mode", "Throttled Periods", "Throttled Time", "Percent"} + + // allCaps is all linux capabilities which is used to configure libcontainer + allCaps []string ) -// configureIsolation configures chroot and creates cgroups -func (e *UniversalExecutor) configureIsolation() error { - if e.command.FSIsolation { - if err := e.configureChroot(); err != nil { - return err - } +// initialize the allCaps var with all capabilities available on the system +func init() { + last := capability.CAP_LAST_CAP + // workaround for RHEL6 which has no /proc/sys/kernel/cap_last_cap + if last == capability.Cap(63) { + last = capability.CAP_BLOCK_SUSPEND } - - if e.command.ResourceLimits || e.command.BasicProcessCgroup { - if err := e.configureCgroups(e.ctx.Task.Resources); err != nil { - return fmt.Errorf("error creating cgroups: %v", err) + for _, cap := range capability.List() { + if cap > last { + continue } + allCaps = append(allCaps, fmt.Sprintf("CAP_%s", strings.ToUpper(cap.String()))) } - return nil } -// applyLimits puts a process in a pre-configured cgroup -func (e *UniversalExecutor) applyLimits(pid int) error { - if !(e.command.ResourceLimits || e.command.BasicProcessCgroup) { - return nil - } +// LibcontainerExecutor implements an Executor with the runc/libcontainer api +type LibcontainerExecutor struct { + id string + command *ExecCommand - // Entering the process in the cgroup - manager := getCgroupManager(e.resConCtx.groups, nil) - if err := manager.Apply(pid); err != nil { - e.logger.Printf("[ERR] executor: error applying pid to cgroup: %v", err) - return err - } + logger hclog.Logger - e.resConCtx.cgPaths = manager.GetPaths() + totalCpuStats *stats.CpuStats + userCpuStats *stats.CpuStats + systemCpuStats *stats.CpuStats + pidCollector *pidCollector - // Don't enter all the cgroups since we will inherit resources limits. Only - // use devices (required by libcontainer) and freezer. Freezer allows us to - // capture all pids and stop any fork/execs from happening while we are - // cleaning up. - if !e.command.ResourceLimits { - // Move the executor into the global cgroup so that the task specific - // cgroup can be destroyed. - nilGroup := &cgroupConfig.Cgroup{} - nilGroup.Path = "/" - nilGroup.Resources = e.resConCtx.groups.Resources - nilManager := getCgroupManager(nilGroup, nil) - err := nilManager.Apply(pid) - if err != nil { - return fmt.Errorf("failed to remove executor pid %d: %v", pid, err) - } - - // Grab the freezer and devices cgroup paths. We do this from the old - // manager after the executor pid has been applied since there is no - // other way to determine what the proper cgroup paths would be. - freezer := &cgroupFs.FreezerGroup{} - devices := &cgroupFs.DevicesGroup{} - freezerName, devicesName := freezer.Name(), devices.Name() - newPath := map[string]string{ - freezerName: e.resConCtx.cgPaths[freezerName], - devicesName: e.resConCtx.cgPaths[devicesName], - } - - // Clear the cgroups paths so that everything is properly cleaned except - // the groups we want our process to stay in. This will delete the - // directories from disk. - manager.Cgroups.Paths = nil - delete(manager.Paths, freezerName) - delete(manager.Paths, devicesName) - if err := manager.Destroy(); err != nil { - e.logger.Printf("[ERR] executor: failed to destroy original: %v", err) - return err - } - - // Update our context such that the new cgroup manager only is tracking - // the paths we care about now. - e.resConCtx.cgPaths = newPath - e.resConCtx.groups.Paths = newPath - - // Apply just the freezer and devices now - manager = getCgroupManager(e.resConCtx.groups, e.resConCtx.cgPaths) - if err := manager.Apply(pid); err != nil { - e.logger.Printf("[ERR] executor: error applying pid to cgroup subset %v: %v", e.resConCtx.cgPaths, err) - return err - } - } - - cgConfig := cgroupConfig.Config{Cgroups: e.resConCtx.groups} - if err := manager.Set(&cgConfig); err != nil { - e.logger.Printf("[ERR] executor: error setting cgroup config: %v", err) - if er := DestroyCgroup(e.resConCtx.groups, e.resConCtx.cgPaths, os.Getpid()); er != nil { - e.logger.Printf("[ERR] executor: error destroying cgroup: %v", er) - } - return err - } - return nil + container libcontainer.Container + userProc *libcontainer.Process + userProcExited chan interface{} + exitState *ProcessState } -// configureCgroups converts a Nomad Resources specification into the equivalent -// cgroup configuration. It returns an error if the resources are invalid. -func (e *UniversalExecutor) configureCgroups(resources *structs.Resources) error { - e.resConCtx.groups = &cgroupConfig.Cgroup{} - e.resConCtx.groups.Resources = &cgroupConfig.Resources{} - cgroupName := uuid.Generate() - e.resConCtx.groups.Path = filepath.Join("/nomad", cgroupName) - - // Allow access to /dev/ - allowAllDevs := true - e.resConCtx.groups.Resources.AllowAllDevices = &allowAllDevs - - // Use a cgroup but don't apply limits - if !e.command.ResourceLimits { - return nil +func NewExecutorWithIsolation(logger hclog.Logger) Executor { + logger = logger.Named("isolated_executor") + if err := shelpers.Init(); err != nil { + logger.Error("unable to initialize stats", "error", err) } - - if resources.MemoryMB > 0 { - // Total amount of memory allowed to consume - e.resConCtx.groups.Resources.Memory = int64(resources.MemoryMB * 1024 * 1024) - // Disable swap to avoid issues on the machine - var memSwappiness uint64 = 0 - e.resConCtx.groups.Resources.MemorySwappiness = &memSwappiness + return &LibcontainerExecutor{ + id: strings.Replace(uuid.Generate(), "-", "_", 0), + logger: logger, + totalCpuStats: stats.NewCpuStats(), + userCpuStats: stats.NewCpuStats(), + systemCpuStats: stats.NewCpuStats(), + pidCollector: newPidCollector(logger), } - - if resources.CPU < 2 { - return fmt.Errorf("resources.CPU must be equal to or greater than 2: %v", resources.CPU) - } - - // Set the relative CPU shares for this cgroup. - e.resConCtx.groups.Resources.CpuShares = uint64(resources.CPU) - - if resources.IOPS != 0 { - // Validate it is in an acceptable range. - if resources.IOPS < 10 || resources.IOPS > 1000 { - return fmt.Errorf("resources.IOPS must be between 10 and 1000: %d", resources.IOPS) - } - - e.resConCtx.groups.Resources.BlkioWeight = uint16(resources.IOPS) - } - - return nil } -// Stats reports the resource utilization of the cgroup. If there is no resource -// isolation we aggregate the resource utilization of all the pids launched by -// the executor. -func (e *UniversalExecutor) Stats() (*cstructs.TaskResourceUsage, error) { - // If we don't use full resource limits fallback to normal collection. It is - // not enough to be in the Cgroup since you must be in the memory, cpu, and - // cpuacct cgroup to gather the correct statistics. - if !e.command.ResourceLimits { - pidStats, err := e.pidStats() - if err != nil { - return nil, err - } - return e.aggregatedResourceUsage(pidStats), nil +// Launch creates a new container in libcontainer and starts a new process with it +func (l *LibcontainerExecutor) Launch(command *ExecCommand) (*ProcessState, error) { + l.logger.Info("launching command", "command", command.Cmd, "args", strings.Join(command.Args, " ")) + // Find the nomad executable to launch the executor process with + bin, err := discover.NomadExecutable() + if err != nil { + return nil, fmt.Errorf("unable to find the nomad binary: %v", err) } - ts := time.Now() - manager := getCgroupManager(e.resConCtx.groups, e.resConCtx.cgPaths) - stats, err := manager.GetStats() + + l.command = command + + // Move to the root cgroup until process is started + subsystems, err := cgroups.GetAllSubsystems() if err != nil { return nil, err } + if err := JoinRootCgroup(subsystems); err != nil { + return nil, err + } + + // create a new factory which will store the container state in the allocDir + factory, err := libcontainer.New( + path.Join(command.TaskDir, "../alloc/container"), + libcontainer.Cgroupfs, + libcontainer.InitArgs(bin, "libcontainer-shim"), + ) + if err != nil { + return nil, fmt.Errorf("failed to create factory: %v", err) + } + + // A container groups processes under the same isolation enforcement + container, err := factory.Create(l.id, newLibcontainerConfig(command)) + if err != nil { + return nil, fmt.Errorf("failed to create container(%s): %v", l.id, err) + } + l.container = container + + combined := append([]string{command.Cmd}, command.Args...) + stdout, err := command.Stdout() + if err != nil { + return nil, err + } + stderr, err := command.Stderr() + if err != nil { + return nil, err + } + + // the task process will be started by the container + process := &libcontainer.Process{ + Args: combined, + Env: command.Env, + Stdout: stdout, + Stderr: stderr, + Init: true, + } + + if command.User != "" { + process.User = command.User + } + l.userProc = process + + l.totalCpuStats = stats.NewCpuStats() + l.userCpuStats = stats.NewCpuStats() + l.systemCpuStats = stats.NewCpuStats() + + // Starts the task + if err := container.Run(process); err != nil { + container.Destroy() + return nil, err + } + + pid, err := process.Pid() + if err != nil { + container.Destroy() + return nil, err + } + + // Join process cgroups + containerState, err := container.State() + if err != nil { + l.logger.Error("error entering user process cgroups", "executor_pid", os.Getpid(), "error", err) + } + if err := cgroups.EnterPid(containerState.CgroupPaths, os.Getpid()); err != nil { + l.logger.Error("error entering user process cgroups", "executor_pid", os.Getpid(), "error", err) + } + + // start a goroutine to wait on the process to complete, so Wait calls can + // be multiplexed + l.userProcExited = make(chan interface{}) + go l.pidCollector.collectPids(l.userProcExited, l.getAllPids) + go l.wait() + + return &ProcessState{ + Pid: pid, + ExitCode: -1, + Time: time.Now(), + }, nil +} + +func (l *LibcontainerExecutor) getAllPids() (map[int]*nomadPid, error) { + pids, err := l.container.Processes() + if err != nil { + return nil, err + } + nPids := make(map[int]*nomadPid) + for _, pid := range pids { + nPids[pid] = &nomadPid{ + pid: pid, + cpuStatsTotal: stats.NewCpuStats(), + cpuStatsUser: stats.NewCpuStats(), + cpuStatsSys: stats.NewCpuStats(), + } + } + return nPids, nil +} + +// Wait waits until a process has exited and returns it's exitcode and errors +func (l *LibcontainerExecutor) Wait() (*ProcessState, error) { + <-l.userProcExited + return l.exitState, nil +} + +func (l *LibcontainerExecutor) wait() { + defer close(l.userProcExited) + + ps, err := l.userProc.Wait() + if err != nil { + // If the process has exited before we called wait an error is returned + // the process state is embedded in the error + if exitErr, ok := err.(*exec.ExitError); ok { + ps = exitErr.ProcessState + } else { + l.logger.Error("failed to call wait on user process", "error", err) + l.exitState = &ProcessState{Pid: 0, ExitCode: 0, Time: time.Now()} + return + } + } + + l.command.Close() + + exitCode := 1 + var signal int + if status, ok := ps.Sys().(syscall.WaitStatus); ok { + exitCode = status.ExitStatus() + if status.Signaled() { + const exitSignalBase = 128 + signal = int(status.Signal()) + exitCode = exitSignalBase + signal + } + } + + l.exitState = &ProcessState{ + Pid: ps.Pid(), + ExitCode: exitCode, + Signal: signal, + Time: time.Now(), + } +} + +// Shutdown stops all processes started and cleans up any resources +// created (such as mountpoints, devices, etc). +func (l *LibcontainerExecutor) Shutdown(signal string, grace time.Duration) error { + if l.container == nil { + return nil + } + + // move executor to root cgroup + subsystems, err := cgroups.GetAllSubsystems() + if err != nil { + return err + } + if err := JoinRootCgroup(subsystems); err != nil { + return err + } + + status, err := l.container.Status() + if err != nil { + return err + } + + defer l.container.Destroy() + + if status == libcontainer.Stopped { + return nil + } + + if grace > 0 { + if signal == "" { + signal = "SIGINT" + } + + sig, ok := signals.SignalLookup[signal] + if !ok { + return fmt.Errorf("error unknown signal given for shutdown: %s", signal) + } + + err = l.container.Signal(sig, false) + if err != nil { + return err + } + + select { + case <-l.userProcExited: + return nil + case <-time.After(grace): + return l.container.Signal(os.Kill, false) + } + } else { + return l.container.Signal(os.Kill, false) + } +} + +// UpdateResources updates the resource isolation with new values to be enforced +func (l *LibcontainerExecutor) UpdateResources(resources *Resources) error { + return nil +} + +// Version returns the api version of the executor +func (l *LibcontainerExecutor) Version() (*ExecutorVersion, error) { + return &ExecutorVersion{Version: ExecutorVersionLatest}, nil +} + +// Stats returns the resource statistics for processes managed by the executor +func (l *LibcontainerExecutor) Stats() (*cstructs.TaskResourceUsage, error) { + lstats, err := l.container.Stats() + if err != nil { + return nil, err + } + + pidStats, err := l.pidCollector.pidStats() + if err != nil { + return nil, err + } + + ts := time.Now() + stats := lstats.CgroupStats // Memory Related Stats swap := stats.MemoryStats.SwapUsage @@ -205,14 +349,14 @@ func (e *UniversalExecutor) Stats() (*cstructs.TaskResourceUsage, error) { userModeTime := float64(stats.CpuStats.CpuUsage.UsageInUsermode) kernelModeTime := float64(stats.CpuStats.CpuUsage.UsageInKernelmode) - totalPercent := e.totalCpuStats.Percent(totalProcessCPUUsage) + totalPercent := l.totalCpuStats.Percent(totalProcessCPUUsage) cs := &cstructs.CpuStats{ - SystemMode: e.systemCpuStats.Percent(kernelModeTime), - UserMode: e.userCpuStats.Percent(userModeTime), + SystemMode: l.systemCpuStats.Percent(kernelModeTime), + UserMode: l.userCpuStats.Percent(userModeTime), Percent: totalPercent, ThrottledPeriods: stats.CpuStats.ThrottlingData.ThrottledPeriods, ThrottledTime: stats.CpuStats.ThrottlingData.ThrottledTime, - TotalTicks: e.systemCpuStats.TicksConsumed(totalPercent), + TotalTicks: l.systemCpuStats.TicksConsumed(totalPercent), Measured: ExecutorCgroupMeasuredCpuStats, } taskResUsage := cstructs.TaskResourceUsage{ @@ -221,181 +365,246 @@ func (e *UniversalExecutor) Stats() (*cstructs.TaskResourceUsage, error) { CpuStats: cs, }, Timestamp: ts.UTC().UnixNano(), + Pids: pidStats, } - if pidStats, err := e.pidStats(); err == nil { - taskResUsage.Pids = pidStats - } + return &taskResUsage, nil } -// runAs takes a user id as a string and looks up the user, and sets the command -// to execute as that user. -func (e *UniversalExecutor) runAs(userid string) error { - u, err := user.Lookup(userid) - if err != nil { - return fmt.Errorf("Failed to identify user %v: %v", userid, err) - } - - // Get the groups the user is a part of - gidStrings, err := u.GroupIds() - if err != nil { - return fmt.Errorf("Unable to lookup user's group membership: %v", err) - } - - gids := make([]uint32, len(gidStrings)) - for _, gidString := range gidStrings { - u, err := strconv.Atoi(gidString) - if err != nil { - return fmt.Errorf("Unable to convert user's group to int %s: %v", gidString, err) - } - - gids = append(gids, uint32(u)) - } - - // Convert the uid and gid - uid, err := strconv.ParseUint(u.Uid, 10, 32) - if err != nil { - return fmt.Errorf("Unable to convert userid to uint32: %s", err) - } - gid, err := strconv.ParseUint(u.Gid, 10, 32) - if err != nil { - return fmt.Errorf("Unable to convert groupid to uint32: %s", err) - } - - // Set the command to run as that user and group. - if e.cmd.SysProcAttr == nil { - e.cmd.SysProcAttr = &syscall.SysProcAttr{} - } - if e.cmd.SysProcAttr.Credential == nil { - e.cmd.SysProcAttr.Credential = &syscall.Credential{} - } - e.cmd.SysProcAttr.Credential.Uid = uint32(uid) - e.cmd.SysProcAttr.Credential.Gid = uint32(gid) - e.cmd.SysProcAttr.Credential.Groups = gids - - e.logger.Printf("[DEBUG] executor: running as user:group %d:%d with group membership in %v", uid, gid, gids) - - return nil +// Signal sends a signal to the process managed by the executor +func (l *LibcontainerExecutor) Signal(s os.Signal) error { + return l.userProc.Signal(s) } -// configureChroot configures a chroot -func (e *UniversalExecutor) configureChroot() error { - if e.cmd.SysProcAttr == nil { - e.cmd.SysProcAttr = &syscall.SysProcAttr{} +// Exec starts an additional process inside the container +func (l *LibcontainerExecutor) Exec(deadline time.Time, cmd string, args []string) ([]byte, int, error) { + combined := append([]string{cmd}, args...) + // Capture output + buf, _ := circbuf.NewBuffer(int64(dstructs.CheckBufSize)) + + process := &libcontainer.Process{ + Args: combined, + Env: l.command.Env, + Stdout: buf, + Stderr: buf, } - e.cmd.SysProcAttr.Chroot = e.ctx.TaskDir - e.cmd.Dir = "/" - e.fsIsolationEnforced = true - return nil -} + err := l.container.Run(process) + if err != nil { + return nil, 0, err + } -// getAllPids returns the pids of all the processes spun up by the executor. We -// use the libcontainer apis to get the pids when the user is using cgroup -// isolation and we scan the entire process table if the user is not using any -// isolation -func (e *UniversalExecutor) getAllPids() (map[int]*nomadPid, error) { - if e.command.ResourceLimits || e.command.BasicProcessCgroup { - manager := getCgroupManager(e.resConCtx.groups, e.resConCtx.cgPaths) - pids, err := manager.GetAllPids() - if err != nil { - return nil, err - } - np := make(map[int]*nomadPid, len(pids)) - for _, pid := range pids { - np[pid] = &nomadPid{ - pid: pid, - cpuStatsTotal: stats.NewCpuStats(), - cpuStatsSys: stats.NewCpuStats(), - cpuStatsUser: stats.NewCpuStats(), + waitCh := make(chan *waitResult) + defer close(waitCh) + go l.handleExecWait(waitCh, process) + + select { + case result := <-waitCh: + ps := result.ps + if result.err != nil { + if exitErr, ok := result.err.(*exec.ExitError); ok { + ps = exitErr.ProcessState + } else { + return nil, 0, result.err } } - return np, nil + var exitCode int + if status, ok := ps.Sys().(syscall.WaitStatus); ok { + exitCode = status.ExitStatus() + } + return buf.Bytes(), exitCode, nil + + case <-time.After(time.Until(deadline)): + process.Signal(os.Kill) + return nil, 0, context.DeadlineExceeded } - allProcesses, err := ps.Processes() - if err != nil { - return nil, err - } - return e.scanPids(os.Getpid(), allProcesses) + } -// destroyCgroup kills all processes in the cgroup and removes the cgroup -// configuration from the host. This function is idempotent. -func DestroyCgroup(groups *cgroupConfig.Cgroup, cgPaths map[string]string, executorPid int) error { - mErrs := new(multierror.Error) - if groups == nil { - return fmt.Errorf("Can't destroy: cgroup configuration empty") +type waitResult struct { + ps *os.ProcessState + err error +} + +func (l *LibcontainerExecutor) handleExecWait(ch chan *waitResult, process *libcontainer.Process) { + ps, err := process.Wait() + ch <- &waitResult{ps, err} +} + +func configureCapabilities(cfg *lconfigs.Config, command *ExecCommand) { + // TODO: allow better control of these + cfg.Capabilities = &lconfigs.Capabilities{ + Bounding: allCaps, + Permitted: allCaps, + Inheritable: allCaps, + Ambient: allCaps, + Effective: allCaps, } - // Move the executor into the global cgroup so that the task specific - // cgroup can be destroyed. - nilGroup := &cgroupConfig.Cgroup{} - nilGroup.Path = "/" - nilGroup.Resources = groups.Resources - nilManager := getCgroupManager(nilGroup, nil) - err := nilManager.Apply(executorPid) - if err != nil && !strings.Contains(err.Error(), "no such process") { - return fmt.Errorf("failed to remove executor pid %d: %v", executorPid, err) +} + +func configureIsolation(cfg *lconfigs.Config, command *ExecCommand) { + defaultMountFlags := syscall.MS_NOEXEC | syscall.MS_NOSUID | syscall.MS_NODEV + + // set the new root directory for the container + cfg.Rootfs = command.TaskDir + + // launch with mount namespace + cfg.Namespaces = lconfigs.Namespaces{ + {Type: lconfigs.NEWNS}, } - // Freeze the Cgroup so that it can not continue to fork/exec. - manager := getCgroupManager(groups, cgPaths) - err = manager.Freeze(cgroupConfig.Frozen) - if err != nil && !strings.Contains(err.Error(), "no such file or directory") { - return fmt.Errorf("failed to freeze cgroup: %v", err) + // paths to mask using a bind mount to /dev/null to prevent reading + cfg.MaskPaths = []string{ + "/proc/kcore", + "/sys/firmware", } - var procs []*os.Process - pids, err := manager.GetAllPids() - if err != nil { - multierror.Append(mErrs, fmt.Errorf("error getting pids: %v", err)) + // paths that should be remounted as readonly inside the container + cfg.ReadonlyPaths = []string{ + "/proc/sys", "/proc/sysrq-trigger", "/proc/irq", "/proc/bus", + } - // Unfreeze the cgroup. - err = manager.Freeze(cgroupConfig.Thawed) - if err != nil && !strings.Contains(err.Error(), "no such file or directory") { - multierror.Append(mErrs, fmt.Errorf("failed to unfreeze cgroup: %v", err)) + cfg.Devices = lconfigs.DefaultAutoCreatedDevices + cfg.Mounts = []*lconfigs.Mount{ + { + Source: "tmpfs", + Destination: "/dev", + Device: "tmpfs", + Flags: syscall.MS_NOSUID | syscall.MS_STRICTATIME, + Data: "mode=755", + }, + { + Source: "devpts", + Destination: "/dev/pts", + Device: "devpts", + Flags: syscall.MS_NOSUID | syscall.MS_NOEXEC, + Data: "newinstance,ptmxmode=0666,mode=0620,gid=5", + }, + { + Device: "tmpfs", + Source: "shm", + Destination: "/dev/shm", + Data: "mode=1777,size=65536k", + Flags: defaultMountFlags, + }, + { + Source: "mqueue", + Destination: "/dev/mqueue", + Device: "mqueue", + Flags: defaultMountFlags, + }, + { + Source: "sysfs", + Destination: "/sys", + Device: "sysfs", + Flags: defaultMountFlags | syscall.MS_RDONLY, + }, + } +} + +func configureCgroups(cfg *lconfigs.Config, command *ExecCommand) error { + + // If resources are not limited then manually create cgroups needed + if !command.ResourceLimits { + return configureBasicCgroups(cfg) + } + + id := uuid.Generate() + cfg.Cgroups.Path = filepath.Join(defaultCgroupParent, id) + if command.Resources.MemoryMB > 0 { + // Total amount of memory allowed to consume + cfg.Cgroups.Resources.Memory = int64(command.Resources.MemoryMB * 1024 * 1024) + // Disable swap to avoid issues on the machine + var memSwappiness uint64 = 0 + cfg.Cgroups.Resources.MemorySwappiness = &memSwappiness + } + + if command.Resources.CPU < 2 { + return fmt.Errorf("resources.CPU must be equal to or greater than 2: %v", command.Resources.CPU) + } + + // Set the relative CPU shares for this cgroup. + cfg.Cgroups.Resources.CpuShares = uint64(command.Resources.CPU) + + if command.Resources.IOPS != 0 { + // Validate it is in an acceptable range. + if command.Resources.IOPS < 10 || command.Resources.IOPS > 1000 { + return fmt.Errorf("resources.IOPS must be between 10 and 1000: %d", command.Resources.IOPS) } - return mErrs.ErrorOrNil() + + cfg.Cgroups.Resources.BlkioWeight = uint16(command.Resources.IOPS) + } + return nil +} + +func configureBasicCgroups(cfg *lconfigs.Config) error { + id := uuid.Generate() + + // Manually create freezer cgroup + cfg.Cgroups.Paths = map[string]string{} + root, err := cgroups.FindCgroupMountpointDir() + if err != nil { + return err } - // Kill the processes in the cgroup - for _, pid := range pids { - proc, err := os.FindProcess(pid) + if _, err := os.Stat(root); err != nil { + return err + } + + freezer := cgroupFs.FreezerGroup{} + subsystem := freezer.Name() + path, err := cgroups.FindCgroupMountpoint(subsystem) + if err != nil { + return fmt.Errorf("failed to find %s cgroup mountpoint: %v", subsystem, err) + } + // Sometimes subsystems can be mounted together as 'cpu,cpuacct'. + path = filepath.Join(root, filepath.Base(path), defaultCgroupParent, id) + + if err = os.MkdirAll(path, 0755); err != nil { + return err + } + + cfg.Cgroups.Paths[subsystem] = path + return nil +} + +func newLibcontainerConfig(command *ExecCommand) *lconfigs.Config { + cfg := &lconfigs.Config{ + Cgroups: &lconfigs.Cgroup{ + Resources: &lconfigs.Resources{ + AllowAllDevices: nil, + MemorySwappiness: nil, + AllowedDevices: lconfigs.DefaultAllowedDevices, + }, + }, + Version: "1.0.0", + } + + configureCapabilities(cfg, command) + configureIsolation(cfg, command) + configureCgroups(cfg, command) + return cfg +} + +// JoinRootCgroup moves the current process to the cgroups of the init process +func JoinRootCgroup(subsystems []string) error { + mErrs := new(multierror.Error) + paths := map[string]string{} + for _, s := range subsystems { + mnt, _, err := cgroups.FindCgroupMountpointAndRoot(s) if err != nil { - multierror.Append(mErrs, fmt.Errorf("error finding process %v: %v", pid, err)) + multierror.Append(mErrs, fmt.Errorf("error getting cgroup path for subsystem: %s", s)) continue } - procs = append(procs, proc) - if e := proc.Kill(); e != nil { - multierror.Append(mErrs, fmt.Errorf("error killing process %v: %v", pid, e)) - } + paths[s] = mnt } - // Unfreeze the cgroug so we can wait. - err = manager.Freeze(cgroupConfig.Thawed) - if err != nil && !strings.Contains(err.Error(), "no such file or directory") { - multierror.Append(mErrs, fmt.Errorf("failed to unfreeze cgroup: %v", err)) + err := cgroups.EnterPid(paths, os.Getpid()) + if err != nil { + multierror.Append(mErrs, err) } - // Wait on the killed processes to ensure they are cleaned up. - for _, proc := range procs { - // Don't capture the error because we expect this to fail for - // processes we didn't fork. - proc.Wait() - } - - // Clear the cgroups paths so that everything is properly cleaned - manager.Cgroups.Paths = nil - - // Remove the cgroup. - if err := manager.Destroy(); err != nil { - multierror.Append(mErrs, fmt.Errorf("failed to delete the cgroup directories: %v", err)) - } return mErrs.ErrorOrNil() } - -// getCgroupManager returns the correct libcontainer cgroup manager. -func getCgroupManager(groups *cgroupConfig.Cgroup, paths map[string]string) *cgroupFs.Manager { - return &cgroupFs.Manager{Cgroups: groups, Paths: paths} -} diff --git a/client/driver/executor/executor_linux_test.go b/client/driver/executor/executor_linux_test.go index 3d423ddd0..d65c33df6 100644 --- a/client/driver/executor/executor_linux_test.go +++ b/client/driver/executor/executor_linux_test.go @@ -1,6 +1,7 @@ package executor import ( + "fmt" "io/ioutil" "os" "path/filepath" @@ -9,20 +10,30 @@ import ( "testing" "time" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/driver/env" - dstructs "github.com/hashicorp/nomad/client/driver/structs" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/client/testutil" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/nomad/mock" + tu "github.com/hashicorp/nomad/testutil" + "github.com/stretchr/testify/require" ) +func init() { + executorFactories["LibcontainerExecutor"] = libcontainerFactory +} + +func libcontainerFactory(l hclog.Logger) Executor { + return NewExecutorWithIsolation(l) +} + // testExecutorContextWithChroot returns an ExecutorContext and AllocDir with // chroot. Use testExecutorContext if you don't need a chroot. // // The caller is responsible for calling AllocDir.Destroy() to cleanup. -func testExecutorContextWithChroot(t *testing.T) (*ExecutorContext, *allocdir.AllocDir) { +func testExecutorCommandWithChroot(t *testing.T) (*ExecCommand, *allocdir.AllocDir) { chrootEnv := map[string]string{ "/etc/ld.so.cache": "/etc/ld.so.cache", "/etc/ld.so.conf": "/etc/ld.so.conf", @@ -41,7 +52,7 @@ func testExecutorContextWithChroot(t *testing.T) (*ExecutorContext, *allocdir.Al task := alloc.Job.TaskGroups[0].Tasks[0] taskEnv := env.NewBuilder(mock.Node(), alloc, task, "global").Build() - allocDir := allocdir.NewAllocDir(testlog.Logger(t), filepath.Join(os.TempDir(), alloc.ID)) + allocDir := allocdir.NewAllocDir(testlog.HCLogger(t), filepath.Join(os.TempDir(), alloc.ID)) if err := allocDir.Build(); err != nil { t.Fatalf("AllocDir.Build() failed: %v", err) } @@ -50,69 +61,69 @@ func testExecutorContextWithChroot(t *testing.T) (*ExecutorContext, *allocdir.Al t.Fatalf("allocDir.NewTaskDir(%q) failed: %v", task.Name, err) } td := allocDir.TaskDirs[task.Name] - ctx := &ExecutorContext{ - TaskEnv: taskEnv, - Task: task, + cmd := &ExecCommand{ + Env: taskEnv.List(), TaskDir: td.Dir, - LogDir: td.LogDir, + Resources: &Resources{ + CPU: task.Resources.CPU, + MemoryMB: task.Resources.MemoryMB, + IOPS: task.Resources.IOPS, + DiskMB: task.Resources.DiskMB, + }, } - return ctx, allocDir + configureTLogging(cmd) + + return cmd, allocDir } func TestExecutor_IsolationAndConstraints(t *testing.T) { t.Parallel() + require := require.New(t) testutil.ExecCompatible(t) - execCmd := ExecCommand{Cmd: "/bin/ls", Args: []string{"-F", "/", "/etc/"}} - ctx, allocDir := testExecutorContextWithChroot(t) + execCmd, allocDir := testExecutorCommandWithChroot(t) + execCmd.Cmd = "/bin/ls" + execCmd.Args = []string{"-F", "/", "/etc/"} defer allocDir.Destroy() - execCmd.FSIsolation = true execCmd.ResourceLimits = true - execCmd.User = dstructs.DefaultUnprivilegedUser - executor := NewExecutor(testlog.Logger(t)) + executor := libcontainerFactory(testlog.HCLogger(t)) + defer executor.Shutdown("SIGKILL", 0) - if err := executor.SetContext(ctx); err != nil { - t.Fatalf("Unexpected error: %v", err) - } + ps, err := executor.Launch(execCmd) + require.NoError(err) + require.NotZero(ps.Pid) - ps, err := executor.LaunchCmd(&execCmd) - if err != nil { - t.Fatalf("error in launching command: %v", err) - } - if ps.Pid == 0 { - t.Fatalf("expected process to start and have non zero pid") - } state, err := executor.Wait() - if err != nil { - t.Fatalf("error in waiting for command: %v", err) - } - if state.ExitCode != 0 { - t.Errorf("exited with non-zero code: %v", state.ExitCode) - } + require.NoError(err) + require.Zero(state.ExitCode) // Check if the resource constraints were applied - memLimits := filepath.Join(ps.IsolationConfig.CgroupPaths["memory"], "memory.limit_in_bytes") - data, err := ioutil.ReadFile(memLimits) - if err != nil { - t.Fatalf("err: %v", err) - } - expectedMemLim := strconv.Itoa(ctx.Task.Resources.MemoryMB * 1024 * 1024) - actualMemLim := strings.TrimSpace(string(data)) - if actualMemLim != expectedMemLim { - t.Fatalf("actual mem limit: %v, expected: %v", string(data), expectedMemLim) - } + if lexec, ok := executor.(*LibcontainerExecutor); ok { + state, err := lexec.container.State() + require.NoError(err) - if err := executor.Exit(); err != nil { - t.Fatalf("error: %v", err) - } + memLimits := filepath.Join(state.CgroupPaths["memory"], "memory.limit_in_bytes") + data, err := ioutil.ReadFile(memLimits) + require.NoError(err) - // Check if Nomad has actually removed the cgroups - if _, err := os.Stat(memLimits); err == nil { - t.Fatalf("file %v hasn't been removed", memLimits) - } + expectedMemLim := strconv.Itoa(execCmd.Resources.MemoryMB * 1024 * 1024) + actualMemLim := strings.TrimSpace(string(data)) + require.Equal(actualMemLim, expectedMemLim) + require.NoError(executor.Shutdown("", 0)) + executor.Wait() + // Check if Nomad has actually removed the cgroups + tu.WaitForResult(func() (bool, error) { + _, err = os.Stat(memLimits) + if err == nil { + return false, fmt.Errorf("expected an error from os.Stat %s", memLimits) + } + return true, nil + }, func(err error) { t.Error(err) }) + + } expected := `/: alloc/ bin/ @@ -123,6 +134,7 @@ lib64/ local/ proc/ secrets/ +sys/ tmp/ usr/ @@ -130,66 +142,43 @@ usr/ ld.so.cache ld.so.conf ld.so.conf.d/` - file := filepath.Join(ctx.LogDir, "web.stdout.0") - output, err := ioutil.ReadFile(file) - if err != nil { - t.Fatalf("Couldn't read file %v", file) - } - - act := strings.TrimSpace(string(output)) - if act != expected { - t.Errorf("Command output incorrectly: want %v; got %v", expected, act) - } + tu.WaitForResult(func() (bool, error) { + output := execCmd.stdout.(*bufferCloser).String() + act := strings.TrimSpace(string(output)) + if act != expected { + return false, fmt.Errorf("Command output incorrectly: want %v; got %v", expected, act) + } + return true, nil + }, func(err error) { t.Error(err) }) } func TestExecutor_ClientCleanup(t *testing.T) { t.Parallel() testutil.ExecCompatible(t) + require := require.New(t) - ctx, allocDir := testExecutorContextWithChroot(t) - ctx.Task.LogConfig.MaxFiles = 1 - ctx.Task.LogConfig.MaxFileSizeMB = 300 + execCmd, allocDir := testExecutorCommandWithChroot(t) defer allocDir.Destroy() - executor := NewExecutor(testlog.Logger(t)) - - if err := executor.SetContext(ctx); err != nil { - t.Fatalf("Unexpected error") - } + executor := libcontainerFactory(testlog.HCLogger(t)) + defer executor.Shutdown("", 0) // Need to run a command which will produce continuous output but not // too quickly to ensure executor.Exit() stops the process. - execCmd := ExecCommand{Cmd: "/bin/bash", Args: []string{"-c", "while true; do /bin/echo X; /bin/sleep 1; done"}} - execCmd.FSIsolation = true + execCmd.Cmd = "/bin/bash" + execCmd.Args = []string{"-c", "while true; do /bin/echo X; /bin/sleep 1; done"} execCmd.ResourceLimits = true - execCmd.User = "nobody" - ps, err := executor.LaunchCmd(&execCmd) - if err != nil { - t.Fatalf("error in launching command: %v", err) - } - if ps.Pid == 0 { - t.Fatalf("expected process to start and have non zero pid") - } + ps, err := executor.Launch(execCmd) + require.NoError(err) + require.NotZero(ps.Pid) time.Sleep(500 * time.Millisecond) - if err := executor.Exit(); err != nil { - t.Fatalf("err: %v", err) - } + require.NoError(executor.Shutdown("SIGINT", 100*time.Millisecond)) + executor.Wait() - file := filepath.Join(ctx.LogDir, "web.stdout.0") - finfo, err := os.Stat(file) - if err != nil { - t.Fatalf("error stating stdout file: %v", err) - } - if finfo.Size() == 0 { - t.Fatal("Nothing in stdout; expected at least one byte.") - } + output := execCmd.stdout.(*bufferCloser).String() + require.NotZero(len(output)) time.Sleep(2 * time.Second) - finfo1, err := os.Stat(file) - if err != nil { - t.Fatalf("error stating stdout file: %v", err) - } - if finfo.Size() != finfo1.Size() { - t.Fatalf("Expected size: %v, actual: %v", finfo.Size(), finfo1.Size()) - } + output1 := execCmd.stdout.(*bufferCloser).String() + require.Equal(len(output), len(output1)) } diff --git a/client/driver/executor/executor_test.go b/client/driver/executor/executor_test.go index ddb5169ba..8fdabea46 100644 --- a/client/driver/executor/executor_test.go +++ b/client/driver/executor/executor_test.go @@ -1,6 +1,8 @@ package executor import ( + "bytes" + "fmt" "io/ioutil" "os" "path/filepath" @@ -9,24 +11,35 @@ import ( "testing" "time" + tu "github.com/hashicorp/nomad/testutil" + + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/driver/env" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/nomad/mock" - tu "github.com/hashicorp/nomad/testutil" - "github.com/mitchellh/go-ps" + "github.com/stretchr/testify/require" ) +var executorFactories = map[string]func(hclog.Logger) Executor{} +var universalFactory = func(l hclog.Logger) Executor { + return NewExecutor(l) +} + +func init() { + executorFactories["UniversalExecutor"] = universalFactory +} + // testExecutorContext returns an ExecutorContext and AllocDir. // // The caller is responsible for calling AllocDir.Destroy() to cleanup. -func testExecutorContext(t *testing.T) (*ExecutorContext, *allocdir.AllocDir) { +func testExecutorCommand(t *testing.T) (*ExecCommand, *allocdir.AllocDir) { alloc := mock.Alloc() task := alloc.Job.TaskGroups[0].Tasks[0] taskEnv := env.NewBuilder(mock.Node(), alloc, task, "global").Build() - allocDir := allocdir.NewAllocDir(testlog.Logger(t), filepath.Join(os.TempDir(), alloc.ID)) + allocDir := allocdir.NewAllocDir(testlog.HCLogger(t), filepath.Join(os.TempDir(), alloc.ID)) if err := allocDir.Build(); err != nil { t.Fatalf("AllocDir.Build() failed: %v", err) } @@ -35,184 +48,176 @@ func testExecutorContext(t *testing.T) (*ExecutorContext, *allocdir.AllocDir) { t.Fatalf("allocDir.NewTaskDir(%q) failed: %v", task.Name, err) } td := allocDir.TaskDirs[task.Name] - ctx := &ExecutorContext{ - TaskEnv: taskEnv, - Task: task, + cmd := &ExecCommand{ + Env: taskEnv.List(), TaskDir: td.Dir, - LogDir: td.LogDir, + Resources: &Resources{ + CPU: task.Resources.CPU, + MemoryMB: task.Resources.MemoryMB, + IOPS: task.Resources.IOPS, + DiskMB: task.Resources.DiskMB, + }, } - return ctx, allocDir + configureTLogging(cmd) + return cmd, allocDir } -func TestExecutor_Start_Invalid(t *testing.T) { - t.Parallel() +type bufferCloser struct { + bytes.Buffer +} + +func (_ *bufferCloser) Close() error { return nil } + +func configureTLogging(cmd *ExecCommand) (stdout bufferCloser, stderr bufferCloser) { + cmd.stdout = &stdout + cmd.stderr = &stderr + return +} + +func TestExecutor_Start_Invalid(pt *testing.T) { + pt.Parallel() invalid := "/bin/foobar" - execCmd := ExecCommand{Cmd: invalid, Args: []string{"1"}} - ctx, allocDir := testExecutorContext(t) - defer allocDir.Destroy() - executor := NewExecutor(testlog.Logger(t)) + for name, factory := range executorFactories { + pt.Run(name, func(t *testing.T) { + t.Parallel() + require := require.New(t) + execCmd, allocDir := testExecutorCommand(t) + execCmd.Cmd = invalid + execCmd.Args = []string{"1"} + defer allocDir.Destroy() + executor := factory(testlog.HCLogger(t)) + defer executor.Shutdown("", 0) - if err := executor.SetContext(ctx); err != nil { - t.Fatalf("Unexpected error") - } - - if _, err := executor.LaunchCmd(&execCmd); err == nil { - t.Fatalf("Expected error") + _, err := executor.Launch(execCmd) + require.Error(err) + }) } } -func TestExecutor_Start_Wait_Failure_Code(t *testing.T) { - t.Parallel() - execCmd := ExecCommand{Cmd: "/bin/date", Args: []string{"fail"}} - ctx, allocDir := testExecutorContext(t) - defer allocDir.Destroy() - executor := NewExecutor(testlog.Logger(t)) +func TestExecutor_Start_Wait_Failure_Code(pt *testing.T) { + pt.Parallel() + for name, factory := range executorFactories { + pt.Run(name, func(t *testing.T) { + t.Parallel() + require := require.New(t) + execCmd, allocDir := testExecutorCommand(t) + execCmd.Cmd = "/bin/date" + execCmd.Args = []string{"fail"} + defer allocDir.Destroy() + executor := factory(testlog.HCLogger(t)) + defer executor.Shutdown("", 0) - if err := executor.SetContext(ctx); err != nil { - t.Fatalf("Unexpected error") - } - - ps, err := executor.LaunchCmd(&execCmd) - if err != nil { - t.Fatalf("Unexpected error") - } - - if ps.Pid == 0 { - t.Fatalf("expected process to start and have non zero pid") - } - ps, _ = executor.Wait() - if ps.ExitCode < 1 { - t.Fatalf("expected exit code to be non zero, actual: %v", ps.ExitCode) - } - if err := executor.Exit(); err != nil { - t.Fatalf("error: %v", err) + ps, err := executor.Launch(execCmd) + require.NoError(err) + require.NotZero(ps.Pid) + ps, _ = executor.Wait() + require.NotZero(ps.ExitCode, "expected exit code to be non zero") + require.NoError(executor.Shutdown("SIGINT", 100*time.Millisecond)) + }) } } -func TestExecutor_Start_Wait(t *testing.T) { - t.Parallel() - execCmd := ExecCommand{Cmd: "/bin/echo", Args: []string{"hello world"}} - ctx, allocDir := testExecutorContext(t) - defer allocDir.Destroy() - executor := NewExecutor(testlog.Logger(t)) +func TestExecutor_Start_Wait(pt *testing.T) { + pt.Parallel() + for name, factory := range executorFactories { + pt.Run(name, func(t *testing.T) { + t.Parallel() + require := require.New(t) + execCmd, allocDir := testExecutorCommand(t) + execCmd.Cmd = "/bin/echo" + execCmd.Args = []string{"hello world"} + defer allocDir.Destroy() + executor := factory(testlog.HCLogger(t)) + defer executor.Shutdown("", 0) - if err := executor.SetContext(ctx); err != nil { - t.Fatalf("Unexpected error") - } + ps, err := executor.Launch(execCmd) + require.NoError(err) + require.NotZero(ps.Pid) - ps, err := executor.LaunchCmd(&execCmd) - if err != nil { - t.Fatalf("error in launching command: %v", err) - } - if ps.Pid == 0 { - t.Fatalf("expected process to start and have non zero pid") - } - ps, err = executor.Wait() - if err != nil { - t.Fatalf("error in waiting for command: %v", err) - } - if err := executor.Exit(); err != nil { - t.Fatalf("error: %v", err) - } + ps, err = executor.Wait() + require.NoError(err) + require.NoError(executor.Shutdown("SIGINT", 100*time.Millisecond)) - expected := "hello world" - file := filepath.Join(ctx.LogDir, "web.stdout.0") - output, err := ioutil.ReadFile(file) - if err != nil { - t.Fatalf("Couldn't read file %v", file) - } - - act := strings.TrimSpace(string(output)) - if act != expected { - t.Fatalf("Command output incorrectly: want %v; got %v", expected, act) + expected := "hello world" + tu.WaitForResult(func() (bool, error) { + output := execCmd.stdout.(*bufferCloser).String() + act := strings.TrimSpace(string(output)) + if expected != act { + return false, fmt.Errorf("expected: '%s' actual: '%s'", expected, act) + } + return true, nil + }, func(err error) { + require.NoError(err) + }) + }) } } -func TestExecutor_WaitExitSignal(t *testing.T) { - t.Parallel() - execCmd := ExecCommand{Cmd: "/bin/sleep", Args: []string{"10000"}} - ctx, allocDir := testExecutorContext(t) - defer allocDir.Destroy() - executor := NewExecutor(testlog.Logger(t)) +func TestExecutor_WaitExitSignal(pt *testing.T) { + pt.Parallel() + for name, factory := range executorFactories { + pt.Run(name, func(t *testing.T) { + t.Parallel() + require := require.New(t) + execCmd, allocDir := testExecutorCommand(t) + execCmd.Cmd = "/bin/sleep" + execCmd.Args = []string{"10000"} + defer allocDir.Destroy() + executor := factory(testlog.HCLogger(t)) + defer executor.Shutdown("", 0) - if err := executor.SetContext(ctx); err != nil { - t.Fatalf("Unexpected error") - } + ps, err := executor.Launch(execCmd) + require.NoError(err) - ps, err := executor.LaunchCmd(&execCmd) - if err != nil { - t.Fatalf("err: %v", err) - } + go func() { + time.Sleep(2 * time.Second) + _, err := executor.Stats() + require.NoError(err) + //require.NotEmpty(ru.Pids) + proc, err := os.FindProcess(ps.Pid) + require.NoError(err) + err = proc.Signal(syscall.SIGKILL) + require.NoError(err) + }() - go func() { - time.Sleep(2 * time.Second) - ru, err := executor.Stats() - if err != nil { - t.Fatalf("err: %v", err) - } - if len(ru.Pids) == 0 { - t.Fatalf("expected pids") - } - proc, err := os.FindProcess(ps.Pid) - if err != nil { - t.Fatalf("err: %v", err) - } - if err := proc.Signal(syscall.SIGKILL); err != nil { - t.Fatalf("err: %v", err) - } - }() - - ps, err = executor.Wait() - if err != nil { - t.Fatalf("err: %v", err) - } - if ps.Signal != int(syscall.SIGKILL) { - t.Fatalf("expected signal: %v, actual: %v", int(syscall.SIGKILL), ps.Signal) + ps, err = executor.Wait() + require.NoError(err) + require.Equal(ps.Signal, int(syscall.SIGKILL)) + }) } } -func TestExecutor_Start_Kill(t *testing.T) { - t.Parallel() - execCmd := ExecCommand{Cmd: "/bin/sleep", Args: []string{"10 && hello world"}} - ctx, allocDir := testExecutorContext(t) - defer allocDir.Destroy() - executor := NewExecutor(testlog.Logger(t)) +func TestExecutor_Start_Kill(pt *testing.T) { + pt.Parallel() + for name, factory := range executorFactories { + pt.Run(name, func(t *testing.T) { + t.Parallel() + require := require.New(t) + execCmd, allocDir := testExecutorCommand(t) + execCmd.Cmd = "/bin/sleep" + execCmd.Args = []string{"10 && hello world"} + defer allocDir.Destroy() + executor := factory(testlog.HCLogger(t)) + defer executor.Shutdown("", 0) - if err := executor.SetContext(ctx); err != nil { - t.Fatalf("Unexpected error") - } + ps, err := executor.Launch(execCmd) + require.NoError(err) + require.NotZero(ps.Pid) - ps, err := executor.LaunchCmd(&execCmd) - if err != nil { - t.Fatalf("error in launching command: %v", err) - } - if ps.Pid == 0 { - t.Fatalf("expected process to start and have non zero pid") - } - ps, err = executor.Wait() - if err != nil { - t.Fatalf("error in waiting for command: %v", err) - } - if err := executor.Exit(); err != nil { - t.Fatalf("error: %v", err) - } + require.NoError(executor.Shutdown("SIGINT", 100*time.Millisecond)) - file := filepath.Join(ctx.LogDir, "web.stdout.0") - time.Sleep(time.Duration(tu.TestMultiplier()*2) * time.Second) - - output, err := ioutil.ReadFile(file) - if err != nil { - t.Fatalf("Couldn't read file %v", file) - } - - expected := "" - act := strings.TrimSpace(string(output)) - if act != expected { - t.Fatalf("Command output incorrectly: want %v; got %v", expected, act) + time.Sleep(time.Duration(tu.TestMultiplier()*2) * time.Second) + output := execCmd.stdout.(*bufferCloser).String() + expected := "" + act := strings.TrimSpace(string(output)) + if act != expected { + t.Fatalf("Command output incorrectly: want %v; got %v", expected, act) + } + }) } } -func TestExecutor_MakeExecutable(t *testing.T) { +func TestUniversalExecutor_MakeExecutable(t *testing.T) { t.Parallel() // Create a temp file f, err := ioutil.TempFile("", "") @@ -226,9 +231,9 @@ func TestExecutor_MakeExecutable(t *testing.T) { f.Chmod(os.FileMode(0610)) // Make a fake executor - executor := NewExecutor(testlog.Logger(t)) + executor := NewExecutor(testlog.HCLogger(t)).(*UniversalExecutor) - err = executor.(*UniversalExecutor).makeExecutable(f.Name()) + err = executor.makeExecutable(f.Name()) if err != nil { t.Fatalf("makeExecutable() failed: %v", err) } @@ -245,44 +250,3 @@ func TestExecutor_MakeExecutable(t *testing.T) { t.Fatalf("expected permissions %v; got %v", exp, act) } } - -func TestScanPids(t *testing.T) { - t.Parallel() - p1 := NewFakeProcess(2, 5) - p2 := NewFakeProcess(10, 2) - p3 := NewFakeProcess(15, 6) - p4 := NewFakeProcess(3, 10) - p5 := NewFakeProcess(20, 18) - - // Make a fake executor - executor := NewExecutor(testlog.Logger(t)).(*UniversalExecutor) - - nomadPids, err := executor.scanPids(5, []ps.Process{p1, p2, p3, p4, p5}) - if err != nil { - t.Fatalf("error: %v", err) - } - if len(nomadPids) != 4 { - t.Fatalf("expected: 4, actual: %v", len(nomadPids)) - } -} - -type FakeProcess struct { - pid int - ppid int -} - -func (f FakeProcess) Pid() int { - return f.pid -} - -func (f FakeProcess) PPid() int { - return f.ppid -} - -func (f FakeProcess) Executable() string { - return "fake" -} - -func NewFakeProcess(pid int, ppid int) ps.Process { - return FakeProcess{pid: pid, ppid: ppid} -} diff --git a/client/driver/executor/executor_universal_linux.go b/client/driver/executor/executor_universal_linux.go new file mode 100644 index 000000000..30fdd9c69 --- /dev/null +++ b/client/driver/executor/executor_universal_linux.go @@ -0,0 +1,156 @@ +package executor + +import ( + "fmt" + "os" + "os/user" + "strconv" + "syscall" + + multierror "github.com/hashicorp/go-multierror" + "github.com/hashicorp/nomad/helper" + "github.com/opencontainers/runc/libcontainer/cgroups" + cgroupFs "github.com/opencontainers/runc/libcontainer/cgroups/fs" + lconfigs "github.com/opencontainers/runc/libcontainer/configs" +) + +// runAs takes a user id as a string and looks up the user, and sets the command +// to execute as that user. +func (e *UniversalExecutor) runAs(userid string) error { + u, err := user.Lookup(userid) + if err != nil { + return fmt.Errorf("Failed to identify user %v: %v", userid, err) + } + + // Get the groups the user is a part of + gidStrings, err := u.GroupIds() + if err != nil { + return fmt.Errorf("Unable to lookup user's group membership: %v", err) + } + + gids := make([]uint32, len(gidStrings)) + for _, gidString := range gidStrings { + u, err := strconv.Atoi(gidString) + if err != nil { + return fmt.Errorf("Unable to convert user's group to int %s: %v", gidString, err) + } + + gids = append(gids, uint32(u)) + } + + // Convert the uid and gid + uid, err := strconv.ParseUint(u.Uid, 10, 32) + if err != nil { + return fmt.Errorf("Unable to convert userid to uint32: %s", err) + } + gid, err := strconv.ParseUint(u.Gid, 10, 32) + if err != nil { + return fmt.Errorf("Unable to convert groupid to uint32: %s", err) + } + + // Set the command to run as that user and group. + if e.childCmd.SysProcAttr == nil { + e.childCmd.SysProcAttr = &syscall.SysProcAttr{} + } + if e.childCmd.SysProcAttr.Credential == nil { + e.childCmd.SysProcAttr.Credential = &syscall.Credential{} + } + e.childCmd.SysProcAttr.Credential.Uid = uint32(uid) + e.childCmd.SysProcAttr.Credential.Gid = uint32(gid) + e.childCmd.SysProcAttr.Credential.Groups = gids + + e.logger.Debug("setting process user", "user", uid, "group", gid, "additional_groups", gids) + + return nil +} + +// configureResourceContainer configured the cgroups to be used to track pids +// created by the executor +func (e *UniversalExecutor) configureResourceContainer(pid int) error { + cfg := &lconfigs.Config{ + Cgroups: &lconfigs.Cgroup{ + Resources: &lconfigs.Resources{ + AllowAllDevices: helper.BoolToPtr(true), + }, + }, + } + + configureBasicCgroups(cfg) + e.resConCtx.groups = cfg.Cgroups + return cgroups.EnterPid(cfg.Cgroups.Paths, pid) +} + +// DestroyCgroup kills all processes in the cgroup and removes the cgroup +// configuration from the host. This function is idempotent. +func DestroyCgroup(groups *lconfigs.Cgroup, executorPid int) error { + mErrs := new(multierror.Error) + if groups == nil { + return fmt.Errorf("Can't destroy: cgroup configuration empty") + } + + // Move the executor into the global cgroup so that the task specific + // cgroup can be destroyed. + path, err := cgroups.GetInitCgroupPath("freezer") + if err != nil { + return err + } + + if err := cgroups.EnterPid(map[string]string{"freezer": path}, executorPid); err != nil { + return err + } + + // Freeze the Cgroup so that it can not continue to fork/exec. + groups.Resources.Freezer = lconfigs.Frozen + freezer := cgroupFs.FreezerGroup{} + if err := freezer.Set(groups.Paths[freezer.Name()], groups); err != nil { + return err + } + + var procs []*os.Process + pids, err := cgroups.GetAllPids(groups.Paths[freezer.Name()]) + if err != nil { + multierror.Append(mErrs, fmt.Errorf("error getting pids: %v", err)) + + // Unfreeze the cgroup. + groups.Resources.Freezer = lconfigs.Thawed + freezer := cgroupFs.FreezerGroup{} + if err := freezer.Set(groups.Paths[freezer.Name()], groups); err != nil { + multierror.Append(mErrs, fmt.Errorf("failed to unfreeze cgroup: %v", err)) + return mErrs.ErrorOrNil() + } + } + + // Kill the processes in the cgroup + for _, pid := range pids { + proc, err := os.FindProcess(pid) + if err != nil { + multierror.Append(mErrs, fmt.Errorf("error finding process %v: %v", pid, err)) + continue + } + + procs = append(procs, proc) + if e := proc.Kill(); e != nil { + multierror.Append(mErrs, fmt.Errorf("error killing process %v: %v", pid, e)) + } + } + + // Unfreeze the cgroug so we can wait. + groups.Resources.Freezer = lconfigs.Thawed + if err := freezer.Set(groups.Paths[freezer.Name()], groups); err != nil { + multierror.Append(mErrs, fmt.Errorf("failed to unfreeze cgroup: %v", err)) + return mErrs.ErrorOrNil() + } + + // Wait on the killed processes to ensure they are cleaned up. + for _, proc := range procs { + // Don't capture the error because we expect this to fail for + // processes we didn't fork. + proc.Wait() + } + + // Remove the cgroup. + if err := cgroups.RemovePaths(groups.Paths); err != nil { + multierror.Append(mErrs, fmt.Errorf("failed to delete the cgroup directories: %v", err)) + } + return mErrs.ErrorOrNil() +} diff --git a/client/driver/executor/executor_unix.go b/client/driver/executor/executor_unix.go index af6f5a488..6f90b7ade 100644 --- a/client/driver/executor/executor_unix.go +++ b/client/driver/executor/executor_unix.go @@ -10,10 +10,10 @@ import ( // configure new process group for child process func (e *UniversalExecutor) setNewProcessGroup() error { - if e.cmd.SysProcAttr == nil { - e.cmd.SysProcAttr = &syscall.SysProcAttr{} + if e.childCmd.SysProcAttr == nil { + e.childCmd.SysProcAttr = &syscall.SysProcAttr{} } - e.cmd.SysProcAttr.Setpgid = true + e.childCmd.SysProcAttr.Setpgid = true return nil } @@ -21,7 +21,7 @@ func (e *UniversalExecutor) setNewProcessGroup() error { func (e *UniversalExecutor) cleanupChildProcesses(proc *os.Process) error { // If new process group was created upon command execution // we can kill the whole process group now to cleanup any leftovers. - if e.cmd.SysProcAttr != nil && e.cmd.SysProcAttr.Setpgid { + if e.childCmd.SysProcAttr != nil && e.childCmd.SysProcAttr.Setpgid { if err := syscall.Kill(-proc.Pid, syscall.SIGKILL); err != nil && err.Error() != noSuchProcessErr { return err } @@ -32,18 +32,13 @@ func (e *UniversalExecutor) cleanupChildProcesses(proc *os.Process) error { // Only send the process a shutdown signal (default INT), doesn't // necessarily kill it. -func (e *UniversalExecutor) shutdownProcess(proc *os.Process) error { - // Set default kill signal, as some drivers don't support configurable - // signals (such as rkt) - var osSignal os.Signal - if e.command.TaskKillSignal != nil { - osSignal = e.command.TaskKillSignal - } else { - osSignal = os.Interrupt +func (e *UniversalExecutor) shutdownProcess(sig os.Signal, proc *os.Process) error { + if sig == nil { + sig = os.Interrupt } - if err := proc.Signal(osSignal); err != nil && err.Error() != finishedErr { - return fmt.Errorf("executor.shutdown error: %v", err) + if err := proc.Signal(sig); err != nil && err.Error() != finishedErr { + return fmt.Errorf("executor shutdown error: %v", err) } return nil diff --git a/client/driver/executor/executor_windows.go b/client/driver/executor/executor_windows.go index b3c027bfb..3b508184f 100644 --- a/client/driver/executor/executor_windows.go +++ b/client/driver/executor/executor_windows.go @@ -59,11 +59,11 @@ func sendCtrlBreak(pid int) error { // Send the process a Ctrl-Break event, allowing it to shutdown by itself // before being Terminate. -func (e *UniversalExecutor) shutdownProcess(proc *os.Process) error { +func (e *UniversalExecutor) shutdownProcess(_ os.Signal, proc *os.Process) error { if err := sendCtrlBreak(proc.Pid); err != nil { - return fmt.Errorf("executor.shutdown error: %v", err) + return fmt.Errorf("executor shutdown error: %v", err) } - e.logger.Printf("[INFO] executor: sent Ctrl-Break to process %v", proc.Pid) + e.logger.Info("sent Ctrl-Break to process", "pid", proc.Pid) return nil } diff --git a/client/driver/executor/pid_collector.go b/client/driver/executor/pid_collector.go new file mode 100644 index 000000000..79d4acbf8 --- /dev/null +++ b/client/driver/executor/pid_collector.go @@ -0,0 +1,219 @@ +package executor + +import ( + "os" + "strconv" + "sync" + "time" + + hclog "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/stats" + cstructs "github.com/hashicorp/nomad/client/structs" + ps "github.com/mitchellh/go-ps" + "github.com/shirou/gopsutil/process" +) + +var ( + // pidScanInterval is the interval at which the executor scans the process + // tree for finding out the pids that the executor and it's child processes + // have forked + pidScanInterval = 5 * time.Second +) + +// pidCollector is a utility that can be embedded in an executor to collect pid +// stats +type pidCollector struct { + pids map[int]*nomadPid + pidLock sync.RWMutex + logger hclog.Logger +} + +// nomadPid holds a pid and it's cpu percentage calculator +type nomadPid struct { + pid int + cpuStatsTotal *stats.CpuStats + cpuStatsUser *stats.CpuStats + cpuStatsSys *stats.CpuStats +} + +// allPidGetter is a func which is used by the pid collector to gather +// stats on +type allPidGetter func() (map[int]*nomadPid, error) + +func newPidCollector(logger hclog.Logger) *pidCollector { + return &pidCollector{ + pids: make(map[int]*nomadPid), + logger: logger.Named("pid_collector"), + } +} + +// collectPids collects the pids of the child processes that the executor is +// running every 5 seconds +func (c *pidCollector) collectPids(stopCh chan interface{}, pidGetter allPidGetter) { + // Fire the timer right away when the executor starts from there on the pids + // are collected every scan interval + timer := time.NewTimer(0) + defer timer.Stop() + for { + select { + case <-timer.C: + pids, err := pidGetter() + if err != nil { + c.logger.Debug("error collecting pids", "error", err) + } + c.pidLock.Lock() + + // Adding pids which are not being tracked + for pid, np := range pids { + if _, ok := c.pids[pid]; !ok { + c.pids[pid] = np + } + } + // Removing pids which are no longer present + for pid := range c.pids { + if _, ok := pids[pid]; !ok { + delete(c.pids, pid) + } + } + c.pidLock.Unlock() + timer.Reset(pidScanInterval) + case <-stopCh: + return + } + } +} + +// scanPids scans all the pids on the machine running the current executor and +// returns the child processes of the executor. +func scanPids(parentPid int, allPids []ps.Process) (map[int]*nomadPid, error) { + processFamily := make(map[int]struct{}) + processFamily[parentPid] = struct{}{} + + // A mapping of pids to their parent pids. It is used to build the process + // tree of the executing task + pidsRemaining := make(map[int]int, len(allPids)) + for _, pid := range allPids { + pidsRemaining[pid.Pid()] = pid.PPid() + } + + for { + // flag to indicate if we have found a match + foundNewPid := false + + for pid, ppid := range pidsRemaining { + _, childPid := processFamily[ppid] + + // checking if the pid is a child of any of the parents + if childPid { + processFamily[pid] = struct{}{} + delete(pidsRemaining, pid) + foundNewPid = true + } + } + + // not scanning anymore if we couldn't find a single match + if !foundNewPid { + break + } + } + + res := make(map[int]*nomadPid) + for pid := range processFamily { + np := nomadPid{ + pid: pid, + cpuStatsTotal: stats.NewCpuStats(), + cpuStatsUser: stats.NewCpuStats(), + cpuStatsSys: stats.NewCpuStats(), + } + res[pid] = &np + } + return res, nil +} + +// pidStats returns the resource usage stats per pid +func (c *pidCollector) pidStats() (map[string]*cstructs.ResourceUsage, error) { + stats := make(map[string]*cstructs.ResourceUsage) + c.pidLock.RLock() + pids := make(map[int]*nomadPid, len(c.pids)) + for k, v := range c.pids { + pids[k] = v + } + c.pidLock.RUnlock() + for pid, np := range pids { + p, err := process.NewProcess(int32(pid)) + if err != nil { + c.logger.Trace("unable to create new process", "pid", pid, "error", err) + continue + } + ms := &cstructs.MemoryStats{} + if memInfo, err := p.MemoryInfo(); err == nil { + ms.RSS = memInfo.RSS + ms.Swap = memInfo.Swap + ms.Measured = ExecutorBasicMeasuredMemStats + } + + cs := &cstructs.CpuStats{} + if cpuStats, err := p.Times(); err == nil { + cs.SystemMode = np.cpuStatsSys.Percent(cpuStats.System * float64(time.Second)) + cs.UserMode = np.cpuStatsUser.Percent(cpuStats.User * float64(time.Second)) + cs.Measured = ExecutorBasicMeasuredCpuStats + + // calculate cpu usage percent + cs.Percent = np.cpuStatsTotal.Percent(cpuStats.Total() * float64(time.Second)) + } + stats[strconv.Itoa(pid)] = &cstructs.ResourceUsage{MemoryStats: ms, CpuStats: cs} + } + + return stats, nil +} + +// aggregatedResourceUsage aggregates the resource usage of all the pids and +// returns a TaskResourceUsage data point +func aggregatedResourceUsage(systemCpuStats *stats.CpuStats, pidStats map[string]*cstructs.ResourceUsage) *cstructs.TaskResourceUsage { + ts := time.Now().UTC().UnixNano() + var ( + systemModeCPU, userModeCPU, percent float64 + totalRSS, totalSwap uint64 + ) + + for _, pidStat := range pidStats { + systemModeCPU += pidStat.CpuStats.SystemMode + userModeCPU += pidStat.CpuStats.UserMode + percent += pidStat.CpuStats.Percent + + totalRSS += pidStat.MemoryStats.RSS + totalSwap += pidStat.MemoryStats.Swap + } + + totalCPU := &cstructs.CpuStats{ + SystemMode: systemModeCPU, + UserMode: userModeCPU, + Percent: percent, + Measured: ExecutorBasicMeasuredCpuStats, + TotalTicks: systemCpuStats.TicksConsumed(percent), + } + + totalMemory := &cstructs.MemoryStats{ + RSS: totalRSS, + Swap: totalSwap, + Measured: ExecutorBasicMeasuredMemStats, + } + + resourceUsage := cstructs.ResourceUsage{ + MemoryStats: totalMemory, + CpuStats: totalCPU, + } + return &cstructs.TaskResourceUsage{ + ResourceUsage: &resourceUsage, + Timestamp: ts, + Pids: pidStats, + } +} + +func getAllPids() (map[int]*nomadPid, error) { + allProcesses, err := ps.Processes() + if err != nil { + return nil, err + } + return scanPids(os.Getpid(), allProcesses) +} diff --git a/client/driver/executor/pid_collector_test.go b/client/driver/executor/pid_collector_test.go new file mode 100644 index 000000000..c42397b68 --- /dev/null +++ b/client/driver/executor/pid_collector_test.go @@ -0,0 +1,45 @@ +package executor + +import ( + "testing" + + ps "github.com/mitchellh/go-ps" +) + +func TestScanPids(t *testing.T) { + t.Parallel() + p1 := NewFakeProcess(2, 5) + p2 := NewFakeProcess(10, 2) + p3 := NewFakeProcess(15, 6) + p4 := NewFakeProcess(3, 10) + p5 := NewFakeProcess(20, 18) + + nomadPids, err := scanPids(5, []ps.Process{p1, p2, p3, p4, p5}) + if err != nil { + t.Fatalf("error: %v", err) + } + if len(nomadPids) != 4 { + t.Fatalf("expected: 4, actual: %v", len(nomadPids)) + } +} + +type FakeProcess struct { + pid int + ppid int +} + +func (f FakeProcess) Pid() int { + return f.pid +} + +func (f FakeProcess) PPid() int { + return f.ppid +} + +func (f FakeProcess) Executable() string { + return "fake" +} + +func NewFakeProcess(pid int, ppid int) ps.Process { + return FakeProcess{pid: pid, ppid: ppid} +} diff --git a/client/driver/executor/resource_container_default.go b/client/driver/executor/resource_container_default.go index 6e9503206..7bd02dbb6 100644 --- a/client/driver/executor/resource_container_default.go +++ b/client/driver/executor/resource_container_default.go @@ -2,23 +2,11 @@ package executor -import ( - dstructs "github.com/hashicorp/nomad/client/driver/structs" -) - // resourceContainerContext is a platform-specific struct for managing a // resource container. type resourceContainerContext struct { } -func clientCleanup(ic *dstructs.IsolationConfig, pid int) error { - return nil -} - func (rc *resourceContainerContext) executorCleanup() error { return nil } - -func (rc *resourceContainerContext) getIsolationConfig() *dstructs.IsolationConfig { - return nil -} diff --git a/client/driver/executor/resource_container_linux.go b/client/driver/executor/resource_container_linux.go index a9efbbc53..5a414c9cf 100644 --- a/client/driver/executor/resource_container_linux.go +++ b/client/driver/executor/resource_container_linux.go @@ -4,39 +4,22 @@ import ( "os" "sync" - dstructs "github.com/hashicorp/nomad/client/driver/structs" cgroupConfig "github.com/opencontainers/runc/libcontainer/configs" ) // resourceContainerContext is a platform-specific struct for managing a // resource container. In the case of Linux, this is used to control Cgroups. type resourceContainerContext struct { - groups *cgroupConfig.Cgroup - cgPaths map[string]string - cgLock sync.Mutex -} - -// clientCleanup removes this host's Cgroup from the Nomad Client's context -func clientCleanup(ic *dstructs.IsolationConfig, pid int) error { - if err := DestroyCgroup(ic.Cgroup, ic.CgroupPaths, pid); err != nil { - return err - } - return nil + groups *cgroupConfig.Cgroup + cgLock sync.Mutex } // cleanup removes this host's Cgroup from within an Executor's context func (rc *resourceContainerContext) executorCleanup() error { rc.cgLock.Lock() defer rc.cgLock.Unlock() - if err := DestroyCgroup(rc.groups, rc.cgPaths, os.Getpid()); err != nil { + if err := DestroyCgroup(rc.groups, os.Getpid()); err != nil { return err } return nil } - -func (rc *resourceContainerContext) getIsolationConfig() *dstructs.IsolationConfig { - return &dstructs.IsolationConfig{ - Cgroup: rc.groups, - CgroupPaths: rc.cgPaths, - } -} diff --git a/client/driver/executor_plugin.go b/client/driver/executor_plugin.go index 01e63343d..b2e30491d 100644 --- a/client/driver/executor_plugin.go +++ b/client/driver/executor_plugin.go @@ -2,16 +2,15 @@ package driver import ( "encoding/gob" - "log" "net/rpc" "os" "syscall" "time" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/go-plugin" "github.com/hashicorp/nomad/client/driver/executor" cstructs "github.com/hashicorp/nomad/client/structs" - "github.com/hashicorp/nomad/nomad/structs" ) // Registering these types since we have to serialize and de-serialize the Task @@ -26,65 +25,53 @@ func init() { type ExecutorRPC struct { client *rpc.Client - logger *log.Logger + logger hclog.Logger } // LaunchCmdArgs wraps a user command and the args for the purposes of RPC -type LaunchCmdArgs struct { +type LaunchArgs struct { Cmd *executor.ExecCommand } -type ExecCmdArgs struct { +// ShutdownArgs wraps shutdown signal and grace period +type ShutdownArgs struct { + Signal string + GracePeriod time.Duration +} + +type ExecArgs struct { Deadline time.Time Name string Args []string } -type ExecCmdReturn struct { +type ExecReturn struct { Output []byte Code int } -func (e *ExecutorRPC) LaunchCmd(cmd *executor.ExecCommand) (*executor.ProcessState, error) { +func (e *ExecutorRPC) Launch(cmd *executor.ExecCommand) (*executor.ProcessState, error) { var ps *executor.ProcessState - err := e.client.Call("Plugin.LaunchCmd", LaunchCmdArgs{Cmd: cmd}, &ps) + err := e.client.Call("Plugin.Launch", LaunchArgs{Cmd: cmd}, &ps) return ps, err } -func (e *ExecutorRPC) LaunchSyslogServer() (*executor.SyslogServerState, error) { - var ss *executor.SyslogServerState - err := e.client.Call("Plugin.LaunchSyslogServer", new(interface{}), &ss) - return ss, err -} - func (e *ExecutorRPC) Wait() (*executor.ProcessState, error) { var ps executor.ProcessState err := e.client.Call("Plugin.Wait", new(interface{}), &ps) return &ps, err } -func (e *ExecutorRPC) ShutDown() error { - return e.client.Call("Plugin.ShutDown", new(interface{}), new(interface{})) +func (e *ExecutorRPC) Kill() error { + return e.client.Call("Plugin.Kill", new(interface{}), new(interface{})) } -func (e *ExecutorRPC) Exit() error { - return e.client.Call("Plugin.Exit", new(interface{}), new(interface{})) +func (e *ExecutorRPC) Shutdown(signal string, grace time.Duration) error { + return e.client.Call("Plugin.Shutdown", &ShutdownArgs{signal, grace}, new(interface{})) } -func (e *ExecutorRPC) SetContext(ctx *executor.ExecutorContext) error { - return e.client.Call("Plugin.SetContext", ctx, new(interface{})) -} - -func (e *ExecutorRPC) UpdateLogConfig(logConfig *structs.LogConfig) error { - return e.client.Call("Plugin.UpdateLogConfig", logConfig, new(interface{})) -} - -func (e *ExecutorRPC) UpdateTask(task *structs.Task) error { - return e.client.Call("Plugin.UpdateTask", task, new(interface{})) -} - -func (e *ExecutorRPC) DeregisterServices() error { - return e.client.Call("Plugin.DeregisterServices", new(interface{}), new(interface{})) +func (e *ExecutorRPC) UpdateResources(resources *executor.Resources) error { + return e.client.Call("Plugin.UpdateResources", resources, new(interface{})) } func (e *ExecutorRPC) Version() (*executor.ExecutorVersion, error) { @@ -104,12 +91,12 @@ func (e *ExecutorRPC) Signal(s os.Signal) error { } func (e *ExecutorRPC) Exec(deadline time.Time, name string, args []string) ([]byte, int, error) { - req := ExecCmdArgs{ + req := ExecArgs{ Deadline: deadline, Name: name, Args: args, } - var resp *ExecCmdReturn + var resp *ExecReturn err := e.client.Call("Plugin.Exec", req, &resp) if resp == nil { return nil, 0, err @@ -119,25 +106,17 @@ func (e *ExecutorRPC) Exec(deadline time.Time, name string, args []string) ([]by type ExecutorRPCServer struct { Impl executor.Executor - logger *log.Logger + logger hclog.Logger } -func (e *ExecutorRPCServer) LaunchCmd(args LaunchCmdArgs, ps *executor.ProcessState) error { - state, err := e.Impl.LaunchCmd(args.Cmd) +func (e *ExecutorRPCServer) Launch(args LaunchArgs, ps *executor.ProcessState) error { + state, err := e.Impl.Launch(args.Cmd) if state != nil { *ps = *state } return err } -func (e *ExecutorRPCServer) LaunchSyslogServer(args interface{}, ss *executor.SyslogServerState) error { - state, err := e.Impl.LaunchSyslogServer() - if state != nil { - *ss = *state - } - return err -} - func (e *ExecutorRPCServer) Wait(args interface{}, ps *executor.ProcessState) error { state, err := e.Impl.Wait() if state != nil { @@ -146,29 +125,12 @@ func (e *ExecutorRPCServer) Wait(args interface{}, ps *executor.ProcessState) er return err } -func (e *ExecutorRPCServer) ShutDown(args interface{}, resp *interface{}) error { - return e.Impl.ShutDown() +func (e *ExecutorRPCServer) Shutdown(args ShutdownArgs, resp *interface{}) error { + return e.Impl.Shutdown(args.Signal, args.GracePeriod) } -func (e *ExecutorRPCServer) Exit(args interface{}, resp *interface{}) error { - return e.Impl.Exit() -} - -func (e *ExecutorRPCServer) SetContext(args *executor.ExecutorContext, resp *interface{}) error { - return e.Impl.SetContext(args) -} - -func (e *ExecutorRPCServer) UpdateLogConfig(args *structs.LogConfig, resp *interface{}) error { - return e.Impl.UpdateLogConfig(args) -} - -func (e *ExecutorRPCServer) UpdateTask(args *structs.Task, resp *interface{}) error { - return e.Impl.UpdateTask(args) -} - -func (e *ExecutorRPCServer) DeregisterServices(args interface{}, resp *interface{}) error { - // In 0.6 this is a noop. Goes away in 0.7. - return nil +func (e *ExecutorRPCServer) UpdateResources(args *executor.Resources, resp *interface{}) error { + return e.Impl.UpdateResources(args) } func (e *ExecutorRPCServer) Version(args interface{}, version *executor.ExecutorVersion) error { @@ -191,9 +153,9 @@ func (e *ExecutorRPCServer) Signal(args os.Signal, resp *interface{}) error { return e.Impl.Signal(args) } -func (e *ExecutorRPCServer) Exec(args ExecCmdArgs, result *ExecCmdReturn) error { +func (e *ExecutorRPCServer) Exec(args ExecArgs, result *ExecReturn) error { out, code, err := e.Impl.Exec(args.Deadline, args.Name, args.Args) - ret := &ExecCmdReturn{ + ret := &ExecReturn{ Output: out, Code: code, } @@ -202,13 +164,18 @@ func (e *ExecutorRPCServer) Exec(args ExecCmdArgs, result *ExecCmdReturn) error } type ExecutorPlugin struct { - logger *log.Logger - Impl *ExecutorRPCServer + logger hclog.Logger + fsIsolation bool + Impl *ExecutorRPCServer } func (p *ExecutorPlugin) Server(*plugin.MuxBroker) (interface{}, error) { if p.Impl == nil { - p.Impl = &ExecutorRPCServer{Impl: executor.NewExecutor(p.logger), logger: p.logger} + if p.fsIsolation { + p.Impl = &ExecutorRPCServer{Impl: executor.NewExecutorWithIsolation(p.logger), logger: p.logger} + } else { + p.Impl = &ExecutorRPCServer{Impl: executor.NewExecutor(p.logger), logger: p.logger} + } } return p.Impl, nil } diff --git a/client/driver/executor_plugin_pre_0_9_0.go b/client/driver/executor_plugin_pre_0_9_0.go new file mode 100644 index 000000000..eceea403d --- /dev/null +++ b/client/driver/executor_plugin_pre_0_9_0.go @@ -0,0 +1,220 @@ +package driver + +/* +import ( + "encoding/gob" + "log" + "net/rpc" + "os" + "syscall" + "time" + + "github.com/hashicorp/go-plugin" + executorv0 "github.com/hashicorp/nomad/client/driver/executorv0" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/nomad/structs" +) + +// Registering these types since we have to serialize and de-serialize the Task +// structs over the wire between drivers and the executorv0. +func init() { + gob.Register([]interface{}{}) + gob.Register(map[string]interface{}{}) + gob.Register([]map[string]string{}) + gob.Register([]map[string]int{}) + gob.Register(syscall.Signal(0x1)) +} + +type ExecutorRPCPre0_9_0 struct { + client *rpc.Client + logger *log.Logger +} + +// LaunchCmdArgs wraps a user command and the args for the purposes of RPC +type LaunchCmdArgs struct { + Cmd *executorv0.ExecCommand +} + +type ExecCmdArgs struct { + Deadline time.Time + Name string + Args []string +} + +type ExecCmdReturn struct { + Output []byte + Code int +} + +func (e *ExecutorRPCPre0_9_0) LaunchCmd(cmd *executorv0.ExecCommand) (*executorv0.ProcessState, error) { + var ps *executorv0.ProcessState + err := e.client.Call("Plugin.LaunchCmd", LaunchCmdArgs{Cmd: cmd}, &ps) + return ps, err +} + +func (e *ExecutorRPCPre0_9_0) LaunchSyslogServer() (*executorv0.SyslogServerState, error) { + var ss *executorv0.SyslogServerState + err := e.client.Call("Plugin.LaunchSyslogServer", new(interface{}), &ss) + return ss, err +} + +func (e *ExecutorRPCPre0_9_0) Wait() (*executorv0.ProcessState, error) { + var ps executorv0.ProcessState + err := e.client.Call("Plugin.Wait", new(interface{}), &ps) + return &ps, err +} + +func (e *ExecutorRPCPre0_9_0) ShutDown() error { + return e.client.Call("Plugin.ShutDown", new(interface{}), new(interface{})) +} + +func (e *ExecutorRPCPre0_9_0) Exit() error { + return e.client.Call("Plugin.Exit", new(interface{}), new(interface{})) +} + +func (e *ExecutorRPCPre0_9_0) SetContext(ctx *executorv0.ExecutorContext) error { + return e.client.Call("Plugin.SetContext", ctx, new(interface{})) +} + +func (e *ExecutorRPCPre0_9_0) UpdateLogConfig(logConfig *executorv0.LogConfig) error { + return e.client.Call("Plugin.UpdateLogConfig", logConfig, new(interface{})) +} + +func (e *ExecutorRPCPre0_9_0) UpdateTask(task *structs.Task) error { + return e.client.Call("Plugin.UpdateTask", task, new(interface{})) +} + +func (e *ExecutorRPCPre0_9_0) DeregisterServices() error { + return e.client.Call("Plugin.DeregisterServices", new(interface{}), new(interface{})) +} + +func (e *ExecutorRPCPre0_9_0) Version() (*executorv0.ExecutorVersion, error) { + var version executorv0.ExecutorVersion + err := e.client.Call("Plugin.Version", new(interface{}), &version) + return &version, err +} + +func (e *ExecutorRPCPre0_9_0) Stats() (*cstructs.TaskResourceUsage, error) { + var resourceUsage cstructs.TaskResourceUsage + err := e.client.Call("Plugin.Stats", new(interface{}), &resourceUsage) + return &resourceUsage, err +} + +func (e *ExecutorRPCPre0_9_0) Signal(s os.Signal) error { + return e.client.Call("Plugin.Signal", &s, new(interface{})) +} + +func (e *ExecutorRPCPre0_9_0) Exec(deadline time.Time, name string, args []string) ([]byte, int, error) { + req := ExecCmdArgs{ + Deadline: deadline, + Name: name, + Args: args, + } + var resp *ExecCmdReturn + err := e.client.Call("Plugin.Exec", req, &resp) + if resp == nil { + return nil, 0, err + } + return resp.Output, resp.Code, err +} + +type ExecutorRPCServerPre0_9_0 struct { + Impl executorv0.Executor + logger *log.Logger +} + +func (e *ExecutorRPCServerPre0_9_0) LaunchCmd(args LaunchCmdArgs, ps *executorv0.ProcessState) error { + state, err := e.Impl.LaunchCmd(args.Cmd) + if state != nil { + *ps = *state + } + return err +} + +func (e *ExecutorRPCServerPre0_9_0) LaunchSyslogServer(args interface{}, ss *executorv0.SyslogServerState) error { + state, err := e.Impl.LaunchSyslogServer() + if state != nil { + *ss = *state + } + return err +} + +func (e *ExecutorRPCServerPre0_9_0) Wait(args interface{}, ps *executorv0.ProcessState) error { + state, err := e.Impl.Wait() + if state != nil { + *ps = *state + } + return err +} + +func (e *ExecutorRPCServerPre0_9_0) ShutDown(args interface{}, resp *interface{}) error { + return e.Impl.ShutDown() +} + +func (e *ExecutorRPCServerPre0_9_0) Exit(args interface{}, resp *interface{}) error { + return e.Impl.Exit() +} + +func (e *ExecutorRPCServerPre0_9_0) SetContext(args *executorv0.ExecutorContext, resp *interface{}) error { + return e.Impl.SetContext(args) +} + +func (e *ExecutorRPCServerPre0_9_0) UpdateLogConfig(args *executorv0.LogConfig, resp *interface{}) error { + return e.Impl.UpdateLogConfig(args) +} + +func (e *ExecutorRPCServerPre0_9_0) UpdateTask(args *structs.Task, resp *interface{}) error { + return e.Impl.UpdateTask(args) +} + +func (e *ExecutorRPCServerPre0_9_0) DeregisterServices(args interface{}, resp *interface{}) error { + // In 0.6 this is a noop. Goes away in 0.7. + return nil +} + +func (e *ExecutorRPCServerPre0_9_0) Version(args interface{}, version *executorv0.ExecutorVersion) error { + ver, err := e.Impl.Version() + if ver != nil { + *version = *ver + } + return err +} + +func (e *ExecutorRPCServerPre0_9_0) Stats(args interface{}, resourceUsage *cstructs.TaskResourceUsage) error { + ru, err := e.Impl.Stats() + if ru != nil { + *resourceUsage = *ru + } + return err +} + +func (e *ExecutorRPCServerPre0_9_0) Signal(args os.Signal, resp *interface{}) error { + return e.Impl.Signal(args) +} + +func (e *ExecutorRPCServerPre0_9_0) Exec(args ExecCmdArgs, result *ExecCmdReturn) error { + out, code, err := e.Impl.Exec(args.Deadline, args.Name, args.Args) + ret := &ExecCmdReturn{ + Output: out, + Code: code, + } + *result = *ret + return err +} + +type ExecutorPluginPre0_9_0 struct { + logger *log.Logger + Impl *ExecutorRPCServerPre0_9_0 +} + +func (p *ExecutorPluginPre0_9_0) Server(*plugin.MuxBroker) (interface{}, error) { + if p.Impl == nil { + p.Impl = &ExecutorRPCServerPre0_9_0{Impl: executorv0.NewExecutor(p.logger), logger: p.logger} + } + return p.Impl, nil +} + +func (p *ExecutorPluginPre0_9_0) Client(b *plugin.MuxBroker, c *rpc.Client) (interface{}, error) { + return &ExecutorRPCPre0_9_0{client: c, logger: p.logger}, nil +} +*/ diff --git a/client/driver/java.go b/client/driver/java.go index e8cbe4813..470463030 100644 --- a/client/driver/java.go +++ b/client/driver/java.go @@ -55,14 +55,14 @@ type JavaDriverConfig struct { // javaHandle is returned from Start/Open as a handle to the PID type javaHandle struct { - pluginClient *plugin.Client - userPid int - executor executor.Executor - isolationConfig *dstructs.IsolationConfig - taskDir string + pluginClient *plugin.Client + userPid int + executor executor.Executor + taskDir string killTimeout time.Duration maxKillTimeout time.Duration + shutdownSignal string version string logger *log.Logger waitCh chan *dstructs.WaitResult @@ -239,8 +239,9 @@ func (d *JavaDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse pluginLogFile := filepath.Join(ctx.TaskDir.Dir, "executor.out") executorConfig := &dstructs.ExecutorConfig{ - LogFile: pluginLogFile, - LogLevel: d.config.LogLevel, + LogFile: pluginLogFile, + LogLevel: d.config.LogLevel, + FSIsolation: true, } execIntf, pluginClient, err := createExecutor(d.config.LogOutput, d.config, executorConfig) @@ -248,25 +249,12 @@ func (d *JavaDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse return nil, err } - // Set the context - executorCtx := &executor.ExecutorContext{ - TaskEnv: ctx.TaskEnv, - Driver: "java", - Task: task, - TaskDir: ctx.TaskDir.Dir, - LogDir: ctx.TaskDir.LogDir, - } - if err := execIntf.SetContext(executorCtx); err != nil { - pluginClient.Kill() - return nil, fmt.Errorf("failed to set executor context: %v", err) - } - absPath, err := GetAbsolutePath("java") if err != nil { return nil, err } - taskKillSignal, err := getTaskKillSignal(task.KillSignal) + _, err = getTaskKillSignal(task.KillSignal) if err != nil { return nil, err } @@ -274,12 +262,20 @@ func (d *JavaDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse execCmd := &executor.ExecCommand{ Cmd: absPath, Args: args, - FSIsolation: true, ResourceLimits: true, User: getExecutorUser(task), - TaskKillSignal: taskKillSignal, + Resources: &executor.Resources{ + CPU: task.Resources.CPU, + MemoryMB: task.Resources.MemoryMB, + IOPS: task.Resources.IOPS, + DiskMB: task.Resources.DiskMB, + }, + Env: ctx.TaskEnv.List(), + TaskDir: ctx.TaskDir.Dir, + StdoutPath: ctx.StdoutFifo, + StderrPath: ctx.StderrFifo, } - ps, err := execIntf.LaunchCmd(execCmd) + ps, err := execIntf.Launch(execCmd) if err != nil { pluginClient.Kill() return nil, err @@ -289,17 +285,17 @@ func (d *JavaDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse // Return a driver handle maxKill := d.DriverContext.config.MaxKillTimeout h := &javaHandle{ - pluginClient: pluginClient, - executor: execIntf, - userPid: ps.Pid, - isolationConfig: ps.IsolationConfig, - taskDir: ctx.TaskDir.Dir, - killTimeout: GetKillTimeout(task.KillTimeout, maxKill), - maxKillTimeout: maxKill, - version: d.config.Version.VersionNumber(), - logger: d.logger, - doneCh: make(chan struct{}), - waitCh: make(chan *dstructs.WaitResult, 1), + pluginClient: pluginClient, + executor: execIntf, + userPid: ps.Pid, + shutdownSignal: task.KillSignal, + taskDir: ctx.TaskDir.Dir, + killTimeout: GetKillTimeout(task.KillTimeout, maxKill), + maxKillTimeout: maxKill, + version: d.config.Version.VersionNumber(), + logger: d.logger, + doneCh: make(chan struct{}), + waitCh: make(chan *dstructs.WaitResult, 1), } go h.run() return &StartResponse{Handle: h}, nil @@ -308,13 +304,13 @@ func (d *JavaDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse func (d *JavaDriver) Cleanup(*ExecContext, *CreatedResources) error { return nil } type javaId struct { - Version string - KillTimeout time.Duration - MaxKillTimeout time.Duration - PluginConfig *PluginReattachConfig - IsolationConfig *dstructs.IsolationConfig - TaskDir string - UserPid int + Version string + KillTimeout time.Duration + MaxKillTimeout time.Duration + PluginConfig *PluginReattachConfig + TaskDir string + UserPid int + ShutdownSignal string } func (d *JavaDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, error) { @@ -334,12 +330,6 @@ func (d *JavaDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, erro if e := destroyPlugin(id.PluginConfig.Pid, id.UserPid); e != nil { merrs.Errors = append(merrs.Errors, fmt.Errorf("error destroying plugin and userpid: %v", e)) } - if id.IsolationConfig != nil { - ePid := pluginConfig.Reattach.Pid - if e := executor.ClientCleanup(id.IsolationConfig, ePid); e != nil { - merrs.Errors = append(merrs.Errors, fmt.Errorf("destroying resource container failed: %v", e)) - } - } return nil, fmt.Errorf("error connecting to plugin: %v", merrs.ErrorOrNil()) } @@ -349,16 +339,16 @@ func (d *JavaDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, erro // Return a driver handle h := &javaHandle{ - pluginClient: pluginClient, - executor: exec, - userPid: id.UserPid, - isolationConfig: id.IsolationConfig, - logger: d.logger, - version: id.Version, - killTimeout: id.KillTimeout, - maxKillTimeout: id.MaxKillTimeout, - doneCh: make(chan struct{}), - waitCh: make(chan *dstructs.WaitResult, 1), + pluginClient: pluginClient, + executor: exec, + userPid: id.UserPid, + shutdownSignal: id.ShutdownSignal, + logger: d.logger, + version: id.Version, + killTimeout: id.KillTimeout, + maxKillTimeout: id.MaxKillTimeout, + doneCh: make(chan struct{}), + waitCh: make(chan *dstructs.WaitResult, 1), } go h.run() return h, nil @@ -366,13 +356,13 @@ func (d *JavaDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, erro func (h *javaHandle) ID() string { id := javaId{ - Version: h.version, - KillTimeout: h.killTimeout, - MaxKillTimeout: h.maxKillTimeout, - PluginConfig: NewPluginReattachConfig(h.pluginClient.ReattachConfig()), - UserPid: h.userPid, - IsolationConfig: h.isolationConfig, - TaskDir: h.taskDir, + Version: h.version, + KillTimeout: h.killTimeout, + MaxKillTimeout: h.maxKillTimeout, + PluginConfig: NewPluginReattachConfig(h.pluginClient.ReattachConfig()), + UserPid: h.userPid, + TaskDir: h.taskDir, + ShutdownSignal: h.shutdownSignal, } data, err := json.Marshal(id) @@ -389,7 +379,12 @@ func (h *javaHandle) WaitCh() chan *dstructs.WaitResult { func (h *javaHandle) Update(task *structs.Task) error { // Store the updated kill timeout. h.killTimeout = GetKillTimeout(task.KillTimeout, h.maxKillTimeout) - h.executor.UpdateTask(task) + h.executor.UpdateResources(&executor.Resources{ + CPU: task.Resources.CPU, + MemoryMB: task.Resources.MemoryMB, + IOPS: task.Resources.IOPS, + DiskMB: task.Resources.DiskMB, + }) // Update is not possible return nil @@ -413,11 +408,11 @@ func (d *javaHandle) Network() *cstructs.DriverNetwork { } func (h *javaHandle) Kill() error { - if err := h.executor.ShutDown(); err != nil { + if err := h.executor.Shutdown(h.shutdownSignal, h.killTimeout); err != nil { if h.pluginClient.Exited() { return nil } - return fmt.Errorf("executor Shutdown failed: %v", err) + return fmt.Errorf("executor Kill failed: %v", err) } select { @@ -426,8 +421,8 @@ func (h *javaHandle) Kill() error { if h.pluginClient.Exited() { break } - if err := h.executor.Exit(); err != nil { - return fmt.Errorf("executor Exit failed: %v", err) + if err := h.executor.Shutdown(h.shutdownSignal, h.killTimeout); err != nil { + return fmt.Errorf("executor Destroy failed: %v", err) } } @@ -442,20 +437,13 @@ func (h *javaHandle) run() { ps, werr := h.executor.Wait() close(h.doneCh) if ps.ExitCode == 0 && werr != nil { - if h.isolationConfig != nil { - ePid := h.pluginClient.ReattachConfig().Pid - if e := executor.ClientCleanup(h.isolationConfig, ePid); e != nil { - h.logger.Printf("[ERR] driver.java: destroying resource container failed: %v", e) - } - } else { - if e := killProcess(h.userPid); e != nil { - h.logger.Printf("[ERR] driver.java: error killing user process: %v", e) - } + if e := killProcess(h.userPid); e != nil { + h.logger.Printf("[ERR] driver.java: error killing user process: %v", e) } } - // Exit the executor - h.executor.Exit() + // Destroy the executor + h.executor.Shutdown(h.shutdownSignal, 0) h.pluginClient.Kill() // Send the results diff --git a/client/driver/java_test.go b/client/driver/java_test.go index 2c8a5a9e0..827f1bed2 100644 --- a/client/driver/java_test.go +++ b/client/driver/java_test.go @@ -43,7 +43,7 @@ func TestJavaDriver_Fingerprint(t *testing.T) { Resources: structs.DefaultResources(), } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewJavaDriver(ctx.DriverCtx) node := &structs.Node{ Attributes: map[string]string{ @@ -100,7 +100,7 @@ func TestJavaDriver_StartOpen_Wait(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewJavaDriver(ctx.DriverCtx) // Copy the test jar into the task's directory @@ -156,7 +156,7 @@ func TestJavaDriver_Start_Wait(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewJavaDriver(ctx.DriverCtx) // Copy the test jar into the task's directory @@ -222,7 +222,7 @@ func TestJavaDriver_Start_Kill_Wait(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewJavaDriver(ctx.DriverCtx) // Copy the test jar into the task's directory @@ -288,7 +288,7 @@ func TestJavaDriver_Signal(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewJavaDriver(ctx.DriverCtx) // Copy the test jar into the task's directory @@ -357,7 +357,7 @@ func TestJavaDriver_User(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewJavaDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -399,7 +399,7 @@ func TestJavaDriver_Start_Wait_Class(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewJavaDriver(ctx.DriverCtx) // Copy the test jar into the task's directory @@ -469,7 +469,7 @@ func TestJavaDriver_Start_Kill(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewJavaDriver(ctx.DriverCtx) // Copy the test jar into the task's directory @@ -506,7 +506,7 @@ func TestJavaDriver_Start_Kill(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewJavaDriver(ctx.DriverCtx) // Copy the test jar into the task's directory diff --git a/client/driver/logging/rotator.go b/client/driver/logging/rotator.go index 152ab5508..7b579a293 100644 --- a/client/driver/logging/rotator.go +++ b/client/driver/logging/rotator.go @@ -5,7 +5,6 @@ import ( "bytes" "fmt" "io/ioutil" - "log" "os" "path/filepath" "sort" @@ -13,6 +12,8 @@ import ( "strings" "sync" "time" + + hclog "github.com/hashicorp/go-hclog" ) const ( @@ -48,7 +49,7 @@ type FileRotator struct { bufLock sync.Mutex flushTicker *time.Ticker - logger *log.Logger + logger hclog.Logger purgeCh chan struct{} doneCh chan struct{} @@ -58,7 +59,8 @@ type FileRotator struct { // NewFileRotator returns a new file rotator func NewFileRotator(path string, baseFile string, maxFiles int, - fileSize int64, logger *log.Logger) (*FileRotator, error) { + fileSize int64, logger hclog.Logger) (*FileRotator, error) { + logger = logger.Named("rotator") rotator := &FileRotator{ MaxFiles: maxFiles, FileSize: fileSize, @@ -71,6 +73,7 @@ func NewFileRotator(path string, baseFile string, maxFiles int, purgeCh: make(chan struct{}, 1), doneCh: make(chan struct{}, 1), } + if err := rotator.lastFile(); err != nil { return nil, err } @@ -93,7 +96,7 @@ func (f *FileRotator) Write(p []byte) (n int, err error) { f.flushBuffer() f.currentFile.Close() if err := f.nextFile(); err != nil { - f.logger.Printf("[ERROR] driver.rotator: error creating next file: %v", err) + f.logger.Error("error creating next file", "err", err) return 0, err } } @@ -141,7 +144,7 @@ func (f *FileRotator) Write(p []byte) (n int, err error) { // Increment the total number of bytes in the file f.currentWr += int64(n) if err != nil { - f.logger.Printf("[ERROR] driver.rotator: error writing to file: %v", err) + f.logger.Error("error writing to file", "err", err) return } } @@ -210,10 +213,11 @@ func (f *FileRotator) lastFile() error { // createFile opens a new or existing file for writing func (f *FileRotator) createFile() error { logFileName := filepath.Join(f.path, fmt.Sprintf("%s.%d", f.baseFileName, f.logFileIdx)) - cFile, err := os.OpenFile(logFileName, os.O_RDWR|os.O_CREATE|os.O_APPEND, 0666) + cFile, err := os.OpenFile(logFileName, os.O_RDWR|os.O_CREATE|os.O_APPEND, 0644) if err != nil { return err } + f.currentFile = cFile fi, err := f.currentFile.Stat() if err != nil { @@ -257,7 +261,7 @@ func (f *FileRotator) purgeOldFiles() { var fIndexes []int files, err := ioutil.ReadDir(f.path) if err != nil { - f.logger.Printf("[ERROR] driver.rotator: error getting directory listing: %v", err) + f.logger.Error("error getting directory listing", "err", err) return } // Inserting all the rotated files in a slice @@ -266,7 +270,7 @@ func (f *FileRotator) purgeOldFiles() { fileIdx := strings.TrimPrefix(fi.Name(), fmt.Sprintf("%s.", f.baseFileName)) n, err := strconv.Atoi(fileIdx) if err != nil { - f.logger.Printf("[ERROR] driver.rotator: error extracting file index: %v", err) + f.logger.Error("error extracting file index", "err", err) continue } fIndexes = append(fIndexes, n) @@ -287,7 +291,7 @@ func (f *FileRotator) purgeOldFiles() { fname := filepath.Join(f.path, fmt.Sprintf("%s.%d", f.baseFileName, fIndex)) err := os.RemoveAll(fname) if err != nil { - f.logger.Printf("[ERROR] driver.rotator: error removing file: %v", err) + f.logger.Error("error removing file", "filename", fname, "err", err) } } f.oldestLogFileIdx = fIndexes[0] diff --git a/client/driver/logging/syslog_parser.go b/client/driver/logging/syslog_parser.go index b12e97e0b..09f1adcfa 100644 --- a/client/driver/logging/syslog_parser.go +++ b/client/driver/logging/syslog_parser.go @@ -4,10 +4,10 @@ package logging import ( "fmt" - "log" "strconv" syslog "github.com/RackSec/srslog" + hclog "github.com/hashicorp/go-hclog" ) // Errors related to parsing priority @@ -41,11 +41,11 @@ type Priority struct { // DockerLogParser parses a line of log message that the docker daemon ships type DockerLogParser struct { - logger *log.Logger + logger hclog.Logger } // NewDockerLogParser creates a new DockerLogParser -func NewDockerLogParser(logger *log.Logger) *DockerLogParser { +func NewDockerLogParser(logger hclog.Logger) *DockerLogParser { return &DockerLogParser{logger: logger} } diff --git a/client/driver/logging/syslog_server.go b/client/driver/logging/syslog_server.go index b8ab8e68e..d7c7a5283 100644 --- a/client/driver/logging/syslog_server.go +++ b/client/driver/logging/syslog_server.go @@ -2,9 +2,10 @@ package logging import ( "bufio" - "log" "net" "sync" + + hclog "github.com/hashicorp/go-hclog" ) // SyslogServer is a server which listens to syslog messages and parses them @@ -17,11 +18,12 @@ type SyslogServer struct { done bool doneLock sync.Mutex - logger *log.Logger + logger hclog.Logger } // NewSyslogServer creates a new syslog server -func NewSyslogServer(l net.Listener, messages chan *SyslogMessage, logger *log.Logger) *SyslogServer { +func NewSyslogServer(l net.Listener, messages chan *SyslogMessage, logger hclog.Logger) *SyslogServer { + logger = logger.Named("logcollector.server") parser := NewDockerLogParser(logger) return &SyslogServer{ listener: l, @@ -48,7 +50,7 @@ func (s *SyslogServer) Start() { return } - s.logger.Printf("[ERR] logcollector.server: error in accepting connection: %v", err) + s.logger.Error("error in accepting connection", "err", err) continue } go s.read(connection) diff --git a/client/driver/logging/universal_collector.go b/client/driver/logging/universal_collector.go index 54d8b36ff..e3399228f 100644 --- a/client/driver/logging/universal_collector.go +++ b/client/driver/logging/universal_collector.go @@ -6,14 +6,13 @@ import ( "fmt" "io" "io/ioutil" - "log" "net" "os" "runtime" syslog "github.com/RackSec/srslog" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" - cstructs "github.com/hashicorp/nomad/client/driver/structs" "github.com/hashicorp/nomad/nomad/structs" ) @@ -41,8 +40,7 @@ type LogCollectorContext struct { // SyslogCollectorState holds the address and isolation information of a launched // syslog server type SyslogCollectorState struct { - IsolationConfig *cstructs.IsolationConfig - Addr string + Addr string } // LogCollector is an interface which allows a driver to launch a log server @@ -64,12 +62,13 @@ type SyslogCollector struct { syslogChan chan *SyslogMessage taskDir string - logger *log.Logger + logger hclog.Logger } // NewSyslogCollector returns an implementation of the SyslogCollector -func NewSyslogCollector(logger *log.Logger) *SyslogCollector { - return &SyslogCollector{logger: logger, syslogChan: make(chan *SyslogMessage, 2048)} +func NewSyslogCollector(logger hclog.Logger) *SyslogCollector { + return &SyslogCollector{logger: logger.Named("syslog-server"), + syslogChan: make(chan *SyslogMessage, 2048)} } // LaunchCollector launches a new syslog server and starts writing log lines to @@ -79,7 +78,7 @@ func (s *SyslogCollector) LaunchCollector(ctx *LogCollectorContext) (*SyslogColl if err != nil { return nil, err } - s.logger.Printf("[DEBUG] syslog-server: launching syslog server on addr: %v", l.Addr().String()) + s.logger.Debug("launching syslog server on addr", "addr", l.Addr().String()) s.ctx = ctx // configuring the task dir if err := s.configureTaskDir(); err != nil { diff --git a/client/driver/lxc_test.go b/client/driver/lxc_test.go index 6fc203a43..898982b7b 100644 --- a/client/driver/lxc_test.go +++ b/client/driver/lxc_test.go @@ -33,7 +33,7 @@ func TestLxcDriver_Fingerprint(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewLxcDriver(ctx.DriverCtx) node := &structs.Node{ @@ -104,7 +104,7 @@ func TestLxcDriver_Start_Wait(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewLxcDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -191,7 +191,7 @@ func TestLxcDriver_Open_Wait(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewLxcDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -290,7 +290,7 @@ func testVolumeConfig(t *testing.T, volConfig []string) error { task.Config["volumes"] = volConfig ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() driver := NewLxcDriver(ctx.DriverCtx) @@ -320,7 +320,7 @@ func TestLxcDriver_Start_NoVolumes(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() // set lxcVolumesConfigOption to false to disallow absolute paths as the source for the bind mount ctx.DriverCtx.config.Options = map[string]string{lxcVolumesConfigOption: "false"} diff --git a/client/driver/mock_driver.go b/client/driver/mock_driver.go index f3b1f2ede..5f2436ea5 100644 --- a/client/driver/mock_driver.go +++ b/client/driver/mock_driver.go @@ -14,6 +14,7 @@ import ( "github.com/mitchellh/mapstructure" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/driver/logging" dstructs "github.com/hashicorp/nomad/client/driver/structs" cstructs "github.com/hashicorp/nomad/client/structs" @@ -447,7 +448,7 @@ func (h *mockDriverHandle) handleLogging() { // Setup a log rotator logFileSize := int64(h.task.LogConfig.MaxFileSizeMB * 1024 * 1024) lro, err := logging.NewFileRotator(h.ctx.TaskDir.LogDir, fmt.Sprintf("%v.stdout", h.taskName), - h.task.LogConfig.MaxFiles, logFileSize, h.logger) + h.task.LogConfig.MaxFiles, logFileSize, hclog.Default()) //TODO: plumb hclog if err != nil { h.exitErr = err close(h.doneCh) diff --git a/client/driver/plugins.go b/client/driver/plugins.go index f29758491..7385a18e1 100644 --- a/client/driver/plugins.go +++ b/client/driver/plugins.go @@ -2,12 +2,10 @@ package driver import ( "io" - "log" "net" - "strings" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/go-plugin" - "github.com/hashicorp/logutils" ) var HandshakeConfig = plugin.HandshakeConfig{ @@ -16,15 +14,15 @@ var HandshakeConfig = plugin.HandshakeConfig{ MagicCookieValue: "e4327c2e01eabfd75a8a67adb114fb34a757d57eee7728d857a8cec6e91a7255", } -func GetPluginMap(w io.Writer, logLevel string) map[string]plugin.Plugin { +func GetPluginMap(w io.Writer, logLevel hclog.Level, fsIsolation bool) map[string]plugin.Plugin { e := new(ExecutorPlugin) - filter := &logutils.LevelFilter{ - Levels: []logutils.LogLevel{"TRACE", "DEBUG", "INFO", "WARN", "ERR"}, - MinLevel: logutils.LogLevel(strings.ToUpper(logLevel)), - Writer: w, - } - e.logger = log.New(filter, "", log.LstdFlags|log.Lmicroseconds) + e.logger = hclog.New(&hclog.LoggerOptions{ + Output: w, + Level: logLevel, + }) + + e.fsIsolation = fsIsolation return map[string]plugin.Plugin{ "executor": e, diff --git a/client/driver/qemu.go b/client/driver/qemu.go index dcbc0d0ec..e325235b2 100644 --- a/client/driver/qemu.go +++ b/client/driver/qemu.go @@ -75,6 +75,7 @@ type qemuHandle struct { userPid int executor executor.Executor monitorPath string + shutdownSignal string killTimeout time.Duration maxKillTimeout time.Duration logger *log.Logger @@ -319,24 +320,22 @@ func (d *QemuDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse if err != nil { return nil, err } - executorCtx := &executor.ExecutorContext{ - TaskEnv: ctx.TaskEnv, - Driver: "qemu", - Task: task, - TaskDir: ctx.TaskDir.Dir, - LogDir: ctx.TaskDir.LogDir, - } - if err := exec.SetContext(executorCtx); err != nil { - pluginClient.Kill() - return nil, fmt.Errorf("failed to set executor context: %v", err) + + _, err = getTaskKillSignal(task.KillSignal) + if err != nil { + return nil, err } execCmd := &executor.ExecCommand{ - Cmd: args[0], - Args: args[1:], - User: task.User, + Cmd: args[0], + Args: args[1:], + User: task.User, + TaskDir: ctx.TaskDir.Dir, + Env: ctx.TaskEnv.List(), + StdoutPath: ctx.StdoutFifo, + StderrPath: ctx.StderrFifo, } - ps, err := exec.LaunchCmd(execCmd) + ps, err := exec.Launch(execCmd) if err != nil { pluginClient.Kill() return nil, err @@ -349,6 +348,7 @@ func (d *QemuDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse pluginClient: pluginClient, executor: exec, userPid: ps.Pid, + shutdownSignal: task.KillSignal, killTimeout: GetKillTimeout(task.KillTimeout, maxKill), maxKillTimeout: maxKill, monitorPath: monitorPath, @@ -373,6 +373,7 @@ type qemuId struct { MaxKillTimeout time.Duration UserPid int PluginConfig *PluginReattachConfig + ShutdownSignal string } func (d *QemuDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, error) { @@ -404,6 +405,7 @@ func (d *QemuDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, erro logger: d.logger, killTimeout: id.KillTimeout, maxKillTimeout: id.MaxKillTimeout, + shutdownSignal: id.ShutdownSignal, version: id.Version, doneCh: make(chan struct{}), waitCh: make(chan *dstructs.WaitResult, 1), @@ -421,6 +423,7 @@ func (h *qemuHandle) ID() string { MaxKillTimeout: h.maxKillTimeout, PluginConfig: NewPluginReattachConfig(h.pluginClient.ReattachConfig()), UserPid: h.userPid, + ShutdownSignal: h.shutdownSignal, } data, err := json.Marshal(id) @@ -437,7 +440,6 @@ func (h *qemuHandle) WaitCh() chan *dstructs.WaitResult { func (h *qemuHandle) Update(task *structs.Task) error { // Store the updated kill timeout. h.killTimeout = GetKillTimeout(task.KillTimeout, h.maxKillTimeout) - h.executor.UpdateTask(task) // Update is not possible return nil @@ -470,7 +472,7 @@ func (h *qemuHandle) Kill() error { // the qemu process as a last resort if gracefulShutdownSent == false { h.logger.Printf("[DEBUG] driver.qemu: graceful shutdown is not enabled, sending an interrupt signal to pid: %d", h.userPid) - if err := h.executor.ShutDown(); err != nil { + if err := h.executor.Signal(os.Interrupt); err != nil { if h.pluginClient.Exited() { return nil } @@ -490,8 +492,8 @@ func (h *qemuHandle) Kill() error { if h.pluginClient.Exited() { return nil } - if err := h.executor.Exit(); err != nil { - return fmt.Errorf("executor Exit failed: %v", err) + if err := h.executor.Shutdown(h.shutdownSignal, h.killTimeout); err != nil { + return fmt.Errorf("executor Destroy failed: %v", err) } return nil } @@ -510,8 +512,8 @@ func (h *qemuHandle) run() { } close(h.doneCh) - // Exit the executor - h.executor.Exit() + // Destroy the executor + h.executor.Shutdown(h.shutdownSignal, 0) h.pluginClient.Kill() // Send the results diff --git a/client/driver/qemu_test.go b/client/driver/qemu_test.go index 51f021068..e8fc57fe5 100644 --- a/client/driver/qemu_test.go +++ b/client/driver/qemu_test.go @@ -33,7 +33,7 @@ func TestQemuDriver_Fingerprint(t *testing.T) { Resources: structs.DefaultResources(), } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewQemuDriver(ctx.DriverCtx) node := &structs.Node{ @@ -100,7 +100,7 @@ func TestQemuDriver_StartOpen_Wait(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewQemuDriver(ctx.DriverCtx) // Copy the test image into the task's directory @@ -181,7 +181,7 @@ func TestQemuDriver_GracefulShutdown(t *testing.T) { ctx := testDriverContexts(t, task) ctx.DriverCtx.config.MaxKillTimeout = killTimeout - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewQemuDriver(ctx.DriverCtx) request := &cstructs.FingerprintRequest{Config: &config.Config{}, Node: ctx.DriverCtx.node} @@ -340,7 +340,7 @@ func TestQemuDriverUser(t *testing.T) { for _, task := range tasks { ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewQemuDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -391,7 +391,7 @@ func TestQemuDriverGetMonitorPathOldQemu(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() // Simulate an older version of qemu which does not support long monitor socket paths ctx.DriverCtx.node.Attributes[qemuDriverVersionAttr] = "2.0.0" @@ -450,7 +450,7 @@ func TestQemuDriverGetMonitorPathNewQemu(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() // Simulate a version of qemu which supports long monitor socket paths ctx.DriverCtx.node.Attributes[qemuDriverVersionAttr] = "2.99.99" diff --git a/client/driver/raw_exec.go b/client/driver/raw_exec.go index 68aff51f8..2001b423d 100644 --- a/client/driver/raw_exec.go +++ b/client/driver/raw_exec.go @@ -50,18 +50,18 @@ type RawExecDriver struct { // rawExecHandle is returned from Start/Open as a handle to the PID type rawExecHandle struct { - version string - pluginClient *plugin.Client - userPid int - executor executor.Executor - isolationConfig *dstructs.IsolationConfig - killTimeout time.Duration - maxKillTimeout time.Duration - logger *log.Logger - waitCh chan *dstructs.WaitResult - doneCh chan struct{} - taskEnv *env.TaskEnv - taskDir *allocdir.TaskDir + version string + pluginClient *plugin.Client + userPid int + executor executor.Executor + killTimeout time.Duration + maxKillTimeout time.Duration + shutdownSignal string + logger *log.Logger + waitCh chan *dstructs.WaitResult + doneCh chan struct{} + taskEnv *env.TaskEnv + taskDir *allocdir.TaskDir } // NewRawExecDriver is used to create a new raw exec driver @@ -151,31 +151,23 @@ func (d *RawExecDriver) Start(ctx *ExecContext, task *structs.Task) (*StartRespo if err != nil { return nil, err } - executorCtx := &executor.ExecutorContext{ - TaskEnv: ctx.TaskEnv, - Driver: "raw_exec", - Task: task, - TaskDir: ctx.TaskDir.Dir, - LogDir: ctx.TaskDir.LogDir, - } - if err := exec.SetContext(executorCtx); err != nil { - pluginClient.Kill() - return nil, fmt.Errorf("failed to set executor context: %v", err) - } - taskKillSignal, err := getTaskKillSignal(task.KillSignal) + _, err = getTaskKillSignal(task.KillSignal) if err != nil { return nil, err } execCmd := &executor.ExecCommand{ Cmd: command, - Args: driverConfig.Args, + Args: ctx.TaskEnv.ParseAndReplace(driverConfig.Args), User: task.User, - TaskKillSignal: taskKillSignal, BasicProcessCgroup: d.useCgroup, + Env: ctx.TaskEnv.List(), + TaskDir: ctx.TaskDir.Dir, + StdoutPath: ctx.StdoutFifo, + StderrPath: ctx.StderrFifo, } - ps, err := exec.LaunchCmd(execCmd) + ps, err := exec.Launch(execCmd) if err != nil { pluginClient.Kill() return nil, err @@ -185,18 +177,18 @@ func (d *RawExecDriver) Start(ctx *ExecContext, task *structs.Task) (*StartRespo // Return a driver handle maxKill := d.DriverContext.config.MaxKillTimeout h := &rawExecHandle{ - pluginClient: pluginClient, - executor: exec, - isolationConfig: ps.IsolationConfig, - userPid: ps.Pid, - killTimeout: GetKillTimeout(task.KillTimeout, maxKill), - maxKillTimeout: maxKill, - version: d.config.Version.VersionNumber(), - logger: d.logger, - doneCh: make(chan struct{}), - waitCh: make(chan *dstructs.WaitResult, 1), - taskEnv: ctx.TaskEnv, - taskDir: ctx.TaskDir, + pluginClient: pluginClient, + executor: exec, + userPid: ps.Pid, + shutdownSignal: task.KillSignal, + killTimeout: GetKillTimeout(task.KillTimeout, maxKill), + maxKillTimeout: maxKill, + version: d.config.Version.VersionNumber(), + logger: d.logger, + doneCh: make(chan struct{}), + waitCh: make(chan *dstructs.WaitResult, 1), + taskEnv: ctx.TaskEnv, + taskDir: ctx.TaskDir, } go h.run() return &StartResponse{Handle: h}, nil @@ -205,12 +197,12 @@ func (d *RawExecDriver) Start(ctx *ExecContext, task *structs.Task) (*StartRespo func (d *RawExecDriver) Cleanup(*ExecContext, *CreatedResources) error { return nil } type rawExecId struct { - Version string - KillTimeout time.Duration - MaxKillTimeout time.Duration - UserPid int - PluginConfig *PluginReattachConfig - IsolationConfig *dstructs.IsolationConfig + Version string + KillTimeout time.Duration + MaxKillTimeout time.Duration + UserPid int + PluginConfig *PluginReattachConfig + ShutdownSignal string } func (d *RawExecDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, error) { @@ -230,12 +222,6 @@ func (d *RawExecDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, e if e := destroyPlugin(id.PluginConfig.Pid, id.UserPid); e != nil { merrs.Errors = append(merrs.Errors, fmt.Errorf("error destroying plugin and userpid: %v", e)) } - if id.IsolationConfig != nil { - ePid := pluginConfig.Reattach.Pid - if e := executor.ClientCleanup(id.IsolationConfig, ePid); e != nil { - merrs.Errors = append(merrs.Errors, fmt.Errorf("destroying resource container failed: %v", e)) - } - } return nil, fmt.Errorf("error connecting to plugin: %v", merrs.ErrorOrNil()) } @@ -244,18 +230,18 @@ func (d *RawExecDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, e // Return a driver handle h := &rawExecHandle{ - pluginClient: pluginClient, - executor: exec, - userPid: id.UserPid, - isolationConfig: id.IsolationConfig, - logger: d.logger, - killTimeout: id.KillTimeout, - maxKillTimeout: id.MaxKillTimeout, - version: id.Version, - doneCh: make(chan struct{}), - waitCh: make(chan *dstructs.WaitResult, 1), - taskEnv: ctx.TaskEnv, - taskDir: ctx.TaskDir, + pluginClient: pluginClient, + executor: exec, + userPid: id.UserPid, + logger: d.logger, + shutdownSignal: id.ShutdownSignal, + killTimeout: id.KillTimeout, + maxKillTimeout: id.MaxKillTimeout, + version: id.Version, + doneCh: make(chan struct{}), + waitCh: make(chan *dstructs.WaitResult, 1), + taskEnv: ctx.TaskEnv, + taskDir: ctx.TaskDir, } go h.run() return h, nil @@ -263,12 +249,12 @@ func (d *RawExecDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, e func (h *rawExecHandle) ID() string { id := rawExecId{ - Version: h.version, - KillTimeout: h.killTimeout, - MaxKillTimeout: h.maxKillTimeout, - PluginConfig: NewPluginReattachConfig(h.pluginClient.ReattachConfig()), - UserPid: h.userPid, - IsolationConfig: h.isolationConfig, + Version: h.version, + KillTimeout: h.killTimeout, + MaxKillTimeout: h.maxKillTimeout, + PluginConfig: NewPluginReattachConfig(h.pluginClient.ReattachConfig()), + UserPid: h.userPid, + ShutdownSignal: h.shutdownSignal, } data, err := json.Marshal(id) @@ -285,14 +271,13 @@ func (h *rawExecHandle) WaitCh() chan *dstructs.WaitResult { func (h *rawExecHandle) Update(task *structs.Task) error { // Store the updated kill timeout. h.killTimeout = GetKillTimeout(task.KillTimeout, h.maxKillTimeout) - h.executor.UpdateTask(task) // Update is not possible return nil } func (h *rawExecHandle) Exec(ctx context.Context, cmd string, args []string) ([]byte, int, error) { - return executor.ExecScript(ctx, h.taskDir.Dir, h.taskEnv, nil, cmd, args) + return executor.ExecScript(ctx, h.taskDir.Dir, h.taskEnv.List(), nil, h.taskEnv.ReplaceEnv(cmd), h.taskEnv.ParseAndReplace(args)) } func (h *rawExecHandle) Signal(s os.Signal) error { @@ -304,7 +289,7 @@ func (d *rawExecHandle) Network() *cstructs.DriverNetwork { } func (h *rawExecHandle) Kill() error { - if err := h.executor.ShutDown(); err != nil { + if err := h.executor.Signal(os.Interrupt); err != nil { if h.pluginClient.Exited() { return nil } @@ -318,7 +303,7 @@ func (h *rawExecHandle) Kill() error { if h.pluginClient.Exited() { return nil } - if err := h.executor.Exit(); err != nil { + if err := h.executor.Shutdown(h.shutdownSignal, h.killTimeout); err != nil { return fmt.Errorf("executor Exit failed: %v", err) } @@ -334,20 +319,13 @@ func (h *rawExecHandle) run() { ps, werr := h.executor.Wait() close(h.doneCh) if ps.ExitCode == 0 && werr != nil { - if h.isolationConfig != nil { - ePid := h.pluginClient.ReattachConfig().Pid - if e := executor.ClientCleanup(h.isolationConfig, ePid); e != nil { - h.logger.Printf("[ERR] driver.raw_exec: destroying resource container failed: %v", e) - } - } else { - if e := killProcess(h.userPid); e != nil { - h.logger.Printf("[ERR] driver.raw_exec: error killing user process: %v", e) - } + if e := killProcess(h.userPid); e != nil { + h.logger.Printf("[ERR] driver.raw_exec: error killing user process: %v", e) } } - // Exit the executor - if err := h.executor.Exit(); err != nil { + // Destroy the executor + if err := h.executor.Shutdown(h.shutdownSignal, 0); err != nil { h.logger.Printf("[ERR] driver.raw_exec: error killing executor: %v", err) } h.pluginClient.Kill() diff --git a/client/driver/raw_exec_test.go b/client/driver/raw_exec_test.go index c47319b07..990045de3 100644 --- a/client/driver/raw_exec_test.go +++ b/client/driver/raw_exec_test.go @@ -30,7 +30,7 @@ func TestRawExecDriver_Fingerprint(t *testing.T) { Resources: structs.DefaultResources(), } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRawExecDriver(ctx.DriverCtx) node := &structs.Node{ Attributes: make(map[string]string), @@ -83,7 +83,7 @@ func TestRawExecDriver_StartOpen_Wait(t *testing.T) { } testtask.SetTaskEnv(task) ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRawExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -130,7 +130,7 @@ func TestRawExecDriver_Start_Wait(t *testing.T) { } testtask.SetTaskEnv(task) ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRawExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -182,7 +182,7 @@ func TestRawExecDriver_Start_Wait_AllocDir(t *testing.T) { testtask.SetTaskEnv(task) ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRawExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -233,7 +233,7 @@ func TestRawExecDriver_Start_Kill_Wait(t *testing.T) { testtask.SetTaskEnv(task) ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRawExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -290,7 +290,7 @@ func TestRawExecDriver_Start_Kill_Wait_Cgroup(t *testing.T) { ctx := testDriverContexts(t, task) ctx.DriverCtx.node.Attributes["unique.cgroup.mountpoint"] = "foo" // Enable cgroups - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRawExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -377,7 +377,7 @@ func TestRawExecDriver_HandlerExec(t *testing.T) { } testtask.SetTaskEnv(task) ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRawExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { diff --git a/client/driver/raw_exec_unix_test.go b/client/driver/raw_exec_unix_test.go index f307d55d2..863da540d 100644 --- a/client/driver/raw_exec_unix_test.go +++ b/client/driver/raw_exec_unix_test.go @@ -3,6 +3,7 @@ package driver import ( + "fmt" "io/ioutil" "path/filepath" "runtime" @@ -14,6 +15,7 @@ import ( "github.com/hashicorp/nomad/helper/testtask" "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/testutil" + "github.com/stretchr/testify/require" ) func TestRawExecDriver_User(t *testing.T) { @@ -38,7 +40,7 @@ func TestRawExecDriver_User(t *testing.T) { testtask.SetTaskEnv(task) ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRawExecDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -73,7 +75,7 @@ func TestRawExecDriver_Signal(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRawExecDriver(ctx.DriverCtx) testFile := filepath.Join(ctx.ExecCtx.TaskDir.Dir, "test.sh") @@ -119,14 +121,17 @@ done // Check the log file to see it exited because of the signal outputFile := filepath.Join(ctx.ExecCtx.TaskDir.LogDir, "signal.stdout.0") - act, err := ioutil.ReadFile(outputFile) - if err != nil { - t.Fatalf("Couldn't read expected output: %v", err) - } - exp := "Terminated." - if strings.TrimSpace(string(act)) != exp { - t.Logf("Read from %v", outputFile) - t.Fatalf("Command outputted %v; want %v", act, exp) - } + testutil.WaitForResult(func() (bool, error) { + act, err := ioutil.ReadFile(outputFile) + if err != nil { + return false, fmt.Errorf("Couldn't read expected output: %v", err) + } + + if strings.TrimSpace(string(act)) != exp { + t.Logf("Read from %v", outputFile) + return false, fmt.Errorf("Command outputted %v; want %v", act, exp) + } + return true, nil + }, func(err error) { require.NoError(t, err) }) } diff --git a/client/driver/rkt.go b/client/driver/rkt.go index 1d297000f..17464425a 100644 --- a/client/driver/rkt.go +++ b/client/driver/rkt.go @@ -105,6 +105,7 @@ type rktHandle struct { logger *log.Logger killTimeout time.Duration maxKillTimeout time.Duration + shutdownSignal string waitCh chan *dstructs.WaitResult doneCh chan struct{} } @@ -117,6 +118,7 @@ type rktPID struct { ExecutorPid int KillTimeout time.Duration MaxKillTimeout time.Duration + ShutdownSignal string } // Retrieve pod status for the pod with the given UUID. @@ -656,16 +658,10 @@ func (d *RktDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse, eb := env.NewEmptyBuilder() filter := strings.Split(d.config.ReadDefault("env.blacklist", config.DefaultEnvBlacklist), ",") rktEnv := eb.SetHostEnvvars(filter).Build() - executorCtx := &executor.ExecutorContext{ - TaskEnv: rktEnv, - Driver: "rkt", - Task: task, - TaskDir: ctx.TaskDir.Dir, - LogDir: ctx.TaskDir.LogDir, - } - if err := execIntf.SetContext(executorCtx); err != nil { - pluginClient.Kill() - return nil, fmt.Errorf("failed to set executor context: %v", err) + + _, err = getTaskKillSignal(task.KillSignal) + if err != nil { + return nil, err } // Enable ResourceLimits to place the executor in a parent cgroup of @@ -675,8 +671,18 @@ func (d *RktDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse, Cmd: absPath, Args: runArgs, ResourceLimits: true, + Resources: &executor.Resources{ + CPU: task.Resources.CPU, + MemoryMB: task.Resources.MemoryMB, + IOPS: task.Resources.IOPS, + DiskMB: task.Resources.DiskMB, + }, + Env: ctx.TaskEnv.List(), + TaskDir: ctx.TaskDir.Dir, + StdoutPath: ctx.StdoutFifo, + StderrPath: ctx.StderrFifo, } - ps, err := execIntf.LaunchCmd(execCmd) + ps, err := execIntf.Launch(execCmd) if err != nil { pluginClient.Kill() return nil, err @@ -694,6 +700,7 @@ func (d *RktDriver) Start(ctx *ExecContext, task *structs.Task) (*StartResponse, logger: d.logger, killTimeout: GetKillTimeout(task.KillTimeout, maxKill), maxKillTimeout: maxKill, + shutdownSignal: task.KillSignal, doneCh: make(chan struct{}), waitCh: make(chan *dstructs.WaitResult, 1), } @@ -762,6 +769,7 @@ func (d *RktDriver) Open(ctx *ExecContext, handleID string) (DriverHandle, error logger: d.logger, killTimeout: id.KillTimeout, maxKillTimeout: id.MaxKillTimeout, + shutdownSignal: id.ShutdownSignal, doneCh: make(chan struct{}), waitCh: make(chan *dstructs.WaitResult, 1), } @@ -777,6 +785,7 @@ func (h *rktHandle) ID() string { KillTimeout: h.killTimeout, MaxKillTimeout: h.maxKillTimeout, ExecutorPid: h.executorPid, + ShutdownSignal: h.shutdownSignal, } data, err := json.Marshal(pid) if err != nil { @@ -792,7 +801,6 @@ func (h *rktHandle) WaitCh() chan *dstructs.WaitResult { func (h *rktHandle) Update(task *structs.Task) error { // Store the updated kill timeout. h.killTimeout = GetKillTimeout(task.KillTimeout, h.maxKillTimeout) - h.executor.UpdateTask(task) // Update is not possible return nil @@ -806,9 +814,9 @@ func (h *rktHandle) Exec(ctx context.Context, cmd string, args []string) ([]byte enterArgs := make([]string, 3+len(args)) enterArgs[0] = "enter" enterArgs[1] = h.uuid - enterArgs[2] = cmd - copy(enterArgs[3:], args) - return executor.ExecScript(ctx, h.taskDir.Dir, h.env, nil, rktCmd, enterArgs) + enterArgs[2] = h.env.ReplaceEnv(cmd) + copy(enterArgs[3:], h.env.ParseAndReplace(args)) + return executor.ExecScript(ctx, h.taskDir.Dir, h.env.List(), nil, rktCmd, enterArgs) } func (h *rktHandle) Signal(s os.Signal) error { @@ -823,13 +831,7 @@ func (d *rktHandle) Network() *cstructs.DriverNetwork { // Kill is used to terminate the task. We send an Interrupt // and then provide a 5 second grace period before doing a Kill. func (h *rktHandle) Kill() error { - h.executor.ShutDown() - select { - case <-h.doneCh: - return nil - case <-time.After(h.killTimeout): - return h.executor.Exit() - } + return h.executor.Shutdown(h.shutdownSignal, h.killTimeout) } func (h *rktHandle) Stats() (*cstructs.TaskResourceUsage, error) { @@ -845,8 +847,8 @@ func (h *rktHandle) run() { } } - // Exit the executor - if err := h.executor.Exit(); err != nil { + // Destroy the executor + if err := h.executor.Shutdown(h.shutdownSignal, 0); err != nil { h.logger.Printf("[ERR] driver.rkt: error killing executor: %v", err) } h.pluginClient.Kill() diff --git a/client/driver/rkt_test.go b/client/driver/rkt_test.go index 5e2f72d72..0d9d4249a 100644 --- a/client/driver/rkt_test.go +++ b/client/driver/rkt_test.go @@ -107,7 +107,7 @@ func TestRktDriver_Start_DNS(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRktDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -158,7 +158,7 @@ func TestRktDriver_Start_Wait(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRktDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -230,7 +230,7 @@ func TestRktDriver_Start_Wait_Skip_Trust(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRktDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -297,7 +297,7 @@ func TestRktDriver_Start_Wait_AllocDir(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRktDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -359,7 +359,7 @@ func TestRktDriver_UserGroup(t *testing.T) { } tctx := testDriverContexts(t, task) - defer tctx.AllocDir.Destroy() + defer tctx.Destroy() d := NewRktDriver(tctx.DriverCtx) _, err := d.Prestart(tctx.ExecCtx, task) @@ -414,7 +414,7 @@ func TestRktTrustPrefix(t *testing.T) { }, } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRktDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -449,7 +449,7 @@ func TestRktTaskValidate(t *testing.T) { Resources: basicResources, } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRktDriver(ctx.DriverCtx) if err := d.Validate(task.Config); err != nil { @@ -492,7 +492,7 @@ func TestRktDriver_PortMapping(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRktDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -556,7 +556,7 @@ func TestRktDriver_PortsMapping_Host(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRktDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -613,7 +613,7 @@ func TestRktDriver_HandlerExec(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRktDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { @@ -685,7 +685,7 @@ func TestRktDriver_Stats(t *testing.T) { } ctx := testDriverContexts(t, task) - defer ctx.AllocDir.Destroy() + defer ctx.Destroy() d := NewRktDriver(ctx.DriverCtx) if _, err := d.Prestart(ctx.ExecCtx, task); err != nil { diff --git a/client/driver/structs/structs.go b/client/driver/structs/structs.go index 6966205f7..5fc4a3a57 100644 --- a/client/driver/structs/structs.go +++ b/client/driver/structs/structs.go @@ -64,4 +64,8 @@ type ExecutorConfig struct { // LogLevel is the level of the logs to putout LogLevel string + + // FSIsolation if set will use an executor implementation that support + // filesystem isolation + FSIsolation bool } diff --git a/client/driver/structs/structs_linux.go b/client/driver/structs/structs_linux.go deleted file mode 100644 index 48f32f0be..000000000 --- a/client/driver/structs/structs_linux.go +++ /dev/null @@ -1,10 +0,0 @@ -package structs - -import cgroupConfig "github.com/opencontainers/runc/libcontainer/configs" - -// IsolationConfig has information about the isolation mechanism the executor -// uses to put resource constraints and isolation on the user process -type IsolationConfig struct { - Cgroup *cgroupConfig.Cgroup - CgroupPaths map[string]string -} diff --git a/client/driver/utils.go b/client/driver/utils.go index 5fba8071f..ef6dd17a9 100644 --- a/client/driver/utils.go +++ b/client/driver/utils.go @@ -11,6 +11,7 @@ import ( "time" "github.com/hashicorp/consul-template/signals" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/go-multierror" "github.com/hashicorp/go-plugin" "github.com/hashicorp/nomad/client/allocdir" @@ -48,7 +49,7 @@ func createExecutor(w io.Writer, clientConfig *config.Config, Cmd: exec.Command(bin, "executor", string(c)), } config.HandshakeConfig = HandshakeConfig - config.Plugins = GetPluginMap(w, clientConfig.LogLevel) + config.Plugins = GetPluginMap(w, hclog.LevelFromString(clientConfig.LogLevel), executorConfig.FSIsolation) config.MaxPort = clientConfig.ClientMaxPort config.MinPort = clientConfig.ClientMinPort @@ -77,7 +78,7 @@ func createExecutorWithConfig(config *plugin.ClientConfig, w io.Writer) (executo // Setting this to DEBUG since the log level at the executor server process // is already set, and this effects only the executor client. - config.Plugins = GetPluginMap(w, "DEBUG") + config.Plugins = GetPluginMap(w, hclog.Debug, false) executorClient := plugin.NewClient(config) rpcClient, err := executorClient.Client() @@ -93,11 +94,6 @@ func createExecutorWithConfig(config *plugin.ClientConfig, w io.Writer) (executo if !ok { return nil, nil, fmt.Errorf("unexpected executor rpc type: %T", raw) } - // 0.6 Upgrade path: Deregister services from the executor as the Nomad - // client agent now handles all Consul interactions. Ignore errors as - // this shouldn't cause the alloc to fail and there's nothing useful to - // do with them. - executorPlugin.DeregisterServices() return executorPlugin, executorClient, nil } diff --git a/client/logmon/client.go b/client/logmon/client.go new file mode 100644 index 000000000..ed9cef160 --- /dev/null +++ b/client/logmon/client.go @@ -0,0 +1,31 @@ +package logmon + +import ( + "context" + + "github.com/hashicorp/nomad/client/logmon/proto" +) + +type logmonClient struct { + client proto.LogMonClient +} + +func (c *logmonClient) Start(cfg *LogConfig) error { + req := &proto.StartRequest{ + LogDir: cfg.LogDir, + StdoutFileName: cfg.StdoutLogFile, + StderrFileName: cfg.StderrLogFile, + MaxFiles: uint32(cfg.MaxFiles), + MaxFileSizeMb: uint32(cfg.MaxFileSizeMB), + StdoutFifo: cfg.StdoutFifo, + StderrFifo: cfg.StderrFifo, + } + _, err := c.client.Start(context.Background(), req) + return err +} + +func (c *logmonClient) Stop() error { + req := &proto.StopRequest{} + _, err := c.client.Stop(context.Background(), req) + return err +} diff --git a/client/logmon/logmon.go b/client/logmon/logmon.go new file mode 100644 index 000000000..2c7609eae --- /dev/null +++ b/client/logmon/logmon.go @@ -0,0 +1,209 @@ +package logmon + +import ( + "fmt" + "io" + "strings" + "time" + + hclog "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/driver/logging" + "github.com/hashicorp/nomad/client/lib/fifo" +) + +const ( + // processOutputCloseTolerance is the length of time we will wait for the + // launched process to close its stdout/stderr before we force close it. If + // data is written after this tolerance, we will not capture it. + processOutputCloseTolerance = 2 * time.Second +) + +type LogConfig struct { + // LogDir is the host path where logs are to be written to + LogDir string + + // StdoutLogFile is the path relative to LogDir for stdout logging + StdoutLogFile string + + // StderrLogFile is the path relative to LogDir for stderr logging + StderrLogFile string + + // StdoutFifo is the path on the host to the stdout pipe + StdoutFifo string + + // StderrFifo is the path on the host to the stderr pipe + StderrFifo string + + // MaxFiles is the max rotated files allowed + MaxFiles int + + // MaxFileSizeMB is the max log file size in MB allowed before rotation occures + MaxFileSizeMB int +} + +type LogMon interface { + Start(*LogConfig) error + Stop() error +} + +func NewLogMon(logger hclog.Logger) LogMon { + return &logmonImpl{ + logger: logger, + } +} + +type logmonImpl struct { + logger hclog.Logger + tl *TaskLogger +} + +func (l *logmonImpl) Start(cfg *LogConfig) error { + tl, err := NewTaskLogger(cfg, l.logger) + if err != nil { + return err + } + l.tl = tl + return nil +} + +func (l *logmonImpl) Stop() error { + l.tl.Close() + return nil +} + +type TaskLogger struct { + config *LogConfig + + // rotator for stdout + lro *logRotatorWrapper + + // rotator for stderr + lre *logRotatorWrapper +} + +func (tl *TaskLogger) Close() { + if tl.lro != nil { + tl.lro.Close() + } + if tl.lre != nil { + tl.lre.Close() + } +} + +func NewTaskLogger(cfg *LogConfig, logger hclog.Logger) (*TaskLogger, error) { + tl := &TaskLogger{config: cfg} + + logFileSize := int64(cfg.MaxFileSizeMB * 1024 * 1024) + lro, err := logging.NewFileRotator(cfg.LogDir, cfg.StdoutLogFile, + cfg.MaxFiles, logFileSize, logger) + if err != nil { + return nil, fmt.Errorf("failed to create stdout logfile for %q: %v", cfg.StdoutLogFile, err) + } + + wrapperOut, err := newLogRotatorWrapper(cfg.StdoutFifo, logger, lro) + if err != nil { + return nil, err + } + + tl.lro = wrapperOut + + lre, err := logging.NewFileRotator(cfg.LogDir, cfg.StderrLogFile, + cfg.MaxFiles, logFileSize, logger) + if err != nil { + return nil, fmt.Errorf("failed to create stderr logfile for %q: %v", cfg.StderrLogFile, err) + } + + wrapperErr, err := newLogRotatorWrapper(cfg.StderrFifo, logger, lre) + if err != nil { + return nil, err + } + + tl.lre = wrapperErr + + return tl, nil + +} + +// logRotatorWrapper wraps our log rotator and exposes a pipe that can feed the +// log rotator data. The processOutWriter should be attached to the process and +// data will be copied from the reader to the rotator. +type logRotatorWrapper struct { + fifoPath string + processOutReader io.ReadCloser + rotatorWriter *logging.FileRotator + hasFinishedCopied chan struct{} + logger hclog.Logger +} + +// newLogRotatorWrapper takes a rotator and returns a wrapper that has the +// processOutWriter to attach to the stdout or stderr of a process. +func newLogRotatorWrapper(path string, logger hclog.Logger, rotator *logging.FileRotator) (*logRotatorWrapper, error) { + logger.Info("opening fifo", "path", path) + f, err := fifo.New(path) + if err != nil { + return nil, fmt.Errorf("failed to create fifo for extracting logs: %v", err) + } + + wrap := &logRotatorWrapper{ + fifoPath: path, + processOutReader: f, + rotatorWriter: rotator, + hasFinishedCopied: make(chan struct{}), + logger: logger, + } + wrap.start() + return wrap, nil +} + +// start starts a goroutine that copies from the pipe into the rotator. This is +// called by the constructor and not the user of the wrapper. +func (l *logRotatorWrapper) start() { + go func() { + defer close(l.hasFinishedCopied) + _, err := io.Copy(l.rotatorWriter, l.processOutReader) + if err != nil { + // Close reader to propagate io error across pipe. + // Note that this may block until the process exits on + // Windows due to + // https://github.com/PowerShell/PowerShell/issues/4254 + // or similar issues. Since this is already running in + // a goroutine its safe to block until the process is + // force-killed. + l.processOutReader.Close() + } + }() + return +} + +// Close closes the rotator and the process writer to ensure that the Wait +// command exits. +func (l *logRotatorWrapper) Close() { + // Wait up to the close tolerance before we force close + select { + case <-l.hasFinishedCopied: + case <-time.After(processOutputCloseTolerance): + } + + // Closing the read side of a pipe may block on Windows if the process + // is being debugged as in: + // https://github.com/PowerShell/PowerShell/issues/4254 + // The pipe will be closed and cleaned up when the process exits. + closeDone := make(chan struct{}) + go func() { + defer close(closeDone) + err := l.processOutReader.Close() + if err != nil && !strings.Contains(err.Error(), "file already closed") { + l.logger.Warn("error closing read-side of process output pipe", "err", err) + } + + }() + + select { + case <-closeDone: + case <-time.After(processOutputCloseTolerance): + l.logger.Warn("timed out waiting for read-side of process output pipe to close") + } + + l.rotatorWriter.Close() + return +} diff --git a/client/logmon/plugin.go b/client/logmon/plugin.go new file mode 100644 index 000000000..3e0db7317 --- /dev/null +++ b/client/logmon/plugin.go @@ -0,0 +1,69 @@ +package logmon + +import ( + "context" + "os/exec" + + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/client/logmon/proto" + "github.com/hashicorp/nomad/helper/discover" + "github.com/hashicorp/nomad/plugins/base" + "google.golang.org/grpc" +) + +// LaunchLogMon an instance of logmon +// TODO: Integrate with base plugin loader +func LaunchLogMon(logger hclog.Logger) (LogMon, *plugin.Client, error) { + logger = logger.Named("logmon-launcher") + bin, err := discover.NomadExecutable() + if err != nil { + return nil, nil, err + } + + client := plugin.NewClient(&plugin.ClientConfig{ + HandshakeConfig: base.Handshake, + Plugins: map[string]plugin.Plugin{ + "logmon": NewPlugin(NewLogMon(hclog.L().Named("logmon"))), + }, + Cmd: exec.Command(bin, "logmon"), + AllowedProtocols: []plugin.Protocol{ + plugin.ProtocolGRPC, + }, + }) + + rpcClient, err := client.Client() + if err != nil { + return nil, nil, err + } + + raw, err := rpcClient.Dispense("logmon") + if err != nil { + return nil, nil, err + } + + l := raw.(LogMon) + return l, client, nil + +} + +type Plugin struct { + plugin.NetRPCUnsupportedPlugin + impl LogMon +} + +func NewPlugin(i LogMon) plugin.Plugin { + return &Plugin{impl: i} +} + +func (p *Plugin) GRPCServer(broker *plugin.GRPCBroker, s *grpc.Server) error { + proto.RegisterLogMonServer(s, &logmonServer{ + impl: p.impl, + broker: broker, + }) + return nil +} + +func (p *Plugin) GRPCClient(ctx context.Context, broker *plugin.GRPCBroker, c *grpc.ClientConn) (interface{}, error) { + return &logmonClient{client: proto.NewLogMonClient(c)}, nil +} diff --git a/client/logmon/proto/logmon.pb.go b/client/logmon/proto/logmon.pb.go new file mode 100644 index 000000000..9103ac270 --- /dev/null +++ b/client/logmon/proto/logmon.pb.go @@ -0,0 +1,338 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: logmon.proto + +package proto + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" + +import ( + context "golang.org/x/net/context" + grpc "google.golang.org/grpc" +) + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +type StartRequest struct { + LogDir string `protobuf:"bytes,1,opt,name=log_dir,json=logDir,proto3" json:"log_dir,omitempty"` + StdoutFileName string `protobuf:"bytes,2,opt,name=stdout_file_name,json=stdoutFileName,proto3" json:"stdout_file_name,omitempty"` + StderrFileName string `protobuf:"bytes,3,opt,name=stderr_file_name,json=stderrFileName,proto3" json:"stderr_file_name,omitempty"` + MaxFiles uint32 `protobuf:"varint,4,opt,name=max_files,json=maxFiles,proto3" json:"max_files,omitempty"` + MaxFileSizeMb uint32 `protobuf:"varint,5,opt,name=max_file_size_mb,json=maxFileSizeMb,proto3" json:"max_file_size_mb,omitempty"` + StdoutFifo string `protobuf:"bytes,6,opt,name=stdout_fifo,json=stdoutFifo,proto3" json:"stdout_fifo,omitempty"` + StderrFifo string `protobuf:"bytes,7,opt,name=stderr_fifo,json=stderrFifo,proto3" json:"stderr_fifo,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *StartRequest) Reset() { *m = StartRequest{} } +func (m *StartRequest) String() string { return proto.CompactTextString(m) } +func (*StartRequest) ProtoMessage() {} +func (*StartRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_logmon_0690d412b596ec9e, []int{0} +} +func (m *StartRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StartRequest.Unmarshal(m, b) +} +func (m *StartRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StartRequest.Marshal(b, m, deterministic) +} +func (dst *StartRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_StartRequest.Merge(dst, src) +} +func (m *StartRequest) XXX_Size() int { + return xxx_messageInfo_StartRequest.Size(m) +} +func (m *StartRequest) XXX_DiscardUnknown() { + xxx_messageInfo_StartRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_StartRequest proto.InternalMessageInfo + +func (m *StartRequest) GetLogDir() string { + if m != nil { + return m.LogDir + } + return "" +} + +func (m *StartRequest) GetStdoutFileName() string { + if m != nil { + return m.StdoutFileName + } + return "" +} + +func (m *StartRequest) GetStderrFileName() string { + if m != nil { + return m.StderrFileName + } + return "" +} + +func (m *StartRequest) GetMaxFiles() uint32 { + if m != nil { + return m.MaxFiles + } + return 0 +} + +func (m *StartRequest) GetMaxFileSizeMb() uint32 { + if m != nil { + return m.MaxFileSizeMb + } + return 0 +} + +func (m *StartRequest) GetStdoutFifo() string { + if m != nil { + return m.StdoutFifo + } + return "" +} + +func (m *StartRequest) GetStderrFifo() string { + if m != nil { + return m.StderrFifo + } + return "" +} + +type StartResponse struct { + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *StartResponse) Reset() { *m = StartResponse{} } +func (m *StartResponse) String() string { return proto.CompactTextString(m) } +func (*StartResponse) ProtoMessage() {} +func (*StartResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_logmon_0690d412b596ec9e, []int{1} +} +func (m *StartResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StartResponse.Unmarshal(m, b) +} +func (m *StartResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StartResponse.Marshal(b, m, deterministic) +} +func (dst *StartResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_StartResponse.Merge(dst, src) +} +func (m *StartResponse) XXX_Size() int { + return xxx_messageInfo_StartResponse.Size(m) +} +func (m *StartResponse) XXX_DiscardUnknown() { + xxx_messageInfo_StartResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_StartResponse proto.InternalMessageInfo + +type StopRequest struct { + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *StopRequest) Reset() { *m = StopRequest{} } +func (m *StopRequest) String() string { return proto.CompactTextString(m) } +func (*StopRequest) ProtoMessage() {} +func (*StopRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_logmon_0690d412b596ec9e, []int{2} +} +func (m *StopRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StopRequest.Unmarshal(m, b) +} +func (m *StopRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StopRequest.Marshal(b, m, deterministic) +} +func (dst *StopRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_StopRequest.Merge(dst, src) +} +func (m *StopRequest) XXX_Size() int { + return xxx_messageInfo_StopRequest.Size(m) +} +func (m *StopRequest) XXX_DiscardUnknown() { + xxx_messageInfo_StopRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_StopRequest proto.InternalMessageInfo + +type StopResponse struct { + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *StopResponse) Reset() { *m = StopResponse{} } +func (m *StopResponse) String() string { return proto.CompactTextString(m) } +func (*StopResponse) ProtoMessage() {} +func (*StopResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_logmon_0690d412b596ec9e, []int{3} +} +func (m *StopResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StopResponse.Unmarshal(m, b) +} +func (m *StopResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StopResponse.Marshal(b, m, deterministic) +} +func (dst *StopResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_StopResponse.Merge(dst, src) +} +func (m *StopResponse) XXX_Size() int { + return xxx_messageInfo_StopResponse.Size(m) +} +func (m *StopResponse) XXX_DiscardUnknown() { + xxx_messageInfo_StopResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_StopResponse proto.InternalMessageInfo + +func init() { + proto.RegisterType((*StartRequest)(nil), "hashicorp.nomad.client.logmon.proto.StartRequest") + proto.RegisterType((*StartResponse)(nil), "hashicorp.nomad.client.logmon.proto.StartResponse") + proto.RegisterType((*StopRequest)(nil), "hashicorp.nomad.client.logmon.proto.StopRequest") + proto.RegisterType((*StopResponse)(nil), "hashicorp.nomad.client.logmon.proto.StopResponse") +} + +// Reference imports to suppress errors if they are not otherwise used. +var _ context.Context +var _ grpc.ClientConn + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the grpc package it is being compiled against. +const _ = grpc.SupportPackageIsVersion4 + +// LogMonClient is the client API for LogMon service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. +type LogMonClient interface { + Start(ctx context.Context, in *StartRequest, opts ...grpc.CallOption) (*StartResponse, error) + Stop(ctx context.Context, in *StopRequest, opts ...grpc.CallOption) (*StopResponse, error) +} + +type logMonClient struct { + cc *grpc.ClientConn +} + +func NewLogMonClient(cc *grpc.ClientConn) LogMonClient { + return &logMonClient{cc} +} + +func (c *logMonClient) Start(ctx context.Context, in *StartRequest, opts ...grpc.CallOption) (*StartResponse, error) { + out := new(StartResponse) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.client.logmon.proto.LogMon/Start", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *logMonClient) Stop(ctx context.Context, in *StopRequest, opts ...grpc.CallOption) (*StopResponse, error) { + out := new(StopResponse) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.client.logmon.proto.LogMon/Stop", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +// LogMonServer is the server API for LogMon service. +type LogMonServer interface { + Start(context.Context, *StartRequest) (*StartResponse, error) + Stop(context.Context, *StopRequest) (*StopResponse, error) +} + +func RegisterLogMonServer(s *grpc.Server, srv LogMonServer) { + s.RegisterService(&_LogMon_serviceDesc, srv) +} + +func _LogMon_Start_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(StartRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(LogMonServer).Start(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/hashicorp.nomad.client.logmon.proto.LogMon/Start", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(LogMonServer).Start(ctx, req.(*StartRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _LogMon_Stop_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(StopRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(LogMonServer).Stop(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/hashicorp.nomad.client.logmon.proto.LogMon/Stop", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(LogMonServer).Stop(ctx, req.(*StopRequest)) + } + return interceptor(ctx, in, info, handler) +} + +var _LogMon_serviceDesc = grpc.ServiceDesc{ + ServiceName: "hashicorp.nomad.client.logmon.proto.LogMon", + HandlerType: (*LogMonServer)(nil), + Methods: []grpc.MethodDesc{ + { + MethodName: "Start", + Handler: _LogMon_Start_Handler, + }, + { + MethodName: "Stop", + Handler: _LogMon_Stop_Handler, + }, + }, + Streams: []grpc.StreamDesc{}, + Metadata: "logmon.proto", +} + +func init() { proto.RegisterFile("logmon.proto", fileDescriptor_logmon_0690d412b596ec9e) } + +var fileDescriptor_logmon_0690d412b596ec9e = []byte{ + // 314 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x90, 0x31, 0x6f, 0xf2, 0x30, + 0x10, 0x86, 0xbf, 0xf0, 0x41, 0x28, 0x07, 0xa1, 0xc8, 0x4b, 0x23, 0x3a, 0x14, 0xa5, 0x43, 0x99, + 0xa2, 0x42, 0xff, 0x41, 0x55, 0x75, 0x2a, 0x1d, 0x60, 0xeb, 0x12, 0x19, 0xb8, 0x04, 0x4b, 0x71, + 0x2e, 0xb5, 0x8d, 0x84, 0x58, 0xfb, 0x6b, 0xfb, 0x2f, 0xaa, 0x18, 0x13, 0x65, 0x84, 0x29, 0xca, + 0xdd, 0xf3, 0xea, 0x1e, 0xbf, 0x30, 0xc8, 0x29, 0x93, 0x54, 0xc4, 0xa5, 0x22, 0x43, 0xec, 0x71, + 0xc7, 0xf5, 0x4e, 0x6c, 0x48, 0x95, 0x71, 0x41, 0x92, 0x6f, 0xe3, 0x4d, 0x2e, 0xb0, 0x30, 0x71, + 0x13, 0x8a, 0x7e, 0x5a, 0x30, 0x58, 0x19, 0xae, 0xcc, 0x12, 0xbf, 0xf7, 0xa8, 0x0d, 0xbb, 0x83, + 0x6e, 0x4e, 0x59, 0xb2, 0x15, 0x2a, 0xf4, 0x26, 0xde, 0xb4, 0xb7, 0xf4, 0x73, 0xca, 0xde, 0x84, + 0x62, 0x53, 0x18, 0x69, 0xb3, 0xa5, 0xbd, 0x49, 0x52, 0x91, 0x63, 0x52, 0x70, 0x89, 0x61, 0xcb, + 0x12, 0xc3, 0xd3, 0xfc, 0x5d, 0xe4, 0xf8, 0xc9, 0x25, 0x3a, 0x12, 0x95, 0x6a, 0x90, 0xff, 0x6b, + 0x12, 0x95, 0xaa, 0xc9, 0x7b, 0xe8, 0x49, 0x7e, 0xb0, 0x98, 0x0e, 0xdb, 0x13, 0x6f, 0x1a, 0x2c, + 0x6f, 0x24, 0x3f, 0x54, 0x7b, 0xcd, 0x9e, 0x60, 0x74, 0x5e, 0x26, 0x5a, 0x1c, 0x31, 0x91, 0xeb, + 0xb0, 0x63, 0x99, 0xc0, 0x31, 0x2b, 0x71, 0xc4, 0xc5, 0x9a, 0x3d, 0x40, 0xbf, 0x36, 0x4b, 0x29, + 0xf4, 0xed, 0x29, 0x38, 0x4b, 0xa5, 0xe4, 0x80, 0x93, 0x50, 0x4a, 0x61, 0xb7, 0x06, 0xac, 0x4b, + 0x4a, 0xd1, 0x2d, 0x04, 0xae, 0x04, 0x5d, 0x52, 0xa1, 0x31, 0x0a, 0xa0, 0xbf, 0x32, 0x54, 0xba, + 0x52, 0xa2, 0x61, 0x55, 0x52, 0xf5, 0x7b, 0x5a, 0xcf, 0x7f, 0x3d, 0xf0, 0x3f, 0x28, 0x5b, 0x50, + 0xc1, 0x4a, 0xe8, 0xd8, 0x28, 0x9b, 0xc5, 0x17, 0xf4, 0x1d, 0x37, 0xbb, 0x1e, 0xcf, 0xaf, 0x89, + 0x38, 0xb3, 0x7f, 0x4c, 0x42, 0xbb, 0x92, 0x61, 0xcf, 0x17, 0xa6, 0xeb, 0x67, 0x8c, 0x67, 0x57, + 0x24, 0xce, 0xe7, 0x5e, 0xbb, 0x5f, 0x1d, 0x3b, 0x5f, 0xfb, 0xf6, 0xf3, 0xf2, 0x17, 0x00, 0x00, + 0xff, 0xff, 0x1b, 0x8f, 0x8b, 0x51, 0x66, 0x02, 0x00, 0x00, +} diff --git a/client/logmon/proto/logmon.proto b/client/logmon/proto/logmon.proto new file mode 100644 index 000000000..a5048ddeb --- /dev/null +++ b/client/logmon/proto/logmon.proto @@ -0,0 +1,25 @@ +syntax = "proto3"; +package hashicorp.nomad.client.logmon.proto; +option go_package = "proto"; + +service LogMon { + rpc Start(StartRequest) returns (StartResponse) {} + rpc Stop(StopRequest) returns (StopResponse) {} +} + +message StartRequest { + string log_dir = 1; + string stdout_file_name = 2; + string stderr_file_name = 3; + uint32 max_files = 4; + uint32 max_file_size_mb = 5; + string stdout_fifo = 6; + string stderr_fifo = 7; +} + +message StartResponse { +} + +message StopRequest {} + +message StopResponse {} diff --git a/client/logmon/server.go b/client/logmon/server.go new file mode 100644 index 000000000..1be46d80a --- /dev/null +++ b/client/logmon/server.go @@ -0,0 +1,36 @@ +package logmon + +import ( + "golang.org/x/net/context" + + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/client/logmon/proto" +) + +type logmonServer struct { + broker *plugin.GRPCBroker + impl LogMon +} + +func (s *logmonServer) Start(ctx context.Context, req *proto.StartRequest) (*proto.StartResponse, error) { + cfg := &LogConfig{ + LogDir: req.LogDir, + StdoutLogFile: req.StdoutFileName, + StderrLogFile: req.StderrFileName, + MaxFiles: int(req.MaxFiles), + MaxFileSizeMB: int(req.MaxFileSizeMb), + StdoutFifo: req.StdoutFifo, + StderrFifo: req.StderrFifo, + } + + err := s.impl.Start(cfg) + if err != nil { + return nil, err + } + resp := &proto.StartResponse{} + return resp, nil +} + +func (s *logmonServer) Stop(ctx context.Context, req *proto.StopRequest) (*proto.StopResponse, error) { + return &proto.StopResponse{}, s.impl.Stop() +} diff --git a/command/commands.go b/command/commands.go index 27f33b159..69c81b39e 100644 --- a/command/commands.go +++ b/command/commands.go @@ -325,6 +325,11 @@ func Commands(metaPtr *Meta, agentUi cli.Ui) map[string]cli.CommandFactory { Meta: meta, }, nil }, + "logmon": func() (cli.Command, error) { + return &LogMonPluginCommand{ + Meta: meta, + }, nil + }, "logs": func() (cli.Command, error) { return &AllocLogsCommand{ Meta: meta, diff --git a/command/executor_plugin.go b/command/executor_plugin.go index 93da6e1b5..2900c94c0 100644 --- a/command/executor_plugin.go +++ b/command/executor_plugin.go @@ -5,6 +5,7 @@ import ( "os" "strings" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/go-plugin" "github.com/hashicorp/nomad/client/driver" @@ -43,7 +44,11 @@ func (e *ExecutorPluginCommand) Run(args []string) int { } plugin.Serve(&plugin.ServeConfig{ HandshakeConfig: driver.HandshakeConfig, - Plugins: driver.GetPluginMap(stdo, executorConfig.LogLevel), + Plugins: driver.GetPluginMap( + stdo, + hclog.LevelFromString(executorConfig.LogLevel), + executorConfig.FSIsolation, + ), }) return 0 } diff --git a/command/logmon_plugin.go b/command/logmon_plugin.go new file mode 100644 index 000000000..f9329e423 --- /dev/null +++ b/command/logmon_plugin.go @@ -0,0 +1,36 @@ +package command + +import ( + "strings" + + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/client/logmon" + "github.com/hashicorp/nomad/plugins/base" +) + +type LogMonPluginCommand struct { + Meta +} + +func (e *LogMonPluginCommand) Help() string { + helpText := ` + This is a command used by Nomad internally to launch the logmon process" + ` + return strings.TrimSpace(helpText) +} + +func (e *LogMonPluginCommand) Synopsis() string { + return "internal - launch a logmon plugin" +} + +func (e *LogMonPluginCommand) Run(args []string) int { + plugin.Serve(&plugin.ServeConfig{ + HandshakeConfig: base.Handshake, + Plugins: map[string]plugin.Plugin{ + "logmon": logmon.NewPlugin(logmon.NewLogMon(hclog.Default().Named("logmon"))), + }, + GRPCServer: plugin.DefaultGRPCServer, + }) + return 0 +} diff --git a/main.go b/main.go index d74bef167..45229139c 100644 --- a/main.go +++ b/main.go @@ -29,6 +29,7 @@ var ( "executor", "keygen", "keyring", + "logmon", "node-drain", "node-status", "server-force-leave", diff --git a/nsenter_linux.go b/nsenter_linux.go new file mode 100644 index 000000000..7579483c4 --- /dev/null +++ b/nsenter_linux.go @@ -0,0 +1,25 @@ +package main + +import ( + "os" + "runtime" + + hclog "github.com/hashicorp/go-hclog" + "github.com/opencontainers/runc/libcontainer" + _ "github.com/opencontainers/runc/libcontainer/nsenter" +) + +// init is only run on linux and is used when the LibcontainerExecutor starts +// a new process. The libcontainer shim takes over the process, setting up the +// configured isolation and limitions before execve into the user process +func init() { + if len(os.Args) > 1 && os.Args[1] == "libcontainer-shim" { + runtime.GOMAXPROCS(1) + runtime.LockOSThread() + factory, _ := libcontainer.New("") + if err := factory.StartInitialization(); err != nil { + hclog.L().Error("failed to initialize libcontainer-shim", "error", err) + } + panic("--this line should have never been executed, congratulations--") + } +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/container_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/container_linux.go index 9f551fd09..451d4dbf4 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/container_linux.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/container_linux.go @@ -452,8 +452,7 @@ func (c *linuxContainer) newParentProcess(p *Process) (parentProcess, error) { } func (c *linuxContainer) commandTemplate(p *Process, childPipe *os.File) (*exec.Cmd, error) { - cmd := exec.Command(c.initPath, c.initArgs[1:]...) - cmd.Args[0] = c.initArgs[0] + cmd := exec.Command(c.initPath, c.initArgs...) cmd.Stdin = p.Stdin cmd.Stdout = p.Stdout cmd.Stderr = p.Stderr diff --git a/vendor/github.com/opencontainers/runc/libcontainer/factory_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/factory_linux.go index 4cbd70cd7..0297333c5 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/factory_linux.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/factory_linux.go @@ -11,6 +11,7 @@ import ( "runtime/debug" "strconv" + "github.com/cyphar/filepath-securejoin" "github.com/opencontainers/runc/libcontainer/cgroups" "github.com/opencontainers/runc/libcontainer/cgroups/fs" "github.com/opencontainers/runc/libcontainer/cgroups/systemd" @@ -42,7 +43,10 @@ func InitArgs(args ...string) func(*LinuxFactory) error { } } - l.InitArgs = args + l.InitPath = args[0] + if len(args) > 1 { + l.InitArgs = args[1:] + } return nil } } @@ -195,7 +199,10 @@ func (l *LinuxFactory) Create(id string, config *configs.Config) (Container, err if err := l.Validator.Validate(config); err != nil { return nil, newGenericError(err, ConfigInvalid) } - containerRoot := filepath.Join(l.Root, id) + containerRoot, err := securejoin.SecureJoin(l.Root, id) + if err != nil { + return nil, err + } if _, err := os.Stat(containerRoot); err == nil { return nil, newGenericError(fmt.Errorf("container with id exists: %v", id), IdInUse) } else if !os.IsNotExist(err) { @@ -229,7 +236,14 @@ func (l *LinuxFactory) Load(id string) (Container, error) { if l.Root == "" { return nil, newGenericError(fmt.Errorf("invalid root"), ConfigInvalid) } - containerRoot := filepath.Join(l.Root, id) + //when load, we need to check id is valid or not. + if err := l.validateID(id); err != nil { + return nil, err + } + containerRoot, err := securejoin.SecureJoin(l.Root, id) + if err != nil { + return nil, err + } state, err := l.loadState(containerRoot, id) if err != nil { return nil, err @@ -339,7 +353,11 @@ func (l *LinuxFactory) StartInitialization() (err error) { } func (l *LinuxFactory) loadState(root, id string) (*State, error) { - f, err := os.Open(filepath.Join(root, stateFilename)) + stateFilePath, err := securejoin.SecureJoin(root, stateFilename) + if err != nil { + return nil, err + } + f, err := os.Open(stateFilePath) if err != nil { if os.IsNotExist(err) { return nil, newGenericError(fmt.Errorf("container %q does not exist", id), ContainerNotExists) @@ -355,7 +373,7 @@ func (l *LinuxFactory) loadState(root, id string) (*State, error) { } func (l *LinuxFactory) validateID(id string) error { - if !idRegex.MatchString(id) { + if !idRegex.MatchString(id) || string(os.PathSeparator)+id != utils.CleanPath(string(os.PathSeparator)+id) { return newGenericError(fmt.Errorf("invalid id format: %v", id), InvalidIdFormat) } diff --git a/vendor/github.com/opencontainers/runc/libcontainer/network_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/network_linux.go index 5075bee4d..569c53f6e 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/network_linux.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/network_linux.go @@ -5,18 +5,15 @@ package libcontainer import ( "fmt" "io/ioutil" - "net" "path/filepath" "strconv" "strings" "github.com/opencontainers/runc/libcontainer/configs" - "github.com/opencontainers/runc/libcontainer/utils" "github.com/vishvananda/netlink" ) var strategies = map[string]networkStrategy{ - "veth": &veth{}, "loopback": &loopback{}, } @@ -103,157 +100,3 @@ func (l *loopback) attach(n *configs.Network) (err error) { func (l *loopback) detach(n *configs.Network) (err error) { return nil } - -// veth is a network strategy that uses a bridge and creates -// a veth pair, one that is attached to the bridge on the host and the other -// is placed inside the container's namespace -type veth struct { -} - -func (v *veth) detach(n *configs.Network) (err error) { - return netlink.LinkSetMaster(&netlink.Device{LinkAttrs: netlink.LinkAttrs{Name: n.HostInterfaceName}}, nil) -} - -// attach a container network interface to an external network -func (v *veth) attach(n *configs.Network) (err error) { - brl, err := netlink.LinkByName(n.Bridge) - if err != nil { - return err - } - br, ok := brl.(*netlink.Bridge) - if !ok { - return fmt.Errorf("Wrong device type %T", brl) - } - host, err := netlink.LinkByName(n.HostInterfaceName) - if err != nil { - return err - } - - if err := netlink.LinkSetMaster(host, br); err != nil { - return err - } - if err := netlink.LinkSetMTU(host, n.Mtu); err != nil { - return err - } - if n.HairpinMode { - if err := netlink.LinkSetHairpin(host, true); err != nil { - return err - } - } - if err := netlink.LinkSetUp(host); err != nil { - return err - } - - return nil -} - -func (v *veth) create(n *network, nspid int) (err error) { - tmpName, err := v.generateTempPeerName() - if err != nil { - return err - } - n.TempVethPeerName = tmpName - if n.Bridge == "" { - return fmt.Errorf("bridge is not specified") - } - veth := &netlink.Veth{ - LinkAttrs: netlink.LinkAttrs{ - Name: n.HostInterfaceName, - TxQLen: n.TxQueueLen, - }, - PeerName: n.TempVethPeerName, - } - if err := netlink.LinkAdd(veth); err != nil { - return err - } - defer func() { - if err != nil { - netlink.LinkDel(veth) - } - }() - if err := v.attach(&n.Network); err != nil { - return err - } - child, err := netlink.LinkByName(n.TempVethPeerName) - if err != nil { - return err - } - return netlink.LinkSetNsPid(child, nspid) -} - -func (v *veth) generateTempPeerName() (string, error) { - return utils.GenerateRandomName("veth", 7) -} - -func (v *veth) initialize(config *network) error { - peer := config.TempVethPeerName - if peer == "" { - return fmt.Errorf("peer is not specified") - } - child, err := netlink.LinkByName(peer) - if err != nil { - return err - } - if err := netlink.LinkSetDown(child); err != nil { - return err - } - if err := netlink.LinkSetName(child, config.Name); err != nil { - return err - } - // get the interface again after we changed the name as the index also changes. - if child, err = netlink.LinkByName(config.Name); err != nil { - return err - } - if config.MacAddress != "" { - mac, err := net.ParseMAC(config.MacAddress) - if err != nil { - return err - } - if err := netlink.LinkSetHardwareAddr(child, mac); err != nil { - return err - } - } - ip, err := netlink.ParseAddr(config.Address) - if err != nil { - return err - } - if err := netlink.AddrAdd(child, ip); err != nil { - return err - } - if config.IPv6Address != "" { - ip6, err := netlink.ParseAddr(config.IPv6Address) - if err != nil { - return err - } - if err := netlink.AddrAdd(child, ip6); err != nil { - return err - } - } - if err := netlink.LinkSetMTU(child, config.Mtu); err != nil { - return err - } - if err := netlink.LinkSetUp(child); err != nil { - return err - } - if config.Gateway != "" { - gw := net.ParseIP(config.Gateway) - if err := netlink.RouteAdd(&netlink.Route{ - Scope: netlink.SCOPE_UNIVERSE, - LinkIndex: child.Attrs().Index, - Gw: gw, - }); err != nil { - return err - } - } - if config.IPv6Gateway != "" { - gw := net.ParseIP(config.IPv6Gateway) - if err := netlink.RouteAdd(&netlink.Route{ - Scope: netlink.SCOPE_UNIVERSE, - LinkIndex: child.Attrs().Index, - Gw: gw, - }); err != nil { - return err - } - } - return nil -} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/nsenter/README.md b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/README.md new file mode 100644 index 000000000..96d327d80 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/README.md @@ -0,0 +1,41 @@ +## nsenter + +The `nsenter` package registers a special init constructor that is called before +the Go runtime has a chance to boot. This provides us the ability to `setns` on +existing namespaces and avoid the issues that the Go runtime has with multiple +threads. This constructor will be called if this package is registered, +imported, in your go application. + +The `nsenter` package will `import "C"` and it uses [cgo](https://golang.org/cmd/cgo/) +package. In cgo, if the import of "C" is immediately preceded by a comment, that comment, +called the preamble, is used as a header when compiling the C parts of the package. +So every time we import package `nsenter`, the C code function `nsexec()` would be +called. And package `nsenter` is now only imported in `main_unix.go`, so every time +before we call `cmd.Start` on linux, that C code would run. + +Because `nsexec()` must be run before the Go runtime in order to use the +Linux kernel namespace, you must `import` this library into a package if +you plan to use `libcontainer` directly. Otherwise Go will not execute +the `nsexec()` constructor, which means that the re-exec will not cause +the namespaces to be joined. You can import it like this: + +```go +import _ "github.com/opencontainers/runc/libcontainer/nsenter" +``` + +`nsexec()` will first get the file descriptor number for the init pipe +from the environment variable `_LIBCONTAINER_INITPIPE` (which was opened +by the parent and kept open across the fork-exec of the `nsexec()` init +process). The init pipe is used to read bootstrap data (namespace paths, +clone flags, uid and gid mappings, and the console path) from the parent +process. `nsexec()` will then call `setns(2)` to join the namespaces +provided in the bootstrap data (if available), `clone(2)` a child process +with the provided clone flags, update the user and group ID mappings, do +some further miscellaneous setup steps, and then send the PID of the +child process to the parent of the `nsexec()` "caller". Finally, +the parent `nsexec()` will exit and the child `nsexec()` process will +return to allow the Go runtime take over. + +NOTE: We do both `setns(2)` and `clone(2)` even if we don't have any +CLONE_NEW* clone flags because we must fork a new process in order to +enter the PID namespace. diff --git a/vendor/github.com/opencontainers/runc/libcontainer/nsenter/namespace.h b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/namespace.h new file mode 100644 index 000000000..9e9bdca05 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/namespace.h @@ -0,0 +1,32 @@ +#ifndef NSENTER_NAMESPACE_H +#define NSENTER_NAMESPACE_H + +#ifndef _GNU_SOURCE +# define _GNU_SOURCE +#endif +#include + +/* All of these are taken from include/uapi/linux/sched.h */ +#ifndef CLONE_NEWNS +# define CLONE_NEWNS 0x00020000 /* New mount namespace group */ +#endif +#ifndef CLONE_NEWCGROUP +# define CLONE_NEWCGROUP 0x02000000 /* New cgroup namespace */ +#endif +#ifndef CLONE_NEWUTS +# define CLONE_NEWUTS 0x04000000 /* New utsname namespace */ +#endif +#ifndef CLONE_NEWIPC +# define CLONE_NEWIPC 0x08000000 /* New ipc namespace */ +#endif +#ifndef CLONE_NEWUSER +# define CLONE_NEWUSER 0x10000000 /* New user namespace */ +#endif +#ifndef CLONE_NEWPID +# define CLONE_NEWPID 0x20000000 /* New pid namespace */ +#endif +#ifndef CLONE_NEWNET +# define CLONE_NEWNET 0x40000000 /* New network namespace */ +#endif + +#endif /* NSENTER_NAMESPACE_H */ diff --git a/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter.go b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter.go new file mode 100644 index 000000000..07f4d63e4 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter.go @@ -0,0 +1,12 @@ +// +build linux,!gccgo + +package nsenter + +/* +#cgo CFLAGS: -Wall +extern void nsexec(); +void __attribute__((constructor)) init(void) { + nsexec(); +} +*/ +import "C" diff --git a/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter_gccgo.go b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter_gccgo.go new file mode 100644 index 000000000..63c7a3ec2 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter_gccgo.go @@ -0,0 +1,25 @@ +// +build linux,gccgo + +package nsenter + +/* +#cgo CFLAGS: -Wall +extern void nsexec(); +void __attribute__((constructor)) init(void) { + nsexec(); +} +*/ +import "C" + +// AlwaysFalse is here to stay false +// (and be exported so the compiler doesn't optimize out its reference) +var AlwaysFalse bool + +func init() { + if AlwaysFalse { + // by referencing this C init() in a noop test, it will ensure the compiler + // links in the C function. + // https://gcc.gnu.org/bugzilla/show_bug.cgi?id=65134 + C.init() + } +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter_unsupported.go b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter_unsupported.go new file mode 100644 index 000000000..ac701ca39 --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsenter_unsupported.go @@ -0,0 +1,5 @@ +// +build !linux !cgo + +package nsenter + +import "C" diff --git a/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsexec.c b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsexec.c new file mode 100644 index 000000000..a4cd1399d --- /dev/null +++ b/vendor/github.com/opencontainers/runc/libcontainer/nsenter/nsexec.c @@ -0,0 +1,980 @@ + +#define _GNU_SOURCE +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include + +#include +#include +#include +#include +#include + +#include +#include +#include + +/* Get all of the CLONE_NEW* flags. */ +#include "namespace.h" + +/* Synchronisation values. */ +enum sync_t { + SYNC_USERMAP_PLS = 0x40, /* Request parent to map our users. */ + SYNC_USERMAP_ACK = 0x41, /* Mapping finished by the parent. */ + SYNC_RECVPID_PLS = 0x42, /* Tell parent we're sending the PID. */ + SYNC_RECVPID_ACK = 0x43, /* PID was correctly received by parent. */ + SYNC_GRANDCHILD = 0x44, /* The grandchild is ready to run. */ + SYNC_CHILD_READY = 0x45, /* The child or grandchild is ready to return. */ + + /* XXX: This doesn't help with segfaults and other such issues. */ + SYNC_ERR = 0xFF, /* Fatal error, no turning back. The error code follows. */ +}; + +/* longjmp() arguments. */ +#define JUMP_PARENT 0x00 +#define JUMP_CHILD 0xA0 +#define JUMP_INIT 0xA1 + +/* JSON buffer. */ +#define JSON_MAX 4096 + +/* Assume the stack grows down, so arguments should be above it. */ +struct clone_t { + /* + * Reserve some space for clone() to locate arguments + * and retcode in this place + */ + char stack[4096] __attribute__ ((aligned(16))); + char stack_ptr[0]; + + /* There's two children. This is used to execute the different code. */ + jmp_buf *env; + int jmpval; +}; + +struct nlconfig_t { + char *data; + + /* Process settings. */ + uint32_t cloneflags; + char *oom_score_adj; + size_t oom_score_adj_len; + + /* User namespace settings. */ + char *uidmap; + size_t uidmap_len; + char *gidmap; + size_t gidmap_len; + char *namespaces; + size_t namespaces_len; + uint8_t is_setgroup; + + /* Rootless container settings. */ + uint8_t is_rootless; + char *uidmappath; + size_t uidmappath_len; + char *gidmappath; + size_t gidmappath_len; +}; + +/* + * List of netlink message types sent to us as part of bootstrapping the init. + * These constants are defined in libcontainer/message_linux.go. + */ +#define INIT_MSG 62000 +#define CLONE_FLAGS_ATTR 27281 +#define NS_PATHS_ATTR 27282 +#define UIDMAP_ATTR 27283 +#define GIDMAP_ATTR 27284 +#define SETGROUP_ATTR 27285 +#define OOM_SCORE_ADJ_ATTR 27286 +#define ROOTLESS_ATTR 27287 +#define UIDMAPPATH_ATTR 27288 +#define GIDMAPPATH_ATTR 27289 + +/* + * Use the raw syscall for versions of glibc which don't include a function for + * it, namely (glibc 2.12). + */ +#if __GLIBC__ == 2 && __GLIBC_MINOR__ < 14 +# define _GNU_SOURCE +# include "syscall.h" +# if !defined(SYS_setns) && defined(__NR_setns) +# define SYS_setns __NR_setns +# endif + +#ifndef SYS_setns +# error "setns(2) syscall not supported by glibc version" +#endif + +int setns(int fd, int nstype) +{ + return syscall(SYS_setns, fd, nstype); +} +#endif + +/* XXX: This is ugly. */ +static int syncfd = -1; + +/* TODO(cyphar): Fix this so it correctly deals with syncT. */ +#define bail(fmt, ...) \ + do { \ + int ret = __COUNTER__ + 1; \ + fprintf(stderr, "nsenter: " fmt ": %m\n", ##__VA_ARGS__); \ + if (syncfd >= 0) { \ + enum sync_t s = SYNC_ERR; \ + if (write(syncfd, &s, sizeof(s)) != sizeof(s)) \ + fprintf(stderr, "nsenter: failed: write(s)"); \ + if (write(syncfd, &ret, sizeof(ret)) != sizeof(ret)) \ + fprintf(stderr, "nsenter: failed: write(ret)"); \ + } \ + exit(ret); \ + } while(0) + +static int write_file(char *data, size_t data_len, char *pathfmt, ...) +{ + int fd, len, ret = 0; + char path[PATH_MAX]; + + va_list ap; + va_start(ap, pathfmt); + len = vsnprintf(path, PATH_MAX, pathfmt, ap); + va_end(ap); + if (len < 0) + return -1; + + fd = open(path, O_RDWR); + if (fd < 0) { + return -1; + } + + len = write(fd, data, data_len); + if (len != data_len) { + ret = -1; + goto out; + } + + out: + close(fd); + return ret; +} + +enum policy_t { + SETGROUPS_DEFAULT = 0, + SETGROUPS_ALLOW, + SETGROUPS_DENY, +}; + +/* This *must* be called before we touch gid_map. */ +static void update_setgroups(int pid, enum policy_t setgroup) +{ + char *policy; + + switch (setgroup) { + case SETGROUPS_ALLOW: + policy = "allow"; + break; + case SETGROUPS_DENY: + policy = "deny"; + break; + case SETGROUPS_DEFAULT: + default: + /* Nothing to do. */ + return; + } + + if (write_file(policy, strlen(policy), "/proc/%d/setgroups", pid) < 0) { + /* + * If the kernel is too old to support /proc/pid/setgroups, + * open(2) or write(2) will return ENOENT. This is fine. + */ + if (errno != ENOENT) + bail("failed to write '%s' to /proc/%d/setgroups", policy, pid); + } +} + +static int try_mapping_tool(const char *app, int pid, char *map, size_t map_len) +{ + int child; + + /* + * If @app is NULL, execve will segfault. Just check it here and bail (if + * we're in this path, the caller is already getting desparate and there + * isn't a backup to this failing). This usually would be a configuration + * or programming issue. + */ + if (!app) + bail("mapping tool not present"); + + child = fork(); + if (child < 0) + bail("failed to fork"); + + if (!child) { +#define MAX_ARGV 20 + char *argv[MAX_ARGV]; + char *envp[] = { NULL }; + char pid_fmt[16]; + int argc = 0; + char *next; + + snprintf(pid_fmt, 16, "%d", pid); + + argv[argc++] = (char *)app; + argv[argc++] = pid_fmt; + /* + * Convert the map string into a list of argument that + * newuidmap/newgidmap can understand. + */ + + while (argc < MAX_ARGV) { + if (*map == '\0') { + argv[argc++] = NULL; + break; + } + argv[argc++] = map; + next = strpbrk(map, "\n "); + if (next == NULL) + break; + *next++ = '\0'; + map = next + strspn(next, "\n "); + } + + execve(app, argv, envp); + bail("failed to execv"); + } else { + int status; + + while (true) { + if (waitpid(child, &status, 0) < 0) { + if (errno == EINTR) + continue; + bail("failed to waitpid"); + } + if (WIFEXITED(status) || WIFSIGNALED(status)) + return WEXITSTATUS(status); + } + } + + return -1; +} + +static void update_uidmap(const char *path, int pid, char *map, size_t map_len) +{ + if (map == NULL || map_len <= 0) + return; + + if (write_file(map, map_len, "/proc/%d/uid_map", pid) < 0) { + if (errno != EPERM) + bail("failed to update /proc/%d/uid_map", pid); + if (try_mapping_tool(path, pid, map, map_len)) + bail("failed to use newuid map on %d", pid); + } +} + +static void update_gidmap(const char *path, int pid, char *map, size_t map_len) +{ + if (map == NULL || map_len <= 0) + return; + + if (write_file(map, map_len, "/proc/%d/gid_map", pid) < 0) { + if (errno != EPERM) + bail("failed to update /proc/%d/gid_map", pid); + if (try_mapping_tool(path, pid, map, map_len)) + bail("failed to use newgid map on %d", pid); + } +} + +static void update_oom_score_adj(char *data, size_t len) +{ + if (data == NULL || len <= 0) + return; + + if (write_file(data, len, "/proc/self/oom_score_adj") < 0) + bail("failed to update /proc/self/oom_score_adj"); +} + +/* A dummy function that just jumps to the given jumpval. */ +static int child_func(void *arg) __attribute__ ((noinline)); +static int child_func(void *arg) +{ + struct clone_t *ca = (struct clone_t *)arg; + longjmp(*ca->env, ca->jmpval); +} + +static int clone_parent(jmp_buf *env, int jmpval) __attribute__ ((noinline)); +static int clone_parent(jmp_buf *env, int jmpval) +{ + struct clone_t ca = { + .env = env, + .jmpval = jmpval, + }; + + return clone(child_func, ca.stack_ptr, CLONE_PARENT | SIGCHLD, &ca); +} + +/* + * Gets the init pipe fd from the environment, which is used to read the + * bootstrap data and tell the parent what the new pid is after we finish + * setting up the environment. + */ +static int initpipe(void) +{ + int pipenum; + char *initpipe, *endptr; + + initpipe = getenv("_LIBCONTAINER_INITPIPE"); + if (initpipe == NULL || *initpipe == '\0') + return -1; + + pipenum = strtol(initpipe, &endptr, 10); + if (*endptr != '\0') + bail("unable to parse _LIBCONTAINER_INITPIPE"); + + return pipenum; +} + +/* Returns the clone(2) flag for a namespace, given the name of a namespace. */ +static int nsflag(char *name) +{ + if (!strcmp(name, "cgroup")) + return CLONE_NEWCGROUP; + else if (!strcmp(name, "ipc")) + return CLONE_NEWIPC; + else if (!strcmp(name, "mnt")) + return CLONE_NEWNS; + else if (!strcmp(name, "net")) + return CLONE_NEWNET; + else if (!strcmp(name, "pid")) + return CLONE_NEWPID; + else if (!strcmp(name, "user")) + return CLONE_NEWUSER; + else if (!strcmp(name, "uts")) + return CLONE_NEWUTS; + + /* If we don't recognise a name, fallback to 0. */ + return 0; +} + +static uint32_t readint32(char *buf) +{ + return *(uint32_t *) buf; +} + +static uint8_t readint8(char *buf) +{ + return *(uint8_t *) buf; +} + +static void nl_parse(int fd, struct nlconfig_t *config) +{ + size_t len, size; + struct nlmsghdr hdr; + char *data, *current; + + /* Retrieve the netlink header. */ + len = read(fd, &hdr, NLMSG_HDRLEN); + if (len != NLMSG_HDRLEN) + bail("invalid netlink header length %zu", len); + + if (hdr.nlmsg_type == NLMSG_ERROR) + bail("failed to read netlink message"); + + if (hdr.nlmsg_type != INIT_MSG) + bail("unexpected msg type %d", hdr.nlmsg_type); + + /* Retrieve data. */ + size = NLMSG_PAYLOAD(&hdr, 0); + current = data = malloc(size); + if (!data) + bail("failed to allocate %zu bytes of memory for nl_payload", size); + + len = read(fd, data, size); + if (len != size) + bail("failed to read netlink payload, %zu != %zu", len, size); + + /* Parse the netlink payload. */ + config->data = data; + while (current < data + size) { + struct nlattr *nlattr = (struct nlattr *)current; + size_t payload_len = nlattr->nla_len - NLA_HDRLEN; + + /* Advance to payload. */ + current += NLA_HDRLEN; + + /* Handle payload. */ + switch (nlattr->nla_type) { + case CLONE_FLAGS_ATTR: + config->cloneflags = readint32(current); + break; + case ROOTLESS_ATTR: + config->is_rootless = readint8(current); + break; + case OOM_SCORE_ADJ_ATTR: + config->oom_score_adj = current; + config->oom_score_adj_len = payload_len; + break; + case NS_PATHS_ATTR: + config->namespaces = current; + config->namespaces_len = payload_len; + break; + case UIDMAP_ATTR: + config->uidmap = current; + config->uidmap_len = payload_len; + break; + case GIDMAP_ATTR: + config->gidmap = current; + config->gidmap_len = payload_len; + break; + case UIDMAPPATH_ATTR: + config->uidmappath = current; + config->uidmappath_len = payload_len; + break; + case GIDMAPPATH_ATTR: + config->gidmappath = current; + config->gidmappath_len = payload_len; + break; + case SETGROUP_ATTR: + config->is_setgroup = readint8(current); + break; + default: + bail("unknown netlink message type %d", nlattr->nla_type); + } + + current += NLA_ALIGN(payload_len); + } +} + +void nl_free(struct nlconfig_t *config) +{ + free(config->data); +} + +void join_namespaces(char *nslist) +{ + int num = 0, i; + char *saveptr = NULL; + char *namespace = strtok_r(nslist, ",", &saveptr); + struct namespace_t { + int fd; + int ns; + char type[PATH_MAX]; + char path[PATH_MAX]; + } *namespaces = NULL; + + if (!namespace || !strlen(namespace) || !strlen(nslist)) + bail("ns paths are empty"); + + /* + * We have to open the file descriptors first, since after + * we join the mnt namespace we might no longer be able to + * access the paths. + */ + do { + int fd; + char *path; + struct namespace_t *ns; + + /* Resize the namespace array. */ + namespaces = realloc(namespaces, ++num * sizeof(struct namespace_t)); + if (!namespaces) + bail("failed to reallocate namespace array"); + ns = &namespaces[num - 1]; + + /* Split 'ns:path'. */ + path = strstr(namespace, ":"); + if (!path) + bail("failed to parse %s", namespace); + *path++ = '\0'; + + fd = open(path, O_RDONLY); + if (fd < 0) + bail("failed to open %s", path); + + ns->fd = fd; + ns->ns = nsflag(namespace); + strncpy(ns->path, path, PATH_MAX - 1); + ns->path[PATH_MAX - 1] = '\0'; + } while ((namespace = strtok_r(NULL, ",", &saveptr)) != NULL); + + /* + * The ordering in which we join namespaces is important. We should + * always join the user namespace *first*. This is all guaranteed + * from the container_linux.go side of this, so we're just going to + * follow the order given to us. + */ + + for (i = 0; i < num; i++) { + struct namespace_t ns = namespaces[i]; + + if (setns(ns.fd, ns.ns) < 0) + bail("failed to setns to %s", ns.path); + + close(ns.fd); + } + + free(namespaces); +} + +void nsexec(void) +{ + int pipenum; + jmp_buf env; + int sync_child_pipe[2], sync_grandchild_pipe[2]; + struct nlconfig_t config = { 0 }; + + /* + * If we don't have an init pipe, just return to the go routine. + * We'll only get an init pipe for start or exec. + */ + pipenum = initpipe(); + if (pipenum == -1) + return; + + /* Parse all of the netlink configuration. */ + nl_parse(pipenum, &config); + + /* Set oom_score_adj. This has to be done before !dumpable because + * /proc/self/oom_score_adj is not writeable unless you're an privileged + * user (if !dumpable is set). All children inherit their parent's + * oom_score_adj value on fork(2) so this will always be propagated + * properly. + */ + update_oom_score_adj(config.oom_score_adj, config.oom_score_adj_len); + + /* + * Make the process non-dumpable, to avoid various race conditions that + * could cause processes in namespaces we're joining to access host + * resources (or potentially execute code). + * + * However, if the number of namespaces we are joining is 0, we are not + * going to be switching to a different security context. Thus setting + * ourselves to be non-dumpable only breaks things (like rootless + * containers), which is the recommendation from the kernel folks. + */ + if (config.namespaces) { + if (prctl(PR_SET_DUMPABLE, 0, 0, 0, 0) < 0) + bail("failed to set process as non-dumpable"); + } + + /* Pipe so we can tell the child when we've finished setting up. */ + if (socketpair(AF_LOCAL, SOCK_STREAM, 0, sync_child_pipe) < 0) + bail("failed to setup sync pipe between parent and child"); + + /* + * We need a new socketpair to sync with grandchild so we don't have + * race condition with child. + */ + if (socketpair(AF_LOCAL, SOCK_STREAM, 0, sync_grandchild_pipe) < 0) + bail("failed to setup sync pipe between parent and grandchild"); + + /* TODO: Currently we aren't dealing with child deaths properly. */ + + /* + * Okay, so this is quite annoying. + * + * In order for this unsharing code to be more extensible we need to split + * up unshare(CLONE_NEWUSER) and clone() in various ways. The ideal case + * would be if we did clone(CLONE_NEWUSER) and the other namespaces + * separately, but because of SELinux issues we cannot really do that. But + * we cannot just dump the namespace flags into clone(...) because several + * usecases (such as rootless containers) require more granularity around + * the namespace setup. In addition, some older kernels had issues where + * CLONE_NEWUSER wasn't handled before other namespaces (but we cannot + * handle this while also dealing with SELinux so we choose SELinux support + * over broken kernel support). + * + * However, if we unshare(2) the user namespace *before* we clone(2), then + * all hell breaks loose. + * + * The parent no longer has permissions to do many things (unshare(2) drops + * all capabilities in your old namespace), and the container cannot be set + * up to have more than one {uid,gid} mapping. This is obviously less than + * ideal. In order to fix this, we have to first clone(2) and then unshare. + * + * Unfortunately, it's not as simple as that. We have to fork to enter the + * PID namespace (the PID namespace only applies to children). Since we'll + * have to double-fork, this clone_parent() call won't be able to get the + * PID of the _actual_ init process (without doing more synchronisation than + * I can deal with at the moment). So we'll just get the parent to send it + * for us, the only job of this process is to update + * /proc/pid/{setgroups,uid_map,gid_map}. + * + * And as a result of the above, we also need to setns(2) in the first child + * because if we join a PID namespace in the topmost parent then our child + * will be in that namespace (and it will not be able to give us a PID value + * that makes sense without resorting to sending things with cmsg). + * + * This also deals with an older issue caused by dumping cloneflags into + * clone(2): On old kernels, CLONE_PARENT didn't work with CLONE_NEWPID, so + * we have to unshare(2) before clone(2) in order to do this. This was fixed + * in upstream commit 1f7f4dde5c945f41a7abc2285be43d918029ecc5, and was + * introduced by 40a0d32d1eaffe6aac7324ca92604b6b3977eb0e. As far as we're + * aware, the last mainline kernel which had this bug was Linux 3.12. + * However, we cannot comment on which kernels the broken patch was + * backported to. + * + * -- Aleksa "what has my life come to?" Sarai + */ + + switch (setjmp(env)) { + /* + * Stage 0: We're in the parent. Our job is just to create a new child + * (stage 1: JUMP_CHILD) process and write its uid_map and + * gid_map. That process will go on to create a new process, then + * it will send us its PID which we will send to the bootstrap + * process. + */ + case JUMP_PARENT:{ + int len; + pid_t child, first_child = -1; + char buf[JSON_MAX]; + bool ready = false; + + /* For debugging. */ + prctl(PR_SET_NAME, (unsigned long)"runc:[0:PARENT]", 0, 0, 0); + + /* Start the process of getting a container. */ + child = clone_parent(&env, JUMP_CHILD); + if (child < 0) + bail("unable to fork: child_func"); + + /* + * State machine for synchronisation with the children. + * + * Father only return when both child and grandchild are + * ready, so we can receive all possible error codes + * generated by children. + */ + while (!ready) { + enum sync_t s; + int ret; + + syncfd = sync_child_pipe[1]; + close(sync_child_pipe[0]); + + if (read(syncfd, &s, sizeof(s)) != sizeof(s)) + bail("failed to sync with child: next state"); + + switch (s) { + case SYNC_ERR: + /* We have to mirror the error code of the child. */ + if (read(syncfd, &ret, sizeof(ret)) != sizeof(ret)) + bail("failed to sync with child: read(error code)"); + + exit(ret); + case SYNC_USERMAP_PLS: + /* + * Enable setgroups(2) if we've been asked to. But we also + * have to explicitly disable setgroups(2) if we're + * creating a rootless container for single-entry mapping. + * i.e. config.is_setgroup == false. + * (this is required since Linux 3.19). + * + * For rootless multi-entry mapping, config.is_setgroup shall be true and + * newuidmap/newgidmap shall be used. + */ + + if (config.is_rootless && !config.is_setgroup) + update_setgroups(child, SETGROUPS_DENY); + + /* Set up mappings. */ + update_uidmap(config.uidmappath, child, config.uidmap, config.uidmap_len); + update_gidmap(config.gidmappath, child, config.gidmap, config.gidmap_len); + + s = SYNC_USERMAP_ACK; + if (write(syncfd, &s, sizeof(s)) != sizeof(s)) { + kill(child, SIGKILL); + bail("failed to sync with child: write(SYNC_USERMAP_ACK)"); + } + break; + case SYNC_RECVPID_PLS:{ + first_child = child; + + /* Get the init_func pid. */ + if (read(syncfd, &child, sizeof(child)) != sizeof(child)) { + kill(first_child, SIGKILL); + bail("failed to sync with child: read(childpid)"); + } + + /* Send ACK. */ + s = SYNC_RECVPID_ACK; + if (write(syncfd, &s, sizeof(s)) != sizeof(s)) { + kill(first_child, SIGKILL); + kill(child, SIGKILL); + bail("failed to sync with child: write(SYNC_RECVPID_ACK)"); + } + } + break; + case SYNC_CHILD_READY: + ready = true; + break; + default: + bail("unexpected sync value: %u", s); + } + } + + /* Now sync with grandchild. */ + + ready = false; + while (!ready) { + enum sync_t s; + int ret; + + syncfd = sync_grandchild_pipe[1]; + close(sync_grandchild_pipe[0]); + + s = SYNC_GRANDCHILD; + if (write(syncfd, &s, sizeof(s)) != sizeof(s)) { + kill(child, SIGKILL); + bail("failed to sync with child: write(SYNC_GRANDCHILD)"); + } + + if (read(syncfd, &s, sizeof(s)) != sizeof(s)) + bail("failed to sync with child: next state"); + + switch (s) { + case SYNC_ERR: + /* We have to mirror the error code of the child. */ + if (read(syncfd, &ret, sizeof(ret)) != sizeof(ret)) + bail("failed to sync with child: read(error code)"); + + exit(ret); + case SYNC_CHILD_READY: + ready = true; + break; + default: + bail("unexpected sync value: %u", s); + } + } + + /* + * Send the init_func pid and the pid of the first child back to our parent. + * + * We need to send both back because we can't reap the first child we created (CLONE_PARENT). + * It becomes the responsibility of our parent to reap the first child. + */ + len = snprintf(buf, JSON_MAX, "{\"pid\": %d, \"pid_first\": %d}\n", child, first_child); + if (len < 0) { + kill(child, SIGKILL); + bail("unable to generate JSON for child pid"); + } + if (write(pipenum, buf, len) != len) { + kill(child, SIGKILL); + bail("unable to send child pid to bootstrapper"); + } + + exit(0); + } + + /* + * Stage 1: We're in the first child process. Our job is to join any + * provided namespaces in the netlink payload and unshare all + * of the requested namespaces. If we've been asked to + * CLONE_NEWUSER, we will ask our parent (stage 0) to set up + * our user mappings for us. Then, we create a new child + * (stage 2: JUMP_INIT) for PID namespace. We then send the + * child's PID to our parent (stage 0). + */ + case JUMP_CHILD:{ + pid_t child; + enum sync_t s; + + /* We're in a child and thus need to tell the parent if we die. */ + syncfd = sync_child_pipe[0]; + close(sync_child_pipe[1]); + + /* For debugging. */ + prctl(PR_SET_NAME, (unsigned long)"runc:[1:CHILD]", 0, 0, 0); + + /* + * We need to setns first. We cannot do this earlier (in stage 0) + * because of the fact that we forked to get here (the PID of + * [stage 2: JUMP_INIT]) would be meaningless). We could send it + * using cmsg(3) but that's just annoying. + */ + if (config.namespaces) + join_namespaces(config.namespaces); + + /* + * Deal with user namespaces first. They are quite special, as they + * affect our ability to unshare other namespaces and are used as + * context for privilege checks. + * + * We don't unshare all namespaces in one go. The reason for this + * is that, while the kernel documentation may claim otherwise, + * there are certain cases where unsharing all namespaces at once + * will result in namespace objects being owned incorrectly. + * Ideally we should just fix these kernel bugs, but it's better to + * be safe than sorry, and fix them separately. + * + * A specific case of this is that the SELinux label of the + * internal kern-mount that mqueue uses will be incorrect if the + * UTS namespace is cloned before the USER namespace is mapped. + * I've also heard of similar problems with the network namespace + * in some scenarios. This also mirrors how LXC deals with this + * problem. + */ + if (config.cloneflags & CLONE_NEWUSER) { + if (unshare(CLONE_NEWUSER) < 0) + bail("failed to unshare user namespace"); + config.cloneflags &= ~CLONE_NEWUSER; + + /* + * We don't have the privileges to do any mapping here (see the + * clone_parent rant). So signal our parent to hook us up. + */ + + /* Switching is only necessary if we joined namespaces. */ + if (config.namespaces) { + if (prctl(PR_SET_DUMPABLE, 1, 0, 0, 0) < 0) + bail("failed to set process as dumpable"); + } + s = SYNC_USERMAP_PLS; + if (write(syncfd, &s, sizeof(s)) != sizeof(s)) + bail("failed to sync with parent: write(SYNC_USERMAP_PLS)"); + + /* ... wait for mapping ... */ + + if (read(syncfd, &s, sizeof(s)) != sizeof(s)) + bail("failed to sync with parent: read(SYNC_USERMAP_ACK)"); + if (s != SYNC_USERMAP_ACK) + bail("failed to sync with parent: SYNC_USERMAP_ACK: got %u", s); + /* Switching is only necessary if we joined namespaces. */ + if (config.namespaces) { + if (prctl(PR_SET_DUMPABLE, 0, 0, 0, 0) < 0) + bail("failed to set process as dumpable"); + } + + /* Become root in the namespace proper. */ + if (setresuid(0, 0, 0) < 0) + bail("failed to become root in user namespace"); + } + + /* + * Unshare all of the namespaces. Note that we don't merge this + * with clone() because there were some old kernel versions where + * clone(CLONE_PARENT | CLONE_NEWPID) was broken, so we'll just do + * it the long way. + */ + if (unshare(config.cloneflags) < 0) + bail("failed to unshare namespaces"); + + /* + * TODO: What about non-namespace clone flags that we're dropping here? + * + * We fork again because of PID namespace, setns(2) or unshare(2) don't + * change the PID namespace of the calling process, because doing so + * would change the caller's idea of its own PID (as reported by getpid()), + * which would break many applications and libraries, so we must fork + * to actually enter the new PID namespace. + */ + child = clone_parent(&env, JUMP_INIT); + if (child < 0) + bail("unable to fork: init_func"); + + /* Send the child to our parent, which knows what it's doing. */ + s = SYNC_RECVPID_PLS; + if (write(syncfd, &s, sizeof(s)) != sizeof(s)) { + kill(child, SIGKILL); + bail("failed to sync with parent: write(SYNC_RECVPID_PLS)"); + } + if (write(syncfd, &child, sizeof(child)) != sizeof(child)) { + kill(child, SIGKILL); + bail("failed to sync with parent: write(childpid)"); + } + + /* ... wait for parent to get the pid ... */ + + if (read(syncfd, &s, sizeof(s)) != sizeof(s)) { + kill(child, SIGKILL); + bail("failed to sync with parent: read(SYNC_RECVPID_ACK)"); + } + if (s != SYNC_RECVPID_ACK) { + kill(child, SIGKILL); + bail("failed to sync with parent: SYNC_RECVPID_ACK: got %u", s); + } + + s = SYNC_CHILD_READY; + if (write(syncfd, &s, sizeof(s)) != sizeof(s)) { + kill(child, SIGKILL); + bail("failed to sync with parent: write(SYNC_CHILD_READY)"); + } + + /* Our work is done. [Stage 2: JUMP_INIT] is doing the rest of the work. */ + exit(0); + } + + /* + * Stage 2: We're the final child process, and the only process that will + * actually return to the Go runtime. Our job is to just do the + * final cleanup steps and then return to the Go runtime to allow + * init_linux.go to run. + */ + case JUMP_INIT:{ + /* + * We're inside the child now, having jumped from the + * start_child() code after forking in the parent. + */ + enum sync_t s; + + /* We're in a child and thus need to tell the parent if we die. */ + syncfd = sync_grandchild_pipe[0]; + close(sync_grandchild_pipe[1]); + close(sync_child_pipe[0]); + close(sync_child_pipe[1]); + + /* For debugging. */ + prctl(PR_SET_NAME, (unsigned long)"runc:[2:INIT]", 0, 0, 0); + + if (read(syncfd, &s, sizeof(s)) != sizeof(s)) + bail("failed to sync with parent: read(SYNC_GRANDCHILD)"); + if (s != SYNC_GRANDCHILD) + bail("failed to sync with parent: SYNC_GRANDCHILD: got %u", s); + + if (setsid() < 0) + bail("setsid failed"); + + if (setuid(0) < 0) + bail("setuid failed"); + + if (setgid(0) < 0) + bail("setgid failed"); + + if (!config.is_rootless && config.is_setgroup) { + if (setgroups(0, NULL) < 0) + bail("setgroups failed"); + } + + s = SYNC_CHILD_READY; + if (write(syncfd, &s, sizeof(s)) != sizeof(s)) + bail("failed to sync with patent: write(SYNC_CHILD_READY)"); + + /* Close sync pipes. */ + close(sync_grandchild_pipe[0]); + + /* Free netlink data. */ + nl_free(&config); + + /* Finish executing, let the Go runtime take over. */ + return; + } + default: + bail("unexpected jump value"); + } + + /* Should never be reached. */ + bail("should never be reached"); +} diff --git a/vendor/github.com/opencontainers/runc/libcontainer/rootfs_linux.go b/vendor/github.com/opencontainers/runc/libcontainer/rootfs_linux.go index cf715d664..dcb340782 100644 --- a/vendor/github.com/opencontainers/runc/libcontainer/rootfs_linux.go +++ b/vendor/github.com/opencontainers/runc/libcontainer/rootfs_linux.go @@ -42,6 +42,12 @@ func needsSetupDev(config *configs.Config) bool { // finalizeRootfs after this function to finish setting up the rootfs. func prepareRootfs(pipe io.ReadWriter, iConfig *initConfig) (err error) { config := iConfig.Config + if config.Rootfs == "/" { + if err := unix.Chdir(config.Rootfs); err != nil { + return newSystemErrorWithCausef(err, "changing dir to %q", config.Rootfs) + } + return nil + } if err := prepareRoot(config); err != nil { return newSystemErrorWithCause(err, "preparing rootfs") } @@ -129,6 +135,9 @@ func prepareRootfs(pipe io.ReadWriter, iConfig *initConfig) (err error) { // finalizeRootfs sets anything to ro if necessary. You must call // prepareRootfs first. func finalizeRootfs(config *configs.Config) (err error) { + if config.Rootfs == "/" { + return nil + } // remount dev as ro if specified for _, m := range config.Mounts { if libcontainerUtils.CleanPath(m.Destination) == "/dev" { @@ -152,6 +161,26 @@ func finalizeRootfs(config *configs.Config) (err error) { return nil } +// /tmp has to be mounted as private to allow MS_MOVE to work in all situations +func prepareTmp(topTmpDir string) (string, error) { + tmpdir, err := ioutil.TempDir(topTmpDir, "runctop") + if err != nil { + return "", err + } + if err := unix.Mount(tmpdir, tmpdir, "bind", unix.MS_BIND, ""); err != nil { + return "", err + } + if err := unix.Mount("", tmpdir, "", uintptr(unix.MS_PRIVATE), ""); err != nil { + return "", err + } + return tmpdir, nil +} + +func cleanupTmp(tmpdir string) error { + unix.Unmount(tmpdir, 0) + return os.RemoveAll(tmpdir) +} + func mountCmd(cmd configs.Command) error { command := exec.Command(cmd.Path, cmd.Args[:]...) command.Env = cmd.Env @@ -199,7 +228,12 @@ func mountToRootfs(m *configs.Mount, rootfs, mountLabel string) error { } } if copyUp { - tmpDir, err = ioutil.TempDir("/tmp", "runctmpdir") + tmpdir, err := prepareTmp("/tmp") + if err != nil { + return newSystemErrorWithCause(err, "tmpcopyup: failed to setup tmpdir") + } + defer cleanupTmp(tmpdir) + tmpDir, err = ioutil.TempDir(tmpdir, "runctmpdir") if err != nil { return newSystemErrorWithCause(err, "tmpcopyup: failed to create tmpdir") } @@ -396,6 +430,7 @@ func checkMountDestination(rootfs, dest string) error { "/proc/stat", "/proc/swaps", "/proc/uptime", + "/proc/loadavg", "/proc/net/dev", } for _, valid := range validDestinations { @@ -412,7 +447,7 @@ func checkMountDestination(rootfs, dest string) error { if err != nil { return err } - if path == "." || !strings.HasPrefix(path, "..") { + if path != "." && !strings.HasPrefix(path, "..") { return fmt.Errorf("%q cannot be mounted because it is located inside %q", dest, invalid) } } diff --git a/vendor/vendor.json b/vendor/vendor.json index 2d9147b03..7af3c8d61 100644 --- a/vendor/vendor.json +++ b/vendor/vendor.json @@ -289,22 +289,23 @@ {"path":"github.com/opencontainers/go-digest","checksumSHA1":"NTperEHVh1uBqfTy9+oKceN4tKI=","revision":"21dfd564fd89c944783d00d069f33e3e7123c448","revisionTime":"2017-01-11T18:16:59Z"}, {"path":"github.com/opencontainers/image-spec/specs-go","checksumSHA1":"ZGlIwSRjdLYCUII7JLE++N4w7Xc=","revision":"89b51c794e9113108a2914e38e66c826a649f2b5","revisionTime":"2017-11-03T11:36:04Z"}, {"path":"github.com/opencontainers/image-spec/specs-go/v1","checksumSHA1":"jdbXRRzeu0njLE9/nCEZG+Yg/Jk=","revision":"89b51c794e9113108a2914e38e66c826a649f2b5","revisionTime":"2017-11-03T11:36:04Z"}, - {"path":"github.com/opencontainers/runc/libcontainer","checksumSHA1":"cqrf+HmlptrK2lRhQ0PkYdOe5cQ=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/apparmor","checksumSHA1":"gVVY8k2G3ws+V1czsfxfuRs8log=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/cgroups","checksumSHA1":"eA9qNw7Tkpi1GLojT/Vxa99aL00=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/cgroups/fs","checksumSHA1":"YgPihDRi3OxI0qv7CxTxrqZuvfU=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/cgroups/systemd","checksumSHA1":"RVQixM4pF56oCNymhNY67I5eS0Y=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/configs","checksumSHA1":"PUv5rdj6oEGJoBij/9Elx8VO6bs=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/configs/validate","checksumSHA1":"YJq+f9izqizSzG/OuVFUOfloNEk=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/criurpc","checksumSHA1":"I1iwXoDUJeDXviilCtkvDpEF/So=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/intelrdt","checksumSHA1":"bAWJX1XUDMd4GqPLSrCkUcdiTbg=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/keys","checksumSHA1":"QXuHZwxlqhoq/oHRJFbTi6+AWLY=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/mount","checksumSHA1":"MJiogPDUU2nFr1fzQU6T+Ry1W8o=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/seccomp","checksumSHA1":"I1Qw/btE1twMqKHpYNsC98cteak=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/stacktrace","checksumSHA1":"yp/kYBgVqKtxlnpq4CmyxLFMAE4=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/system","checksumSHA1":"cjg/UcueM1/2/ExZ3N7010sa+hI=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/user","checksumSHA1":"XtLpcP6ca9SQG218re7E7UcOj3Y=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, - {"path":"github.com/opencontainers/runc/libcontainer/utils","checksumSHA1":"Sb296YW4V+esieanrx4Nzt2L5lU=","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, + {"path":"github.com/opencontainers/runc/libcontainer","checksumSHA1":"6qP/ejjcc/+HelbVHmtl6cyUZmc=","origin":"github.com/hashicorp/runc/libcontainer","revision":"8df81be073884e4e4c613d893f5877310136820f","revisionTime":"2018-09-10T14:23:11Z","version":"nomad","versionExact":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/apparmor","checksumSHA1":"gVVY8k2G3ws+V1czsfxfuRs8log=","origin":"github.com/hashicorp/runc/libcontainer/apparmor","revision":"8df81be073884e4e4c613d893f5877310136820f","revisionTime":"2018-09-10T14:23:11Z","version":"nomad","versionExact":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/cgroups","checksumSHA1":"eA9qNw7Tkpi1GLojT/Vxa99aL00=","origin":"github.com/hashicorp/runc/libcontainer/cgroups","revision":"8df81be073884e4e4c613d893f5877310136820f","revisionTime":"2018-09-10T14:23:11Z","version":"nomad","versionExact":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/cgroups/fs","checksumSHA1":"YgPihDRi3OxI0qv7CxTxrqZuvfU=","origin":"github.com/hashicorp/runc/libcontainer/cgroups/fs","revision":"8df81be073884e4e4c613d893f5877310136820f","revisionTime":"2018-09-10T14:23:11Z","version":"nomad","versionExact":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/cgroups/systemd","checksumSHA1":"RVQixM4pF56oCNymhNY67I5eS0Y=","origin":"github.com/hashicorp/runc/libcontainer/cgroups/systemd","revision":"8df81be073884e4e4c613d893f5877310136820f","revisionTime":"2018-09-10T14:23:11Z","version":"nomad","versionExact":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/configs","checksumSHA1":"PUv5rdj6oEGJoBij/9Elx8VO6bs=","origin":"github.com/hashicorp/runc/libcontainer/configs","revision":"8df81be073884e4e4c613d893f5877310136820f","revisionTime":"2018-09-10T14:23:11Z","version":"nomad","versionExact":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/configs/validate","checksumSHA1":"YJq+f9izqizSzG/OuVFUOfloNEk=","origin":"github.com/hashicorp/runc/libcontainer/configs/validate","revision":"8df81be073884e4e4c613d893f5877310136820f","revisionTime":"2018-09-10T14:23:11Z","version":"nomad","versionExact":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/criurpc","checksumSHA1":"I1iwXoDUJeDXviilCtkvDpEF/So=","origin":"github.com/hashicorp/runc/libcontainer/criurpc","revision":"8df81be073884e4e4c613d893f5877310136820f","revisionTime":"2018-09-10T14:23:11Z","version":"nomad","versionExact":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/intelrdt","checksumSHA1":"bAWJX1XUDMd4GqPLSrCkUcdiTbg=","origin":"github.com/hashicorp/runc/libcontainer/intelrdt","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z","version":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/keys","checksumSHA1":"QXuHZwxlqhoq/oHRJFbTi6+AWLY=","origin":"github.com/hashicorp/runc/libcontainer/keys","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z","version":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/mount","checksumSHA1":"MJiogPDUU2nFr1fzQU6T+Ry1W8o=","origin":"github.com/hashicorp/runc/libcontainer/mount","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z","version":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/nsenter","checksumSHA1":"5p0YhO25gaLG+GUdTzqgvcRHjkE=","origin":"github.com/hashicorp/runc/libcontainer/nsenter","revision":"8df81be073884e4e4c613d893f5877310136820f","revisionTime":"2018-09-10T14:23:11Z","version":"nomad","versionExact":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/seccomp","checksumSHA1":"I1Qw/btE1twMqKHpYNsC98cteak=","origin":"github.com/hashicorp/runc/libcontainer/seccomp","revision":"8df81be073884e4e4c613d893f5877310136820f","revisionTime":"2018-09-10T14:23:11Z","version":"nomad","versionExact":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/stacktrace","checksumSHA1":"yp/kYBgVqKtxlnpq4CmyxLFMAE4=","origin":"github.com/hashicorp/runc/libcontainer/stacktrace","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z","version":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/system","checksumSHA1":"cjg/UcueM1/2/ExZ3N7010sa+hI=","origin":"github.com/hashicorp/runc/libcontainer/system","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z","version":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/user","checksumSHA1":"XtLpcP6ca9SQG218re7E7UcOj3Y=","origin":"github.com/hashicorp/runc/libcontainer/user","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z","version":"nomad"}, + {"path":"github.com/opencontainers/runc/libcontainer/utils","checksumSHA1":"Sb296YW4V+esieanrx4Nzt2L5lU=","origin":"github.com/hashicorp/runc/libcontainer/utils","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z","version":"nomad"}, {"path":"github.com/opencontainers/runtime-spec/specs-go","checksumSHA1":"AMYc2X2O/IL6EGrq6lTl5vEhLiY=","origin":"github.com/opencontainers/runc/vendor/github.com/opencontainers/runtime-spec/specs-go","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, {"path":"github.com/opencontainers/selinux/go-selinux","checksumSHA1":"j9efF9bPmCCag+LzqwjyB8a44B8=","origin":"github.com/opencontainers/runc/vendor/github.com/opencontainers/selinux/go-selinux","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, {"path":"github.com/opencontainers/selinux/go-selinux/label","checksumSHA1":"QbeVoKIoaJWZDH8V/588i8/Pjjs=","origin":"github.com/opencontainers/runc/vendor/github.com/opencontainers/selinux/go-selinux/label","revision":"459bfaec1fc6c17d8bfb12d0a0f69e7e7271ed2a","revisionTime":"2018-08-23T14:46:37Z"}, From 62c1285afccdf6d1aa1d2ca39dc2fdb6f40b6ca2 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 25 Sep 2018 10:56:12 -0700 Subject: [PATCH 072/152] tr: add comments and cleanup call signature From review comments on #4649 left post-merge. --- client/allocrunnerv2/taskrunner/stats_hook.go | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/stats_hook.go b/client/allocrunnerv2/taskrunner/stats_hook.go index a309e7c34..46dbee926 100644 --- a/client/allocrunnerv2/taskrunner/stats_hook.go +++ b/client/allocrunnerv2/taskrunner/stats_hook.go @@ -12,10 +12,13 @@ import ( cstructs "github.com/hashicorp/nomad/client/structs" ) +// StatsUpdater is the interface required by the StatsHook to update stats. +// Satisfied by TaskRunner. type StatsUpdater interface { UpdateStats(*cstructs.TaskResourceUsage) } +// statsHook manages the task stats collection goroutine. type statsHook struct { updater StatsUpdater interval time.Duration @@ -52,7 +55,7 @@ func (h *statsHook) Poststart(ctx context.Context, req *interfaces.TaskPoststart } h.stopCh = make(chan struct{}) - go h.collectResourceUsageStats(h.logger, req.DriverStats, h.stopCh) + go h.collectResourceUsageStats(req.DriverStats, h.stopCh) return nil } @@ -77,7 +80,7 @@ func (h *statsHook) Exited(context.Context, *interfaces.TaskExitedRequest, *inte // collectResourceUsageStats starts collecting resource usage stats of a Task. // Collection ends when the passed channel is closed -func (h *statsHook) collectResourceUsageStats(logger hclog.Logger, handle interfaces.DriverStats, stopCh <-chan struct{}) { +func (h *statsHook) collectResourceUsageStats(handle interfaces.DriverStats, stopCh <-chan struct{}) { // start collecting the stats right away and then start collecting every // collection interval next := time.NewTimer(0) From bcc5c4a8bda44ceb49907e75d7156c87e436d7c1 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 26 Sep 2018 13:33:37 -0400 Subject: [PATCH 073/152] clientv2: base driver plugin (#4671) Driver plugin framework to facilitate development of driver plugins. Implementing plugins only need to implement the DriverPlugin interface. The framework proxies this interface to the go-plugin GRPC interface generated from the driver.proto spec. A testing harness is provided to allow implementing drivers to test the full lifecycle of the driver plugin. An example use: func TestMyDriver(t *testing.T) { harness := NewDriverHarness(t, &MyDiverPlugin{}) // The harness implements the DriverPlugin interface and can be used as such taskHandle, err := harness.StartTask(...) } --- client/driver/executor/executor.go | 2 +- plugins/base/plugin.go | 5 + plugins/drivers/client.go | 302 +++++++ plugins/drivers/driver.go | 212 +++++ plugins/drivers/errors.go | 5 + plugins/drivers/plugin.go | 41 + plugins/drivers/plugin_test.go | 242 ++++++ plugins/drivers/{base => }/proto/driver.pb.go | 774 ++++++++---------- plugins/drivers/{base => }/proto/driver.proto | 31 +- plugins/drivers/server.go | 254 ++++++ plugins/drivers/task_handle.go | 30 + plugins/drivers/testing.go | 112 +++ plugins/drivers/testing_test.go | 24 + plugins/drivers/utils.go | 296 +++++++ plugins/drivers/utils/utils.go | 99 +++ plugins/drivers/utils/utils_unix.go | 18 + plugins/drivers/utils/utils_windows.go | 8 + 17 files changed, 1988 insertions(+), 467 deletions(-) create mode 100644 plugins/drivers/client.go create mode 100644 plugins/drivers/driver.go create mode 100644 plugins/drivers/errors.go create mode 100644 plugins/drivers/plugin.go create mode 100644 plugins/drivers/plugin_test.go rename plugins/drivers/{base => }/proto/driver.pb.go (77%) rename plugins/drivers/{base => }/proto/driver.proto (95%) create mode 100644 plugins/drivers/server.go create mode 100644 plugins/drivers/task_handle.go create mode 100644 plugins/drivers/testing.go create mode 100644 plugins/drivers/testing_test.go create mode 100644 plugins/drivers/utils.go create mode 100644 plugins/drivers/utils/utils.go create mode 100644 plugins/drivers/utils/utils_unix.go create mode 100644 plugins/drivers/utils/utils_windows.go diff --git a/client/driver/executor/executor.go b/client/driver/executor/executor.go index 280f107c5..62c3b71af 100644 --- a/client/driver/executor/executor.go +++ b/client/driver/executor/executor.go @@ -68,7 +68,7 @@ type Executor interface { // Version returns the executor API version Version() (*ExecutorVersion, error) - // Stats fetchs process usage stats for the executor and each pid if availble + // Stats fetchs process usage stats for the executor and each pid if available Stats() (*cstructs.TaskResourceUsage, error) // Signal sends the given signal to the user process diff --git a/plugins/base/plugin.go b/plugins/base/plugin.go index 79c151ea4..32a369900 100644 --- a/plugins/base/plugin.go +++ b/plugins/base/plugin.go @@ -61,3 +61,8 @@ var MsgpackHandle = func() *codec.MsgpackHandle { func MsgPackDecode(buf []byte, out interface{}) error { return codec.NewDecoder(bytes.NewReader(buf), MsgpackHandle).Decode(out) } + +// MsgPackEncode is used to encode an object to MsgPack +func MsgPackEncode(b *[]byte, in interface{}) error { + return codec.NewEncoderBytes(b, MsgpackHandle).Encode(in) +} diff --git a/plugins/drivers/client.go b/plugins/drivers/client.go new file mode 100644 index 000000000..1d1631caf --- /dev/null +++ b/plugins/drivers/client.go @@ -0,0 +1,302 @@ +package drivers + +import ( + "errors" + "fmt" + "io" + "time" + + "github.com/golang/protobuf/ptypes" + hclog "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/drivers/proto" + "github.com/hashicorp/nomad/plugins/shared/hclspec" + "golang.org/x/net/context" +) + +var _ DriverPlugin = &driverPluginClient{} + +type driverPluginClient struct { + base.BasePluginClient + + client proto.DriverClient + logger hclog.Logger +} + +func (d *driverPluginClient) TaskConfigSchema() (*hclspec.Spec, error) { + req := &proto.TaskConfigSchemaRequest{} + + resp, err := d.client.TaskConfigSchema(context.Background(), req) + if err != nil { + return nil, err + } + + return resp.Spec, nil +} + +func (d *driverPluginClient) Capabilities() (*Capabilities, error) { + req := &proto.CapabilitiesRequest{} + + resp, err := d.client.Capabilities(context.Background(), req) + if err != nil { + return nil, err + } + + caps := &Capabilities{} + if resp.Capabilities != nil { + caps.SendSignals = resp.Capabilities.SendSignals + caps.Exec = resp.Capabilities.Exec + + switch resp.Capabilities.FsIsolation { + case proto.DriverCapabilities_NONE: + caps.FSIsolation = FSIsolationNone + case proto.DriverCapabilities_CHROOT: + caps.FSIsolation = FSIsolationChroot + case proto.DriverCapabilities_IMAGE: + caps.FSIsolation = FSIsolationImage + default: + caps.FSIsolation = FSIsolationNone + } + } + + return caps, nil +} + +// Fingerprint the driver, return a chan that will be pushed to periodically and on changes to health +func (d *driverPluginClient) Fingerprint(ctx context.Context) (<-chan *Fingerprint, error) { + req := &proto.FingerprintRequest{} + + stream, err := d.client.Fingerprint(context.Background(), req) + if err != nil { + return nil, err + } + + ch := make(chan *Fingerprint) + go d.handleFingerprint(ch, stream) + + return ch, nil +} + +func (d *driverPluginClient) handleFingerprint(ch chan *Fingerprint, stream proto.Driver_FingerprintClient) { + defer close(ch) + for { + pb, err := stream.Recv() + if err == io.EOF { + break + } + if err != nil { + d.logger.Error("error receiving stream from Fingerprint driver RPC", "error", err) + ch <- &Fingerprint{Err: fmt.Errorf("error from RPC stream: %v", err)} + break + } + f := &Fingerprint{ + Attributes: pb.Attributes, + Health: healthStateFromProto(pb.Health), + HealthDescription: pb.HealthDescription, + } + ch <- f + } + +} + +// RecoverTask does internal state recovery to be able to control the task of +// the given TaskHandle +func (d *driverPluginClient) RecoverTask(h *TaskHandle) error { + req := &proto.RecoverTaskRequest{Handle: taskHandleToProto(h)} + + _, err := d.client.RecoverTask(context.Background(), req) + return err +} + +// StartTask starts execution of a task with the given TaskConfig. A TaskHandle +// is returned to the caller that can be used to recover state of the task, +// should the driver crash or exit prematurely. +func (d *driverPluginClient) StartTask(c *TaskConfig) (*TaskHandle, error) { + req := &proto.StartTaskRequest{ + Task: taskConfigToProto(c), + } + + resp, err := d.client.StartTask(context.Background(), req) + if err != nil { + return nil, err + } + + return taskHandleFromProto(resp.Handle), nil +} + +// WaitTask returns a channel that will have an ExitResult pushed to it once when the task +// exits on its own or is killed. If WaitTask is called after the task has exited, the channel +// will immedialy return the ExitResult. WaitTask can be called multiple times for +// the same task without issue. +func (d *driverPluginClient) WaitTask(ctx context.Context, id string) (<-chan *ExitResult, error) { + ch := make(chan *ExitResult) + go d.handleWaitTask(ctx, id, ch) + return ch, nil +} + +func (d *driverPluginClient) handleWaitTask(ctx context.Context, id string, ch chan *ExitResult) { + defer close(ch) + var result ExitResult + req := &proto.WaitTaskRequest{ + TaskId: id, + } + + resp, err := d.client.WaitTask(ctx, req) + if err != nil { + result.Err = err + } else { + result.ExitCode = int(resp.Result.ExitCode) + result.Signal = int(resp.Result.Signal) + result.OOMKilled = resp.Result.OomKilled + if len(resp.Err) > 0 { + result.Err = errors.New(resp.Err) + } + } + ch <- &result +} + +// StopTask stops the task with the given taskID. A timeout and signal can be +// given to control a graceful termination of the task. The driver will send the +// given signal to the task and wait for the given timeout for it to exit. If the +// task does not exit within the timeout it will be forcefully killed. +func (d *driverPluginClient) StopTask(taskID string, timeout time.Duration, signal string) error { + req := &proto.StopTaskRequest{ + TaskId: taskID, + Timeout: ptypes.DurationProto(timeout), + Signal: signal, + } + + _, err := d.client.StopTask(context.Background(), req) + return err +} + +// DestroyTask removes the task from the driver's in memory state. The task +// cannot be running unless force is set to true. If force is set to true the +// driver will forcefully terminate the task before removing it. +func (d *driverPluginClient) DestroyTask(taskID string, force bool) error { + req := &proto.DestroyTaskRequest{ + TaskId: taskID, + Force: force, + } + + _, err := d.client.DestroyTask(context.Background(), req) + return err +} + +// InspectTask returns status information for a task +func (d *driverPluginClient) InspectTask(taskID string) (*TaskStatus, error) { + req := &proto.InspectTaskRequest{TaskId: taskID} + + resp, err := d.client.InspectTask(context.Background(), req) + if err != nil { + return nil, err + } + + status, err := taskStatusFromProto(resp.Task) + if err != nil { + return nil, err + } + + if resp.Driver != nil { + status.DriverAttributes = resp.Driver.Attributes + } + if resp.NetworkOverride != nil { + status.NetworkOverride = &NetworkOverride{ + PortMap: resp.NetworkOverride.PortMap, + Addr: resp.NetworkOverride.Addr, + AutoAdvertise: resp.NetworkOverride.AutoAdvertise, + } + } + + return status, nil +} + +// TaskStats returns resource usage statistics for the task +func (d *driverPluginClient) TaskStats(taskID string) (*TaskStats, error) { + req := &proto.TaskStatsRequest{TaskId: taskID} + + resp, err := d.client.TaskStats(context.Background(), req) + if err != nil { + return nil, err + } + + stats, err := taskStatsFromProto(resp.Stats) + if err != nil { + return nil, err + } + + return stats, nil +} + +// TaskEvents returns a channel that will receive events from the driver about all +// tasks such as lifecycle events, terminal errors, etc. +func (d *driverPluginClient) TaskEvents(ctx context.Context) (<-chan *TaskEvent, error) { + req := &proto.TaskEventsRequest{} + stream, err := d.client.TaskEvents(ctx, req) + if err != nil { + return nil, err + } + + ch := make(chan *TaskEvent) + go d.handleTaskEvents(ch, stream) + return ch, nil +} + +func (d *driverPluginClient) handleTaskEvents(ch chan *TaskEvent, stream proto.Driver_TaskEventsClient) { + defer close(ch) + for { + ev, err := stream.Recv() + if err == io.EOF { + break + } + if err != nil { + d.logger.Error("error receiving stream from TaskEvents driver RPC", "error", err) + ch <- &TaskEvent{Err: err} + break + } + timestamp, _ := ptypes.Timestamp(ev.Timestamp) + event := &TaskEvent{ + TaskID: ev.TaskId, + Annotations: ev.Annotations, + Message: ev.Message, + Timestamp: timestamp, + } + ch <- event + } +} + +// SignalTask will send the given signal to the specified task +func (d *driverPluginClient) SignalTask(taskID string, signal string) error { + req := &proto.SignalTaskRequest{ + TaskId: taskID, + Signal: signal, + } + _, err := d.client.SignalTask(context.Background(), req) + return err +} + +// ExecTask will run the given command within the execution context of the task. +// The driver will wait for the given timeout for the command to complete before +// terminating it. The stdout and stderr of the command will be return to the caller, +// along with other exit information such as exit code. +func (d *driverPluginClient) ExecTask(taskID string, cmd []string, timeout time.Duration) (*ExecTaskResult, error) { + req := &proto.ExecTaskRequest{ + TaskId: taskID, + Command: cmd, + Timeout: ptypes.DurationProto(timeout), + } + + resp, err := d.client.ExecTask(context.Background(), req) + if err != nil { + return nil, err + } + + result := &ExecTaskResult{ + Stdout: resp.Stdout, + Stderr: resp.Stderr, + ExitResult: exitResultFromProto(resp.Result), + } + + return result, nil + +} diff --git a/plugins/drivers/driver.go b/plugins/drivers/driver.go new file mode 100644 index 000000000..94c14851f --- /dev/null +++ b/plugins/drivers/driver.go @@ -0,0 +1,212 @@ +package drivers + +import ( + "fmt" + "path/filepath" + "sort" + "time" + + "github.com/hashicorp/nomad/client/allocdir" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/shared/hclspec" + "golang.org/x/net/context" +) + +// DriverPlugin is the interface with drivers will implement. It is also +// implemented by a plugin client which proxies the calls to go-plugin. See +// the proto/driver.proto file for detailed information about each RPC and +// message structure. +type DriverPlugin interface { + base.BasePlugin + + TaskConfigSchema() (*hclspec.Spec, error) + Capabilities() (*Capabilities, error) + Fingerprint(context.Context) (<-chan *Fingerprint, error) + + RecoverTask(*TaskHandle) error + StartTask(*TaskConfig) (*TaskHandle, error) + WaitTask(ctx context.Context, taskID string) (<-chan *ExitResult, error) + StopTask(taskID string, timeout time.Duration, signal string) error + DestroyTask(taskID string, force bool) error + InspectTask(taskID string) (*TaskStatus, error) + TaskStats(taskID string) (*TaskStats, error) + TaskEvents(context.Context) (<-chan *TaskEvent, error) + + SignalTask(taskID string, signal string) error + ExecTask(taskID string, cmd []string, timeout time.Duration) (*ExecTaskResult, error) +} + +// DriverSignalTaskNotSupported can be embedded by drivers which don't support +// the SignalTask RPC. This satisfies the SignalTask func requirement for the +// DriverPlugin interface. +type DriverSignalTaskNotSupported struct{} + +func (_ DriverSignalTaskNotSupported) SignalTask(taskID, signal string) error { + return fmt.Errorf("SignalTask is not supported by this driver") +} + +// DriverExecTaskNotSupported can be embedded by drivers which don't support +// the ExecTask RPC. This satisfies the ExecTask func requirement of the +// DriverPlugin interface. +type DriverExecTaskNotSupported struct{} + +func (_ DriverExecTaskNotSupported) ExecTask(taskID, signal string) error { + return fmt.Errorf("ExecTask is not supported by this driver") +} + +type HealthState string + +var ( + HealthStateUndetected = HealthState("undetected") + HealthStateUnhealthy = HealthState("unhealthy") + HealthStateHealthy = HealthState("healthy") +) + +type Fingerprint struct { + Attributes map[string]string + Health HealthState + HealthDescription string + + // Err is set by the plugin if an error occured during fingerprinting + Err error +} + +type FSIsolation string + +var ( + FSIsolationNone = FSIsolation("none") + FSIsolationChroot = FSIsolation("chroot") + FSIsolationImage = FSIsolation("image") +) + +type Capabilities struct { + // SendSignals marks the driver as being able to send signals + SendSignals bool + + // Exec marks the driver as being able to execute arbitrary commands + // such as health checks. Used by the ScriptExecutor interface. + Exec bool + + //FSIsolation indicates what kind of filesystem isolation the driver supports. + FSIsolation FSIsolation +} + +type TaskConfig struct { + ID string + Name string + Env map[string]string + Resources Resources + Devices []DeviceConfig + Mounts []MountConfig + User string + AllocDir string + rawDriverConfig []byte +} + +func (tc *TaskConfig) EnvList() []string { + l := make([]string, 0, len(tc.Env)) + for k, v := range tc.Env { + l = append(l, k+"="+v) + } + + sort.Strings(l) + return l +} + +func (tc *TaskConfig) TaskDir() *allocdir.TaskDir { + taskDir := filepath.Join(tc.AllocDir, tc.Name) + return &allocdir.TaskDir{ + Dir: taskDir, + SharedAllocDir: filepath.Join(tc.AllocDir, allocdir.SharedAllocName), + LogDir: filepath.Join(tc.AllocDir, allocdir.SharedAllocName, allocdir.LogDirName), + SharedTaskDir: filepath.Join(taskDir, allocdir.SharedAllocName), + LocalDir: filepath.Join(taskDir, allocdir.TaskLocal), + SecretsDir: filepath.Join(taskDir, allocdir.TaskSecrets), + } +} + +func (tc *TaskConfig) DecodeDriverConfig(t interface{}) error { + return base.MsgPackDecode(tc.rawDriverConfig, t) +} + +func (tc *TaskConfig) EncodeDriverConfig(t interface{}) error { + return base.MsgPackEncode(&tc.rawDriverConfig, t) +} + +type Resources struct { + CPUPeriod int64 + CPUQuota int64 + CPUShares int64 + MemoryLimitBytes int64 + OOMScoreAdj int64 + CpusetCPUs string + CpusetMems string +} + +type DeviceConfig struct { + TaskPath string + HostPath string + Permissions string +} + +type MountConfig struct { + TaskPath string + HostPath string + Readonly bool +} + +const ( + TaskStateUnknown TaskState = "unknown" + TaskStateRunning TaskState = "running" + TaskStateExited TaskState = "exited" +) + +type TaskState string + +type NetworkOverride struct { + PortMap map[string]int32 + Addr string + AutoAdvertise bool +} + +type ExitResult struct { + ExitCode int + Signal int + OOMKilled bool + Err error +} + +type TaskStatus struct { + ID string + Name string + State TaskState + StartedAt time.Time + CompletedAt time.Time + ExitResult *ExitResult + DriverAttributes map[string]string + NetworkOverride *NetworkOverride +} + +type TaskStats struct { + ID string + Timestamp int64 + AggResourceUsage *cstructs.ResourceUsage + ResourceUsageByPid map[string]*cstructs.ResourceUsage +} + +type TaskEvent struct { + TaskID string + Timestamp time.Time + Message string + Annotations map[string]string + + // Err is only used if an error occured while consuming the RPC stream + Err error +} + +type ExecTaskResult struct { + Stdout []byte + Stderr []byte + ExitResult *ExitResult +} diff --git a/plugins/drivers/errors.go b/plugins/drivers/errors.go new file mode 100644 index 000000000..f67a9819e --- /dev/null +++ b/plugins/drivers/errors.go @@ -0,0 +1,5 @@ +package drivers + +import "fmt" + +var ErrTaskNotFound = fmt.Errorf("task not found for given id") diff --git a/plugins/drivers/plugin.go b/plugins/drivers/plugin.go new file mode 100644 index 000000000..10af11368 --- /dev/null +++ b/plugins/drivers/plugin.go @@ -0,0 +1,41 @@ +package drivers + +import ( + "context" + + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/plugins/drivers/proto" + "google.golang.org/grpc" +) + +// PluginDriver wraps a DriverPlugin and implements go-plugins GRPCPlugin +// interface to expose the the interface over gRPC +type PluginDriver struct { + plugin.NetRPCUnsupportedPlugin + impl DriverPlugin + logger hclog.Logger +} + +func NewDriverPlugin(d DriverPlugin, logger hclog.Logger) plugin.GRPCPlugin { + return &PluginDriver{ + impl: d, + logger: logger.Named("driver_plugin"), + } +} + +func (p *PluginDriver) GRPCServer(broker *plugin.GRPCBroker, s *grpc.Server) error { + proto.RegisterDriverServer(s, &driverPluginServer{ + impl: p.impl, + broker: broker, + logger: p.logger, + }) + return nil +} + +func (p *PluginDriver) GRPCClient(ctx context.Context, broker *plugin.GRPCBroker, c *grpc.ClientConn) (interface{}, error) { + return &driverPluginClient{ + client: proto.NewDriverClient(c), + logger: p.logger, + }, nil +} diff --git a/plugins/drivers/plugin_test.go b/plugins/drivers/plugin_test.go new file mode 100644 index 000000000..c08d14150 --- /dev/null +++ b/plugins/drivers/plugin_test.go @@ -0,0 +1,242 @@ +package drivers + +import ( + "bytes" + "sync" + "testing" + "time" + + "github.com/hashicorp/nomad/nomad/structs" + "github.com/stretchr/testify/require" + "github.com/ugorji/go/codec" + "golang.org/x/net/context" +) + +type testDriverState struct { + Pid int + Log string +} + +func TestBaseDriver_Fingerprint(t *testing.T) { + t.Parallel() + require := require.New(t) + + fingerprints := []*Fingerprint{ + { + Attributes: map[string]string{"foo": "bar"}, + Health: HealthStateUnhealthy, + HealthDescription: "starting up", + }, + { + Attributes: map[string]string{"foo": "bar"}, + Health: HealthStateHealthy, + HealthDescription: "running", + }, + } + + var complete bool + impl := &MockDriver{ + FingerprintF: func(ctx context.Context) (<-chan *Fingerprint, error) { + ch := make(chan *Fingerprint) + go func() { + defer close(ch) + ch <- fingerprints[0] + time.Sleep(500 * time.Millisecond) + ch <- fingerprints[1] + complete = true + }() + return ch, nil + }, + } + + harness := NewDriverHarness(t, impl) + defer harness.Kill() + + ch, err := harness.Fingerprint(context.Background()) + require.NoError(err) + + var wg sync.WaitGroup + wg.Add(1) + go func() { + defer wg.Done() + select { + case f := <-ch: + require.Exactly(f, fingerprints[0]) + case <-time.After(1 * time.Second): + require.Fail("did not receive fingerprint[0]") + } + select { + case f := <-ch: + require.Exactly(f, fingerprints[1]) + case <-time.After(1 * time.Second): + require.Fail("did not receive fingerprint[1]") + } + }() + require.False(complete) + wg.Wait() + require.True(complete) + +} + +func TestBaseDriver_RecoverTask(t *testing.T) { + t.Parallel() + require := require.New(t) + + // build driver state and encode it into proto msg + state := testDriverState{Pid: 1, Log: "foo"} + var buf bytes.Buffer + enc := codec.NewEncoder(&buf, structs.MsgpackHandle) + enc.Encode(state) + + // mock the RecoverTask driver call + impl := &MockDriver{ + RecoverTaskF: func(h *TaskHandle) error { + var actual testDriverState + require.NoError(h.GetDriverState(&actual)) + require.Equal(state, actual) + return nil + }, + } + + harness := NewDriverHarness(t, impl) + defer harness.Kill() + + handle := &TaskHandle{ + driverState: buf.Bytes(), + } + err := harness.RecoverTask(handle) + require.NoError(err) +} + +func TestBaseDriver_StartTask(t *testing.T) { + t.Parallel() + require := require.New(t) + + cfg := &TaskConfig{ + ID: "foo", + } + state := &testDriverState{Pid: 1, Log: "log"} + var handle *TaskHandle + impl := &MockDriver{ + StartTaskF: func(c *TaskConfig) (*TaskHandle, error) { + handle = NewTaskHandle("test") + handle.Config = c + handle.State = TaskStateRunning + handle.SetDriverState(state) + return handle, nil + }, + } + + harness := NewDriverHarness(t, impl) + defer harness.Kill() + resp, err := harness.StartTask(cfg) + require.NoError(err) + require.Equal(cfg.ID, resp.Config.ID) + require.Equal(handle.State, resp.State) + + var actualState testDriverState + require.NoError(resp.GetDriverState(&actualState)) + require.Equal(*state, actualState) + +} + +func TestBaseDriver_WaitTask(t *testing.T) { + t.Parallel() + require := require.New(t) + + result := &ExitResult{ExitCode: 1, Signal: 9} + + signalTask := make(chan struct{}) + + impl := &MockDriver{ + WaitTaskF: func(_ context.Context, id string) (<-chan *ExitResult, error) { + ch := make(chan *ExitResult) + go func() { + <-signalTask + ch <- result + }() + return ch, nil + }, + } + + harness := NewDriverHarness(t, impl) + defer harness.Kill() + var wg sync.WaitGroup + wg.Add(1) + var finished bool + go func() { + defer wg.Done() + ch, err := harness.WaitTask(context.TODO(), "foo") + require.NoError(err) + actualResult := <-ch + finished = true + require.Exactly(result, actualResult) + }() + require.False(finished) + close(signalTask) + wg.Wait() + require.True(finished) +} + +func TestBaseDriver_TaskEvents(t *testing.T) { + t.Parallel() + require := require.New(t) + + now := time.Now().UTC().Truncate(time.Millisecond) + events := []*TaskEvent{ + { + TaskID: "abc", + Timestamp: now, + Annotations: map[string]string{"foo": "bar"}, + Message: "starting", + }, + { + TaskID: "xyz", + Timestamp: now.Add(2 * time.Second), + Annotations: map[string]string{"foo": "bar"}, + Message: "starting", + }, + { + TaskID: "xyz", + Timestamp: now.Add(3 * time.Second), + Annotations: map[string]string{"foo": "bar"}, + Message: "running", + }, + { + TaskID: "abc", + Timestamp: now.Add(4 * time.Second), + Annotations: map[string]string{"foo": "bar"}, + Message: "running", + }, + } + + impl := &MockDriver{ + TaskEventsF: func(ctx context.Context) (<-chan *TaskEvent, error) { + ch := make(chan *TaskEvent) + go func() { + defer close(ch) + for _, event := range events { + ch <- event + } + }() + return ch, nil + }, + } + + harness := NewDriverHarness(t, impl) + defer harness.Kill() + + ch, err := harness.TaskEvents(context.Background()) + require.NoError(err) + + for _, event := range events { + select { + case actual := <-ch: + require.Exactly(actual, event) + case <-time.After(500 * time.Millisecond): + require.Fail("failed to receive event") + + } + } + +} diff --git a/plugins/drivers/base/proto/driver.pb.go b/plugins/drivers/proto/driver.pb.go similarity index 77% rename from plugins/drivers/base/proto/driver.pb.go rename to plugins/drivers/proto/driver.pb.go index e170302c9..0c9542e5e 100644 --- a/plugins/drivers/base/proto/driver.pb.go +++ b/plugins/drivers/proto/driver.pb.go @@ -49,7 +49,7 @@ func (x TaskState) String() string { return proto.EnumName(TaskState_name, int32(x)) } func (TaskState) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{0} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{0} } type FingerprintResponse_HealthState int32 @@ -75,7 +75,7 @@ func (x FingerprintResponse_HealthState) String() string { return proto.EnumName(FingerprintResponse_HealthState_name, int32(x)) } func (FingerprintResponse_HealthState) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{5, 0} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{5, 0} } type StartTaskResponse_Result int32 @@ -101,7 +101,7 @@ func (x StartTaskResponse_Result) String() string { return proto.EnumName(StartTaskResponse_Result_name, int32(x)) } func (StartTaskResponse_Result) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{9, 0} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{9, 0} } type DriverCapabilities_FSIsolation int32 @@ -127,7 +127,7 @@ func (x DriverCapabilities_FSIsolation) String() string { return proto.EnumName(DriverCapabilities_FSIsolation_name, int32(x)) } func (DriverCapabilities_FSIsolation) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{27, 0} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{25, 0} } type CPUUsage_Fields int32 @@ -162,7 +162,7 @@ func (x CPUUsage_Fields) String() string { return proto.EnumName(CPUUsage_Fields_name, int32(x)) } func (CPUUsage_Fields) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{43, 0} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{41, 0} } type MemoryUsage_Fields int32 @@ -170,7 +170,7 @@ type MemoryUsage_Fields int32 const ( MemoryUsage_RSS MemoryUsage_Fields = 0 MemoryUsage_CACHE MemoryUsage_Fields = 1 - MemoryUsage_MAX_UASGE MemoryUsage_Fields = 2 + MemoryUsage_MAX_USAGE MemoryUsage_Fields = 2 MemoryUsage_KERNEL_USAGE MemoryUsage_Fields = 3 MemoryUsage_KERNEL_MAX_USAGE MemoryUsage_Fields = 4 ) @@ -178,14 +178,14 @@ const ( var MemoryUsage_Fields_name = map[int32]string{ 0: "RSS", 1: "CACHE", - 2: "MAX_UASGE", + 2: "MAX_USAGE", 3: "KERNEL_USAGE", 4: "KERNEL_MAX_USAGE", } var MemoryUsage_Fields_value = map[string]int32{ "RSS": 0, "CACHE": 1, - "MAX_UASGE": 2, + "MAX_USAGE": 2, "KERNEL_USAGE": 3, "KERNEL_MAX_USAGE": 4, } @@ -194,7 +194,7 @@ func (x MemoryUsage_Fields) String() string { return proto.EnumName(MemoryUsage_Fields_name, int32(x)) } func (MemoryUsage_Fields) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{44, 0} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{42, 0} } type TaskConfigSchemaRequest struct { @@ -207,7 +207,7 @@ func (m *TaskConfigSchemaRequest) Reset() { *m = TaskConfigSchemaRequest func (m *TaskConfigSchemaRequest) String() string { return proto.CompactTextString(m) } func (*TaskConfigSchemaRequest) ProtoMessage() {} func (*TaskConfigSchemaRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{0} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{0} } func (m *TaskConfigSchemaRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskConfigSchemaRequest.Unmarshal(m, b) @@ -239,7 +239,7 @@ func (m *TaskConfigSchemaResponse) Reset() { *m = TaskConfigSchemaRespon func (m *TaskConfigSchemaResponse) String() string { return proto.CompactTextString(m) } func (*TaskConfigSchemaResponse) ProtoMessage() {} func (*TaskConfigSchemaResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{1} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{1} } func (m *TaskConfigSchemaResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskConfigSchemaResponse.Unmarshal(m, b) @@ -276,7 +276,7 @@ func (m *CapabilitiesRequest) Reset() { *m = CapabilitiesRequest{} } func (m *CapabilitiesRequest) String() string { return proto.CompactTextString(m) } func (*CapabilitiesRequest) ProtoMessage() {} func (*CapabilitiesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{2} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{2} } func (m *CapabilitiesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CapabilitiesRequest.Unmarshal(m, b) @@ -311,7 +311,7 @@ func (m *CapabilitiesResponse) Reset() { *m = CapabilitiesResponse{} } func (m *CapabilitiesResponse) String() string { return proto.CompactTextString(m) } func (*CapabilitiesResponse) ProtoMessage() {} func (*CapabilitiesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{3} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{3} } func (m *CapabilitiesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CapabilitiesResponse.Unmarshal(m, b) @@ -348,7 +348,7 @@ func (m *FingerprintRequest) Reset() { *m = FingerprintRequest{} } func (m *FingerprintRequest) String() string { return proto.CompactTextString(m) } func (*FingerprintRequest) ProtoMessage() {} func (*FingerprintRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{4} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{4} } func (m *FingerprintRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_FingerprintRequest.Unmarshal(m, b) @@ -378,7 +378,7 @@ type FingerprintResponse struct { // * UNHEALTHY: driver dependencies are met but the driver is unable to // perform operations due to some other problem // * HEALTHY: driver is able to perform all operations - Health FingerprintResponse_HealthState `protobuf:"varint,2,opt,name=health,proto3,enum=hashicorp.nomad.plugins.drivers.base.proto.FingerprintResponse_HealthState" json:"health,omitempty"` + Health FingerprintResponse_HealthState `protobuf:"varint,2,opt,name=health,proto3,enum=hashicorp.nomad.plugins.drivers.proto.FingerprintResponse_HealthState" json:"health,omitempty"` // HealthDescription is a human readable message describing the current // state of driver health HealthDescription string `protobuf:"bytes,3,opt,name=health_description,json=healthDescription,proto3" json:"health_description,omitempty"` @@ -391,7 +391,7 @@ func (m *FingerprintResponse) Reset() { *m = FingerprintResponse{} } func (m *FingerprintResponse) String() string { return proto.CompactTextString(m) } func (*FingerprintResponse) ProtoMessage() {} func (*FingerprintResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{5} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{5} } func (m *FingerprintResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_FingerprintResponse.Unmarshal(m, b) @@ -446,7 +446,7 @@ func (m *RecoverTaskRequest) Reset() { *m = RecoverTaskRequest{} } func (m *RecoverTaskRequest) String() string { return proto.CompactTextString(m) } func (*RecoverTaskRequest) ProtoMessage() {} func (*RecoverTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{6} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{6} } func (m *RecoverTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecoverTaskRequest.Unmarshal(m, b) @@ -490,7 +490,7 @@ func (m *RecoverTaskResponse) Reset() { *m = RecoverTaskResponse{} } func (m *RecoverTaskResponse) String() string { return proto.CompactTextString(m) } func (*RecoverTaskResponse) ProtoMessage() {} func (*RecoverTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{7} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{7} } func (m *RecoverTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecoverTaskResponse.Unmarshal(m, b) @@ -522,7 +522,7 @@ func (m *StartTaskRequest) Reset() { *m = StartTaskRequest{} } func (m *StartTaskRequest) String() string { return proto.CompactTextString(m) } func (*StartTaskRequest) ProtoMessage() {} func (*StartTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{8} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{8} } func (m *StartTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StartTaskRequest.Unmarshal(m, b) @@ -558,7 +558,7 @@ type StartTaskResponse struct { // * FATAL: A fatal error occurred and is not likely to succeed if retried // // If Result is not successful, the DriverErrorMsg will be set. - Result StartTaskResponse_Result `protobuf:"varint,1,opt,name=result,proto3,enum=hashicorp.nomad.plugins.drivers.base.proto.StartTaskResponse_Result" json:"result,omitempty"` + Result StartTaskResponse_Result `protobuf:"varint,1,opt,name=result,proto3,enum=hashicorp.nomad.plugins.drivers.proto.StartTaskResponse_Result" json:"result,omitempty"` // DriverErrorMsg is set if an error occurred DriverErrorMsg string `protobuf:"bytes,2,opt,name=driver_error_msg,json=driverErrorMsg,proto3" json:"driver_error_msg,omitempty"` // Handle is opaque to the client, but must be stored in order to recover @@ -576,7 +576,7 @@ func (m *StartTaskResponse) Reset() { *m = StartTaskResponse{} } func (m *StartTaskResponse) String() string { return proto.CompactTextString(m) } func (*StartTaskResponse) ProtoMessage() {} func (*StartTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{9} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{9} } func (m *StartTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StartTaskResponse.Unmarshal(m, b) @@ -636,7 +636,7 @@ func (m *WaitTaskRequest) Reset() { *m = WaitTaskRequest{} } func (m *WaitTaskRequest) String() string { return proto.CompactTextString(m) } func (*WaitTaskRequest) ProtoMessage() {} func (*WaitTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{10} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{10} } func (m *WaitTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WaitTaskRequest.Unmarshal(m, b) @@ -677,7 +677,7 @@ func (m *WaitTaskResponse) Reset() { *m = WaitTaskResponse{} } func (m *WaitTaskResponse) String() string { return proto.CompactTextString(m) } func (*WaitTaskResponse) ProtoMessage() {} func (*WaitTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{11} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{11} } func (m *WaitTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WaitTaskResponse.Unmarshal(m, b) @@ -729,7 +729,7 @@ func (m *StopTaskRequest) Reset() { *m = StopTaskRequest{} } func (m *StopTaskRequest) String() string { return proto.CompactTextString(m) } func (*StopTaskRequest) ProtoMessage() {} func (*StopTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{12} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{12} } func (m *StopTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StopTaskRequest.Unmarshal(m, b) @@ -780,7 +780,7 @@ func (m *StopTaskResponse) Reset() { *m = StopTaskResponse{} } func (m *StopTaskResponse) String() string { return proto.CompactTextString(m) } func (*StopTaskResponse) ProtoMessage() {} func (*StopTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{13} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{13} } func (m *StopTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StopTaskResponse.Unmarshal(m, b) @@ -802,7 +802,9 @@ var xxx_messageInfo_StopTaskResponse proto.InternalMessageInfo type DestroyTaskRequest struct { // TaskId is the ID of the target task - TaskId string `protobuf:"bytes,1,opt,name=task_id,json=taskId,proto3" json:"task_id,omitempty"` + TaskId string `protobuf:"bytes,1,opt,name=task_id,json=taskId,proto3" json:"task_id,omitempty"` + // Force destroys the task even if it is still in a running state + Force bool `protobuf:"varint,2,opt,name=force,proto3" json:"force,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -812,7 +814,7 @@ func (m *DestroyTaskRequest) Reset() { *m = DestroyTaskRequest{} } func (m *DestroyTaskRequest) String() string { return proto.CompactTextString(m) } func (*DestroyTaskRequest) ProtoMessage() {} func (*DestroyTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{14} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{14} } func (m *DestroyTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DestroyTaskRequest.Unmarshal(m, b) @@ -839,6 +841,13 @@ func (m *DestroyTaskRequest) GetTaskId() string { return "" } +func (m *DestroyTaskRequest) GetForce() bool { + if m != nil { + return m.Force + } + return false +} + type DestroyTaskResponse struct { XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` @@ -849,7 +858,7 @@ func (m *DestroyTaskResponse) Reset() { *m = DestroyTaskResponse{} } func (m *DestroyTaskResponse) String() string { return proto.CompactTextString(m) } func (*DestroyTaskResponse) ProtoMessage() {} func (*DestroyTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{15} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{15} } func (m *DestroyTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DestroyTaskResponse.Unmarshal(m, b) @@ -869,75 +878,6 @@ func (m *DestroyTaskResponse) XXX_DiscardUnknown() { var xxx_messageInfo_DestroyTaskResponse proto.InternalMessageInfo -type ListTasksRequest struct { - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *ListTasksRequest) Reset() { *m = ListTasksRequest{} } -func (m *ListTasksRequest) String() string { return proto.CompactTextString(m) } -func (*ListTasksRequest) ProtoMessage() {} -func (*ListTasksRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{16} -} -func (m *ListTasksRequest) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_ListTasksRequest.Unmarshal(m, b) -} -func (m *ListTasksRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_ListTasksRequest.Marshal(b, m, deterministic) -} -func (dst *ListTasksRequest) XXX_Merge(src proto.Message) { - xxx_messageInfo_ListTasksRequest.Merge(dst, src) -} -func (m *ListTasksRequest) XXX_Size() int { - return xxx_messageInfo_ListTasksRequest.Size(m) -} -func (m *ListTasksRequest) XXX_DiscardUnknown() { - xxx_messageInfo_ListTasksRequest.DiscardUnknown(m) -} - -var xxx_messageInfo_ListTasksRequest proto.InternalMessageInfo - -type ListTasksResponse struct { - // Tasks includes a list of summary information for each task - Tasks []*TaskStatus `protobuf:"bytes,1,rep,name=tasks,proto3" json:"tasks,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *ListTasksResponse) Reset() { *m = ListTasksResponse{} } -func (m *ListTasksResponse) String() string { return proto.CompactTextString(m) } -func (*ListTasksResponse) ProtoMessage() {} -func (*ListTasksResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{17} -} -func (m *ListTasksResponse) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_ListTasksResponse.Unmarshal(m, b) -} -func (m *ListTasksResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_ListTasksResponse.Marshal(b, m, deterministic) -} -func (dst *ListTasksResponse) XXX_Merge(src proto.Message) { - xxx_messageInfo_ListTasksResponse.Merge(dst, src) -} -func (m *ListTasksResponse) XXX_Size() int { - return xxx_messageInfo_ListTasksResponse.Size(m) -} -func (m *ListTasksResponse) XXX_DiscardUnknown() { - xxx_messageInfo_ListTasksResponse.DiscardUnknown(m) -} - -var xxx_messageInfo_ListTasksResponse proto.InternalMessageInfo - -func (m *ListTasksResponse) GetTasks() []*TaskStatus { - if m != nil { - return m.Tasks - } - return nil -} - type InspectTaskRequest struct { // TaskId is the ID of the target task TaskId string `protobuf:"bytes,1,opt,name=task_id,json=taskId,proto3" json:"task_id,omitempty"` @@ -950,7 +890,7 @@ func (m *InspectTaskRequest) Reset() { *m = InspectTaskRequest{} } func (m *InspectTaskRequest) String() string { return proto.CompactTextString(m) } func (*InspectTaskRequest) ProtoMessage() {} func (*InspectTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{18} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{16} } func (m *InspectTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectTaskRequest.Unmarshal(m, b) @@ -993,7 +933,7 @@ func (m *InspectTaskResponse) Reset() { *m = InspectTaskResponse{} } func (m *InspectTaskResponse) String() string { return proto.CompactTextString(m) } func (*InspectTaskResponse) ProtoMessage() {} func (*InspectTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{19} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{17} } func (m *InspectTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectTaskResponse.Unmarshal(m, b) @@ -1046,7 +986,7 @@ func (m *TaskStatsRequest) Reset() { *m = TaskStatsRequest{} } func (m *TaskStatsRequest) String() string { return proto.CompactTextString(m) } func (*TaskStatsRequest) ProtoMessage() {} func (*TaskStatsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{20} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{18} } func (m *TaskStatsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStatsRequest.Unmarshal(m, b) @@ -1085,7 +1025,7 @@ func (m *TaskStatsResponse) Reset() { *m = TaskStatsResponse{} } func (m *TaskStatsResponse) String() string { return proto.CompactTextString(m) } func (*TaskStatsResponse) ProtoMessage() {} func (*TaskStatsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{21} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{19} } func (m *TaskStatsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStatsResponse.Unmarshal(m, b) @@ -1122,7 +1062,7 @@ func (m *TaskEventsRequest) Reset() { *m = TaskEventsRequest{} } func (m *TaskEventsRequest) String() string { return proto.CompactTextString(m) } func (*TaskEventsRequest) ProtoMessage() {} func (*TaskEventsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{22} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{20} } func (m *TaskEventsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskEventsRequest.Unmarshal(m, b) @@ -1156,7 +1096,7 @@ func (m *SignalTaskRequest) Reset() { *m = SignalTaskRequest{} } func (m *SignalTaskRequest) String() string { return proto.CompactTextString(m) } func (*SignalTaskRequest) ProtoMessage() {} func (*SignalTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{23} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{21} } func (m *SignalTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SignalTaskRequest.Unmarshal(m, b) @@ -1200,7 +1140,7 @@ func (m *SignalTaskResponse) Reset() { *m = SignalTaskResponse{} } func (m *SignalTaskResponse) String() string { return proto.CompactTextString(m) } func (*SignalTaskResponse) ProtoMessage() {} func (*SignalTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{24} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{22} } func (m *SignalTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SignalTaskResponse.Unmarshal(m, b) @@ -1237,7 +1177,7 @@ func (m *ExecTaskRequest) Reset() { *m = ExecTaskRequest{} } func (m *ExecTaskRequest) String() string { return proto.CompactTextString(m) } func (*ExecTaskRequest) ProtoMessage() {} func (*ExecTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{25} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{23} } func (m *ExecTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExecTaskRequest.Unmarshal(m, b) @@ -1294,7 +1234,7 @@ func (m *ExecTaskResponse) Reset() { *m = ExecTaskResponse{} } func (m *ExecTaskResponse) String() string { return proto.CompactTextString(m) } func (*ExecTaskResponse) ProtoMessage() {} func (*ExecTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{26} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{24} } func (m *ExecTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExecTaskResponse.Unmarshal(m, b) @@ -1343,7 +1283,7 @@ type DriverCapabilities struct { // in the task's execution environment. Exec bool `protobuf:"varint,2,opt,name=exec,proto3" json:"exec,omitempty"` // FsIsolation indicates what kind of filesystem isolation a driver supports. - FsIsolation DriverCapabilities_FSIsolation `protobuf:"varint,3,opt,name=fs_isolation,json=fsIsolation,proto3,enum=hashicorp.nomad.plugins.drivers.base.proto.DriverCapabilities_FSIsolation" json:"fs_isolation,omitempty"` + FsIsolation DriverCapabilities_FSIsolation `protobuf:"varint,3,opt,name=fs_isolation,json=fsIsolation,proto3,enum=hashicorp.nomad.plugins.drivers.proto.DriverCapabilities_FSIsolation" json:"fs_isolation,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1353,7 +1293,7 @@ func (m *DriverCapabilities) Reset() { *m = DriverCapabilities{} } func (m *DriverCapabilities) String() string { return proto.CompactTextString(m) } func (*DriverCapabilities) ProtoMessage() {} func (*DriverCapabilities) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{27} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{25} } func (m *DriverCapabilities) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DriverCapabilities.Unmarshal(m, b) @@ -1424,7 +1364,7 @@ func (m *TaskConfig) Reset() { *m = TaskConfig{} } func (m *TaskConfig) String() string { return proto.CompactTextString(m) } func (*TaskConfig) ProtoMessage() {} func (*TaskConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{28} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{26} } func (m *TaskConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskConfig.Unmarshal(m, b) @@ -1521,7 +1461,7 @@ func (m *Resources) Reset() { *m = Resources{} } func (m *Resources) String() string { return proto.CompactTextString(m) } func (*Resources) ProtoMessage() {} func (*Resources) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{29} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{27} } func (m *Resources) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Resources.Unmarshal(m, b) @@ -1570,7 +1510,7 @@ func (m *RawResources) Reset() { *m = RawResources{} } func (m *RawResources) String() string { return proto.CompactTextString(m) } func (*RawResources) ProtoMessage() {} func (*RawResources) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{30} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{28} } func (m *RawResources) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RawResources.Unmarshal(m, b) @@ -1641,7 +1581,7 @@ func (m *NetworkResource) Reset() { *m = NetworkResource{} } func (m *NetworkResource) String() string { return proto.CompactTextString(m) } func (*NetworkResource) ProtoMessage() {} func (*NetworkResource) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{31} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{29} } func (m *NetworkResource) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NetworkResource.Unmarshal(m, b) @@ -1715,7 +1655,7 @@ func (m *NetworkPort) Reset() { *m = NetworkPort{} } func (m *NetworkPort) String() string { return proto.CompactTextString(m) } func (*NetworkPort) ProtoMessage() {} func (*NetworkPort) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{32} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{30} } func (m *NetworkPort) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NetworkPort.Unmarshal(m, b) @@ -1773,7 +1713,7 @@ func (m *LinuxResources) Reset() { *m = LinuxResources{} } func (m *LinuxResources) String() string { return proto.CompactTextString(m) } func (*LinuxResources) ProtoMessage() {} func (*LinuxResources) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{33} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{31} } func (m *LinuxResources) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LinuxResources.Unmarshal(m, b) @@ -1858,7 +1798,7 @@ func (m *Mount) Reset() { *m = Mount{} } func (m *Mount) String() string { return proto.CompactTextString(m) } func (*Mount) ProtoMessage() {} func (*Mount) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{34} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{32} } func (m *Mount) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mount.Unmarshal(m, b) @@ -1921,7 +1861,7 @@ func (m *Device) Reset() { *m = Device{} } func (m *Device) String() string { return proto.CompactTextString(m) } func (*Device) ProtoMessage() {} func (*Device) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{35} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{33} } func (m *Device) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Device.Unmarshal(m, b) @@ -1967,7 +1907,7 @@ type TaskHandle struct { // Config is the TaskConfig for the task Config *TaskConfig `protobuf:"bytes,1,opt,name=config,proto3" json:"config,omitempty"` // State is the state of the task's execution - State TaskState `protobuf:"varint,2,opt,name=state,proto3,enum=hashicorp.nomad.plugins.drivers.base.proto.TaskState" json:"state,omitempty"` + State TaskState `protobuf:"varint,2,opt,name=state,proto3,enum=hashicorp.nomad.plugins.drivers.proto.TaskState" json:"state,omitempty"` // DriverState is the encoded state for the specific driver DriverState []byte `protobuf:"bytes,3,opt,name=driver_state,json=driverState,proto3" json:"driver_state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1979,7 +1919,7 @@ func (m *TaskHandle) Reset() { *m = TaskHandle{} } func (m *TaskHandle) String() string { return proto.CompactTextString(m) } func (*TaskHandle) ProtoMessage() {} func (*TaskHandle) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{36} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{34} } func (m *TaskHandle) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskHandle.Unmarshal(m, b) @@ -2039,7 +1979,7 @@ func (m *NetworkOverride) Reset() { *m = NetworkOverride{} } func (m *NetworkOverride) String() string { return proto.CompactTextString(m) } func (*NetworkOverride) ProtoMessage() {} func (*NetworkOverride) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{37} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{35} } func (m *NetworkOverride) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NetworkOverride.Unmarshal(m, b) @@ -2097,7 +2037,7 @@ func (m *ExitResult) Reset() { *m = ExitResult{} } func (m *ExitResult) String() string { return proto.CompactTextString(m) } func (*ExitResult) ProtoMessage() {} func (*ExitResult) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{38} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{36} } func (m *ExitResult) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExitResult.Unmarshal(m, b) @@ -2143,17 +2083,14 @@ type TaskStatus struct { Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` // State is the state of the task's execution - State TaskState `protobuf:"varint,3,opt,name=state,proto3,enum=hashicorp.nomad.plugins.drivers.base.proto.TaskState" json:"state,omitempty"` - // SizeOnDiskMb is the disk space the driver reports the task is consuming - // in megabytes. - SizeOnDiskMb int64 `protobuf:"varint,4,opt,name=size_on_disk_mb,json=sizeOnDiskMb,proto3" json:"size_on_disk_mb,omitempty"` + State TaskState `protobuf:"varint,3,opt,name=state,proto3,enum=hashicorp.nomad.plugins.drivers.proto.TaskState" json:"state,omitempty"` // StartedAt is the timestamp when the task was started - StartedAt *timestamp.Timestamp `protobuf:"bytes,5,opt,name=started_at,json=startedAt,proto3" json:"started_at,omitempty"` + StartedAt *timestamp.Timestamp `protobuf:"bytes,4,opt,name=started_at,json=startedAt,proto3" json:"started_at,omitempty"` // CompletedAt is the timestamp when the task exited. // If the task is still running, CompletedAt will not be set - CompletedAt *timestamp.Timestamp `protobuf:"bytes,6,opt,name=completed_at,json=completedAt,proto3" json:"completed_at,omitempty"` + CompletedAt *timestamp.Timestamp `protobuf:"bytes,5,opt,name=completed_at,json=completedAt,proto3" json:"completed_at,omitempty"` // Result is set when CompletedAt is set. - Result *ExitResult `protobuf:"bytes,7,opt,name=result,proto3" json:"result,omitempty"` + Result *ExitResult `protobuf:"bytes,6,opt,name=result,proto3" json:"result,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2163,7 +2100,7 @@ func (m *TaskStatus) Reset() { *m = TaskStatus{} } func (m *TaskStatus) String() string { return proto.CompactTextString(m) } func (*TaskStatus) ProtoMessage() {} func (*TaskStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{39} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{37} } func (m *TaskStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStatus.Unmarshal(m, b) @@ -2204,13 +2141,6 @@ func (m *TaskStatus) GetState() TaskState { return TaskState_UNKNOWN } -func (m *TaskStatus) GetSizeOnDiskMb() int64 { - if m != nil { - return m.SizeOnDiskMb - } - return 0 -} - func (m *TaskStatus) GetStartedAt() *timestamp.Timestamp { if m != nil { return m.StartedAt @@ -2245,7 +2175,7 @@ func (m *TaskDriverStatus) Reset() { *m = TaskDriverStatus{} } func (m *TaskDriverStatus) String() string { return proto.CompactTextString(m) } func (*TaskDriverStatus) ProtoMessage() {} func (*TaskDriverStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{40} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{38} } func (m *TaskDriverStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskDriverStatus.Unmarshal(m, b) @@ -2290,7 +2220,7 @@ func (m *TaskStats) Reset() { *m = TaskStats{} } func (m *TaskStats) String() string { return proto.CompactTextString(m) } func (*TaskStats) ProtoMessage() {} func (*TaskStats) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{41} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{39} } func (m *TaskStats) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStats.Unmarshal(m, b) @@ -2352,7 +2282,7 @@ func (m *TaskResourceUsage) Reset() { *m = TaskResourceUsage{} } func (m *TaskResourceUsage) String() string { return proto.CompactTextString(m) } func (*TaskResourceUsage) ProtoMessage() {} func (*TaskResourceUsage) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{42} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{40} } func (m *TaskResourceUsage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskResourceUsage.Unmarshal(m, b) @@ -2394,7 +2324,7 @@ type CPUUsage struct { ThrottledTime uint64 `protobuf:"varint,5,opt,name=throttled_time,json=throttledTime,proto3" json:"throttled_time,omitempty"` Percent float64 `protobuf:"fixed64,6,opt,name=percent,proto3" json:"percent,omitempty"` // MeasuredFields indicates which fields were actually sampled - MeasuredFields []CPUUsage_Fields `protobuf:"varint,7,rep,packed,name=measured_fields,json=measuredFields,proto3,enum=hashicorp.nomad.plugins.drivers.base.proto.CPUUsage_Fields" json:"measured_fields,omitempty"` + MeasuredFields []CPUUsage_Fields `protobuf:"varint,7,rep,packed,name=measured_fields,json=measuredFields,proto3,enum=hashicorp.nomad.plugins.drivers.proto.CPUUsage_Fields" json:"measured_fields,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2404,7 +2334,7 @@ func (m *CPUUsage) Reset() { *m = CPUUsage{} } func (m *CPUUsage) String() string { return proto.CompactTextString(m) } func (*CPUUsage) ProtoMessage() {} func (*CPUUsage) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{43} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{41} } func (m *CPUUsage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CPUUsage.Unmarshal(m, b) @@ -2480,7 +2410,7 @@ type MemoryUsage struct { KernelUsage uint64 `protobuf:"varint,4,opt,name=kernel_usage,json=kernelUsage,proto3" json:"kernel_usage,omitempty"` KernelMaxUsage uint64 `protobuf:"varint,5,opt,name=kernel_max_usage,json=kernelMaxUsage,proto3" json:"kernel_max_usage,omitempty"` // MeasuredFields indicates which fields were actually sampled - MeasuredFields []MemoryUsage_Fields `protobuf:"varint,6,rep,packed,name=measured_fields,json=measuredFields,proto3,enum=hashicorp.nomad.plugins.drivers.base.proto.MemoryUsage_Fields" json:"measured_fields,omitempty"` + MeasuredFields []MemoryUsage_Fields `protobuf:"varint,6,rep,packed,name=measured_fields,json=measuredFields,proto3,enum=hashicorp.nomad.plugins.drivers.proto.MemoryUsage_Fields" json:"measured_fields,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2490,7 +2420,7 @@ func (m *MemoryUsage) Reset() { *m = MemoryUsage{} } func (m *MemoryUsage) String() string { return proto.CompactTextString(m) } func (*MemoryUsage) ProtoMessage() {} func (*MemoryUsage) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{44} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{42} } func (m *MemoryUsage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MemoryUsage.Unmarshal(m, b) @@ -2570,7 +2500,7 @@ func (m *DriverTaskEvent) Reset() { *m = DriverTaskEvent{} } func (m *DriverTaskEvent) String() string { return proto.CompactTextString(m) } func (*DriverTaskEvent) ProtoMessage() {} func (*DriverTaskEvent) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_4db9b3af49fa7db9, []int{45} + return fileDescriptor_driver_c5fa1f582257ecd9, []int{43} } func (m *DriverTaskEvent) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DriverTaskEvent.Unmarshal(m, b) @@ -2619,64 +2549,62 @@ func (m *DriverTaskEvent) GetAnnotations() map[string]string { } func init() { - proto.RegisterType((*TaskConfigSchemaRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskConfigSchemaRequest") - proto.RegisterType((*TaskConfigSchemaResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskConfigSchemaResponse") - proto.RegisterType((*CapabilitiesRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.CapabilitiesRequest") - proto.RegisterType((*CapabilitiesResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.CapabilitiesResponse") - proto.RegisterType((*FingerprintRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.FingerprintRequest") - proto.RegisterType((*FingerprintResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.FingerprintResponse") - proto.RegisterMapType((map[string]string)(nil), "hashicorp.nomad.plugins.drivers.base.proto.FingerprintResponse.AttributesEntry") - proto.RegisterType((*RecoverTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.RecoverTaskRequest") - proto.RegisterType((*RecoverTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.RecoverTaskResponse") - proto.RegisterType((*StartTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.StartTaskRequest") - proto.RegisterType((*StartTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.StartTaskResponse") - proto.RegisterType((*WaitTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.WaitTaskRequest") - proto.RegisterType((*WaitTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.WaitTaskResponse") - proto.RegisterType((*StopTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.StopTaskRequest") - proto.RegisterType((*StopTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.StopTaskResponse") - proto.RegisterType((*DestroyTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.DestroyTaskRequest") - proto.RegisterType((*DestroyTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.DestroyTaskResponse") - proto.RegisterType((*ListTasksRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.ListTasksRequest") - proto.RegisterType((*ListTasksResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.ListTasksResponse") - proto.RegisterType((*InspectTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.InspectTaskRequest") - proto.RegisterType((*InspectTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.InspectTaskResponse") - proto.RegisterType((*TaskStatsRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskStatsRequest") - proto.RegisterType((*TaskStatsResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskStatsResponse") - proto.RegisterType((*TaskEventsRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskEventsRequest") - proto.RegisterType((*SignalTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.SignalTaskRequest") - proto.RegisterType((*SignalTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.SignalTaskResponse") - proto.RegisterType((*ExecTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.base.proto.ExecTaskRequest") - proto.RegisterType((*ExecTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.base.proto.ExecTaskResponse") - proto.RegisterType((*DriverCapabilities)(nil), "hashicorp.nomad.plugins.drivers.base.proto.DriverCapabilities") - proto.RegisterType((*TaskConfig)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskConfig") - proto.RegisterMapType((map[string]string)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskConfig.EnvEntry") - proto.RegisterType((*Resources)(nil), "hashicorp.nomad.plugins.drivers.base.proto.Resources") - proto.RegisterType((*RawResources)(nil), "hashicorp.nomad.plugins.drivers.base.proto.RawResources") - proto.RegisterType((*NetworkResource)(nil), "hashicorp.nomad.plugins.drivers.base.proto.NetworkResource") - proto.RegisterType((*NetworkPort)(nil), "hashicorp.nomad.plugins.drivers.base.proto.NetworkPort") - proto.RegisterType((*LinuxResources)(nil), "hashicorp.nomad.plugins.drivers.base.proto.LinuxResources") - proto.RegisterType((*Mount)(nil), "hashicorp.nomad.plugins.drivers.base.proto.Mount") - proto.RegisterType((*Device)(nil), "hashicorp.nomad.plugins.drivers.base.proto.Device") - proto.RegisterType((*TaskHandle)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskHandle") - proto.RegisterType((*NetworkOverride)(nil), "hashicorp.nomad.plugins.drivers.base.proto.NetworkOverride") - proto.RegisterMapType((map[string]int32)(nil), "hashicorp.nomad.plugins.drivers.base.proto.NetworkOverride.PortMapEntry") - proto.RegisterType((*ExitResult)(nil), "hashicorp.nomad.plugins.drivers.base.proto.ExitResult") - proto.RegisterType((*TaskStatus)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskStatus") - proto.RegisterType((*TaskDriverStatus)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskDriverStatus") - proto.RegisterMapType((map[string]string)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskDriverStatus.AttributesEntry") - proto.RegisterType((*TaskStats)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskStats") - proto.RegisterMapType((map[string]*TaskResourceUsage)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskStats.ResourceUsageByPidEntry") - proto.RegisterType((*TaskResourceUsage)(nil), "hashicorp.nomad.plugins.drivers.base.proto.TaskResourceUsage") - proto.RegisterType((*CPUUsage)(nil), "hashicorp.nomad.plugins.drivers.base.proto.CPUUsage") - proto.RegisterType((*MemoryUsage)(nil), "hashicorp.nomad.plugins.drivers.base.proto.MemoryUsage") - proto.RegisterType((*DriverTaskEvent)(nil), "hashicorp.nomad.plugins.drivers.base.proto.DriverTaskEvent") - proto.RegisterMapType((map[string]string)(nil), "hashicorp.nomad.plugins.drivers.base.proto.DriverTaskEvent.AnnotationsEntry") - proto.RegisterEnum("hashicorp.nomad.plugins.drivers.base.proto.TaskState", TaskState_name, TaskState_value) - proto.RegisterEnum("hashicorp.nomad.plugins.drivers.base.proto.FingerprintResponse_HealthState", FingerprintResponse_HealthState_name, FingerprintResponse_HealthState_value) - proto.RegisterEnum("hashicorp.nomad.plugins.drivers.base.proto.StartTaskResponse_Result", StartTaskResponse_Result_name, StartTaskResponse_Result_value) - proto.RegisterEnum("hashicorp.nomad.plugins.drivers.base.proto.DriverCapabilities_FSIsolation", DriverCapabilities_FSIsolation_name, DriverCapabilities_FSIsolation_value) - proto.RegisterEnum("hashicorp.nomad.plugins.drivers.base.proto.CPUUsage_Fields", CPUUsage_Fields_name, CPUUsage_Fields_value) - proto.RegisterEnum("hashicorp.nomad.plugins.drivers.base.proto.MemoryUsage_Fields", MemoryUsage_Fields_name, MemoryUsage_Fields_value) + proto.RegisterType((*TaskConfigSchemaRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskConfigSchemaRequest") + proto.RegisterType((*TaskConfigSchemaResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskConfigSchemaResponse") + proto.RegisterType((*CapabilitiesRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.CapabilitiesRequest") + proto.RegisterType((*CapabilitiesResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.CapabilitiesResponse") + proto.RegisterType((*FingerprintRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.FingerprintRequest") + proto.RegisterType((*FingerprintResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.FingerprintResponse") + proto.RegisterMapType((map[string]string)(nil), "hashicorp.nomad.plugins.drivers.proto.FingerprintResponse.AttributesEntry") + proto.RegisterType((*RecoverTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.RecoverTaskRequest") + proto.RegisterType((*RecoverTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.RecoverTaskResponse") + proto.RegisterType((*StartTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.StartTaskRequest") + proto.RegisterType((*StartTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.StartTaskResponse") + proto.RegisterType((*WaitTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.WaitTaskRequest") + proto.RegisterType((*WaitTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.WaitTaskResponse") + proto.RegisterType((*StopTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.StopTaskRequest") + proto.RegisterType((*StopTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.StopTaskResponse") + proto.RegisterType((*DestroyTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.DestroyTaskRequest") + proto.RegisterType((*DestroyTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.DestroyTaskResponse") + proto.RegisterType((*InspectTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.InspectTaskRequest") + proto.RegisterType((*InspectTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.InspectTaskResponse") + proto.RegisterType((*TaskStatsRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskStatsRequest") + proto.RegisterType((*TaskStatsResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskStatsResponse") + proto.RegisterType((*TaskEventsRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskEventsRequest") + proto.RegisterType((*SignalTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.SignalTaskRequest") + proto.RegisterType((*SignalTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.SignalTaskResponse") + proto.RegisterType((*ExecTaskRequest)(nil), "hashicorp.nomad.plugins.drivers.proto.ExecTaskRequest") + proto.RegisterType((*ExecTaskResponse)(nil), "hashicorp.nomad.plugins.drivers.proto.ExecTaskResponse") + proto.RegisterType((*DriverCapabilities)(nil), "hashicorp.nomad.plugins.drivers.proto.DriverCapabilities") + proto.RegisterType((*TaskConfig)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskConfig") + proto.RegisterMapType((map[string]string)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskConfig.EnvEntry") + proto.RegisterType((*Resources)(nil), "hashicorp.nomad.plugins.drivers.proto.Resources") + proto.RegisterType((*RawResources)(nil), "hashicorp.nomad.plugins.drivers.proto.RawResources") + proto.RegisterType((*NetworkResource)(nil), "hashicorp.nomad.plugins.drivers.proto.NetworkResource") + proto.RegisterType((*NetworkPort)(nil), "hashicorp.nomad.plugins.drivers.proto.NetworkPort") + proto.RegisterType((*LinuxResources)(nil), "hashicorp.nomad.plugins.drivers.proto.LinuxResources") + proto.RegisterType((*Mount)(nil), "hashicorp.nomad.plugins.drivers.proto.Mount") + proto.RegisterType((*Device)(nil), "hashicorp.nomad.plugins.drivers.proto.Device") + proto.RegisterType((*TaskHandle)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskHandle") + proto.RegisterType((*NetworkOverride)(nil), "hashicorp.nomad.plugins.drivers.proto.NetworkOverride") + proto.RegisterMapType((map[string]int32)(nil), "hashicorp.nomad.plugins.drivers.proto.NetworkOverride.PortMapEntry") + proto.RegisterType((*ExitResult)(nil), "hashicorp.nomad.plugins.drivers.proto.ExitResult") + proto.RegisterType((*TaskStatus)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskStatus") + proto.RegisterType((*TaskDriverStatus)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskDriverStatus") + proto.RegisterMapType((map[string]string)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskDriverStatus.AttributesEntry") + proto.RegisterType((*TaskStats)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskStats") + proto.RegisterMapType((map[string]*TaskResourceUsage)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskStats.ResourceUsageByPidEntry") + proto.RegisterType((*TaskResourceUsage)(nil), "hashicorp.nomad.plugins.drivers.proto.TaskResourceUsage") + proto.RegisterType((*CPUUsage)(nil), "hashicorp.nomad.plugins.drivers.proto.CPUUsage") + proto.RegisterType((*MemoryUsage)(nil), "hashicorp.nomad.plugins.drivers.proto.MemoryUsage") + proto.RegisterType((*DriverTaskEvent)(nil), "hashicorp.nomad.plugins.drivers.proto.DriverTaskEvent") + proto.RegisterMapType((map[string]string)(nil), "hashicorp.nomad.plugins.drivers.proto.DriverTaskEvent.AnnotationsEntry") + proto.RegisterEnum("hashicorp.nomad.plugins.drivers.proto.TaskState", TaskState_name, TaskState_value) + proto.RegisterEnum("hashicorp.nomad.plugins.drivers.proto.FingerprintResponse_HealthState", FingerprintResponse_HealthState_name, FingerprintResponse_HealthState_value) + proto.RegisterEnum("hashicorp.nomad.plugins.drivers.proto.StartTaskResponse_Result", StartTaskResponse_Result_name, StartTaskResponse_Result_value) + proto.RegisterEnum("hashicorp.nomad.plugins.drivers.proto.DriverCapabilities_FSIsolation", DriverCapabilities_FSIsolation_name, DriverCapabilities_FSIsolation_value) + proto.RegisterEnum("hashicorp.nomad.plugins.drivers.proto.CPUUsage_Fields", CPUUsage_Fields_name, CPUUsage_Fields_value) + proto.RegisterEnum("hashicorp.nomad.plugins.drivers.proto.MemoryUsage_Fields", MemoryUsage_Fields_name, MemoryUsage_Fields_value) } // Reference imports to suppress errors if they are not otherwise used. @@ -2721,11 +2649,8 @@ type DriverClient interface { StopTask(ctx context.Context, in *StopTaskRequest, opts ...grpc.CallOption) (*StopTaskResponse, error) // DestroyTask removes the task from the driver's internal state and cleans // up any additional resources created by the driver. It cannot be called - // on a running task. + // on a running task, unless force is set to true. DestroyTask(ctx context.Context, in *DestroyTaskRequest, opts ...grpc.CallOption) (*DestroyTaskResponse, error) - // ListTasks returns a list of summary information of all the tasks the - // driver is tracking. - ListTasks(ctx context.Context, in *ListTasksRequest, opts ...grpc.CallOption) (*ListTasksResponse, error) // InspectTask returns detailed information for the given task InspectTask(ctx context.Context, in *InspectTaskRequest, opts ...grpc.CallOption) (*InspectTaskResponse, error) // TaskStats collects and returns runtime metrics for the given task @@ -2749,7 +2674,7 @@ func NewDriverClient(cc *grpc.ClientConn) DriverClient { func (c *driverClient) TaskConfigSchema(ctx context.Context, in *TaskConfigSchemaRequest, opts ...grpc.CallOption) (*TaskConfigSchemaResponse, error) { out := new(TaskConfigSchemaResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/TaskConfigSchema", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.proto.Driver/TaskConfigSchema", in, out, opts...) if err != nil { return nil, err } @@ -2758,7 +2683,7 @@ func (c *driverClient) TaskConfigSchema(ctx context.Context, in *TaskConfigSchem func (c *driverClient) Capabilities(ctx context.Context, in *CapabilitiesRequest, opts ...grpc.CallOption) (*CapabilitiesResponse, error) { out := new(CapabilitiesResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/Capabilities", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.proto.Driver/Capabilities", in, out, opts...) if err != nil { return nil, err } @@ -2766,7 +2691,7 @@ func (c *driverClient) Capabilities(ctx context.Context, in *CapabilitiesRequest } func (c *driverClient) Fingerprint(ctx context.Context, in *FingerprintRequest, opts ...grpc.CallOption) (Driver_FingerprintClient, error) { - stream, err := c.cc.NewStream(ctx, &_Driver_serviceDesc.Streams[0], "/hashicorp.nomad.plugins.drivers.base.proto.Driver/Fingerprint", opts...) + stream, err := c.cc.NewStream(ctx, &_Driver_serviceDesc.Streams[0], "/hashicorp.nomad.plugins.drivers.proto.Driver/Fingerprint", opts...) if err != nil { return nil, err } @@ -2799,7 +2724,7 @@ func (x *driverFingerprintClient) Recv() (*FingerprintResponse, error) { func (c *driverClient) RecoverTask(ctx context.Context, in *RecoverTaskRequest, opts ...grpc.CallOption) (*RecoverTaskResponse, error) { out := new(RecoverTaskResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/RecoverTask", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.proto.Driver/RecoverTask", in, out, opts...) if err != nil { return nil, err } @@ -2808,7 +2733,7 @@ func (c *driverClient) RecoverTask(ctx context.Context, in *RecoverTaskRequest, func (c *driverClient) StartTask(ctx context.Context, in *StartTaskRequest, opts ...grpc.CallOption) (*StartTaskResponse, error) { out := new(StartTaskResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/StartTask", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.proto.Driver/StartTask", in, out, opts...) if err != nil { return nil, err } @@ -2817,7 +2742,7 @@ func (c *driverClient) StartTask(ctx context.Context, in *StartTaskRequest, opts func (c *driverClient) WaitTask(ctx context.Context, in *WaitTaskRequest, opts ...grpc.CallOption) (*WaitTaskResponse, error) { out := new(WaitTaskResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/WaitTask", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.proto.Driver/WaitTask", in, out, opts...) if err != nil { return nil, err } @@ -2826,7 +2751,7 @@ func (c *driverClient) WaitTask(ctx context.Context, in *WaitTaskRequest, opts . func (c *driverClient) StopTask(ctx context.Context, in *StopTaskRequest, opts ...grpc.CallOption) (*StopTaskResponse, error) { out := new(StopTaskResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/StopTask", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.proto.Driver/StopTask", in, out, opts...) if err != nil { return nil, err } @@ -2835,16 +2760,7 @@ func (c *driverClient) StopTask(ctx context.Context, in *StopTaskRequest, opts . func (c *driverClient) DestroyTask(ctx context.Context, in *DestroyTaskRequest, opts ...grpc.CallOption) (*DestroyTaskResponse, error) { out := new(DestroyTaskResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/DestroyTask", in, out, opts...) - if err != nil { - return nil, err - } - return out, nil -} - -func (c *driverClient) ListTasks(ctx context.Context, in *ListTasksRequest, opts ...grpc.CallOption) (*ListTasksResponse, error) { - out := new(ListTasksResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/ListTasks", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.proto.Driver/DestroyTask", in, out, opts...) if err != nil { return nil, err } @@ -2853,7 +2769,7 @@ func (c *driverClient) ListTasks(ctx context.Context, in *ListTasksRequest, opts func (c *driverClient) InspectTask(ctx context.Context, in *InspectTaskRequest, opts ...grpc.CallOption) (*InspectTaskResponse, error) { out := new(InspectTaskResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/InspectTask", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.proto.Driver/InspectTask", in, out, opts...) if err != nil { return nil, err } @@ -2862,7 +2778,7 @@ func (c *driverClient) InspectTask(ctx context.Context, in *InspectTaskRequest, func (c *driverClient) TaskStats(ctx context.Context, in *TaskStatsRequest, opts ...grpc.CallOption) (*TaskStatsResponse, error) { out := new(TaskStatsResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/TaskStats", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.proto.Driver/TaskStats", in, out, opts...) if err != nil { return nil, err } @@ -2870,7 +2786,7 @@ func (c *driverClient) TaskStats(ctx context.Context, in *TaskStatsRequest, opts } func (c *driverClient) TaskEvents(ctx context.Context, in *TaskEventsRequest, opts ...grpc.CallOption) (Driver_TaskEventsClient, error) { - stream, err := c.cc.NewStream(ctx, &_Driver_serviceDesc.Streams[1], "/hashicorp.nomad.plugins.drivers.base.proto.Driver/TaskEvents", opts...) + stream, err := c.cc.NewStream(ctx, &_Driver_serviceDesc.Streams[1], "/hashicorp.nomad.plugins.drivers.proto.Driver/TaskEvents", opts...) if err != nil { return nil, err } @@ -2903,7 +2819,7 @@ func (x *driverTaskEventsClient) Recv() (*DriverTaskEvent, error) { func (c *driverClient) SignalTask(ctx context.Context, in *SignalTaskRequest, opts ...grpc.CallOption) (*SignalTaskResponse, error) { out := new(SignalTaskResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/SignalTask", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.proto.Driver/SignalTask", in, out, opts...) if err != nil { return nil, err } @@ -2912,7 +2828,7 @@ func (c *driverClient) SignalTask(ctx context.Context, in *SignalTaskRequest, op func (c *driverClient) ExecTask(ctx context.Context, in *ExecTaskRequest, opts ...grpc.CallOption) (*ExecTaskResponse, error) { out := new(ExecTaskResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.base.proto.Driver/ExecTask", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.plugins.drivers.proto.Driver/ExecTask", in, out, opts...) if err != nil { return nil, err } @@ -2951,11 +2867,8 @@ type DriverServer interface { StopTask(context.Context, *StopTaskRequest) (*StopTaskResponse, error) // DestroyTask removes the task from the driver's internal state and cleans // up any additional resources created by the driver. It cannot be called - // on a running task. + // on a running task, unless force is set to true. DestroyTask(context.Context, *DestroyTaskRequest) (*DestroyTaskResponse, error) - // ListTasks returns a list of summary information of all the tasks the - // driver is tracking. - ListTasks(context.Context, *ListTasksRequest) (*ListTasksResponse, error) // InspectTask returns detailed information for the given task InspectTask(context.Context, *InspectTaskRequest) (*InspectTaskResponse, error) // TaskStats collects and returns runtime metrics for the given task @@ -2983,7 +2896,7 @@ func _Driver_TaskConfigSchema_Handler(srv interface{}, ctx context.Context, dec } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/TaskConfigSchema", + FullMethod: "/hashicorp.nomad.plugins.drivers.proto.Driver/TaskConfigSchema", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { return srv.(DriverServer).TaskConfigSchema(ctx, req.(*TaskConfigSchemaRequest)) @@ -3001,7 +2914,7 @@ func _Driver_Capabilities_Handler(srv interface{}, ctx context.Context, dec func } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/Capabilities", + FullMethod: "/hashicorp.nomad.plugins.drivers.proto.Driver/Capabilities", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { return srv.(DriverServer).Capabilities(ctx, req.(*CapabilitiesRequest)) @@ -3040,7 +2953,7 @@ func _Driver_RecoverTask_Handler(srv interface{}, ctx context.Context, dec func( } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/RecoverTask", + FullMethod: "/hashicorp.nomad.plugins.drivers.proto.Driver/RecoverTask", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { return srv.(DriverServer).RecoverTask(ctx, req.(*RecoverTaskRequest)) @@ -3058,7 +2971,7 @@ func _Driver_StartTask_Handler(srv interface{}, ctx context.Context, dec func(in } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/StartTask", + FullMethod: "/hashicorp.nomad.plugins.drivers.proto.Driver/StartTask", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { return srv.(DriverServer).StartTask(ctx, req.(*StartTaskRequest)) @@ -3076,7 +2989,7 @@ func _Driver_WaitTask_Handler(srv interface{}, ctx context.Context, dec func(int } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/WaitTask", + FullMethod: "/hashicorp.nomad.plugins.drivers.proto.Driver/WaitTask", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { return srv.(DriverServer).WaitTask(ctx, req.(*WaitTaskRequest)) @@ -3094,7 +3007,7 @@ func _Driver_StopTask_Handler(srv interface{}, ctx context.Context, dec func(int } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/StopTask", + FullMethod: "/hashicorp.nomad.plugins.drivers.proto.Driver/StopTask", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { return srv.(DriverServer).StopTask(ctx, req.(*StopTaskRequest)) @@ -3112,7 +3025,7 @@ func _Driver_DestroyTask_Handler(srv interface{}, ctx context.Context, dec func( } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/DestroyTask", + FullMethod: "/hashicorp.nomad.plugins.drivers.proto.Driver/DestroyTask", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { return srv.(DriverServer).DestroyTask(ctx, req.(*DestroyTaskRequest)) @@ -3120,24 +3033,6 @@ func _Driver_DestroyTask_Handler(srv interface{}, ctx context.Context, dec func( return interceptor(ctx, in, info, handler) } -func _Driver_ListTasks_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { - in := new(ListTasksRequest) - if err := dec(in); err != nil { - return nil, err - } - if interceptor == nil { - return srv.(DriverServer).ListTasks(ctx, in) - } - info := &grpc.UnaryServerInfo{ - Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/ListTasks", - } - handler := func(ctx context.Context, req interface{}) (interface{}, error) { - return srv.(DriverServer).ListTasks(ctx, req.(*ListTasksRequest)) - } - return interceptor(ctx, in, info, handler) -} - func _Driver_InspectTask_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { in := new(InspectTaskRequest) if err := dec(in); err != nil { @@ -3148,7 +3043,7 @@ func _Driver_InspectTask_Handler(srv interface{}, ctx context.Context, dec func( } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/InspectTask", + FullMethod: "/hashicorp.nomad.plugins.drivers.proto.Driver/InspectTask", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { return srv.(DriverServer).InspectTask(ctx, req.(*InspectTaskRequest)) @@ -3166,7 +3061,7 @@ func _Driver_TaskStats_Handler(srv interface{}, ctx context.Context, dec func(in } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/TaskStats", + FullMethod: "/hashicorp.nomad.plugins.drivers.proto.Driver/TaskStats", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { return srv.(DriverServer).TaskStats(ctx, req.(*TaskStatsRequest)) @@ -3205,7 +3100,7 @@ func _Driver_SignalTask_Handler(srv interface{}, ctx context.Context, dec func(i } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/SignalTask", + FullMethod: "/hashicorp.nomad.plugins.drivers.proto.Driver/SignalTask", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { return srv.(DriverServer).SignalTask(ctx, req.(*SignalTaskRequest)) @@ -3223,7 +3118,7 @@ func _Driver_ExecTask_Handler(srv interface{}, ctx context.Context, dec func(int } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.plugins.drivers.base.proto.Driver/ExecTask", + FullMethod: "/hashicorp.nomad.plugins.drivers.proto.Driver/ExecTask", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { return srv.(DriverServer).ExecTask(ctx, req.(*ExecTaskRequest)) @@ -3232,7 +3127,7 @@ func _Driver_ExecTask_Handler(srv interface{}, ctx context.Context, dec func(int } var _Driver_serviceDesc = grpc.ServiceDesc{ - ServiceName: "hashicorp.nomad.plugins.drivers.base.proto.Driver", + ServiceName: "hashicorp.nomad.plugins.drivers.proto.Driver", HandlerType: (*DriverServer)(nil), Methods: []grpc.MethodDesc{ { @@ -3263,10 +3158,6 @@ var _Driver_serviceDesc = grpc.ServiceDesc{ MethodName: "DestroyTask", Handler: _Driver_DestroyTask_Handler, }, - { - MethodName: "ListTasks", - Handler: _Driver_ListTasks_Handler, - }, { MethodName: "InspectTask", Handler: _Driver_InspectTask_Handler, @@ -3299,184 +3190,179 @@ var _Driver_serviceDesc = grpc.ServiceDesc{ Metadata: "driver.proto", } -func init() { proto.RegisterFile("driver.proto", fileDescriptor_driver_4db9b3af49fa7db9) } +func init() { proto.RegisterFile("driver.proto", fileDescriptor_driver_c5fa1f582257ecd9) } -var fileDescriptor_driver_4db9b3af49fa7db9 = []byte{ - // 2810 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xb4, 0x59, 0x5f, 0x6f, 0xe3, 0xc6, - 0x11, 0x37, 0xf5, 0xcf, 0xd2, 0x48, 0x96, 0x79, 0x7b, 0x77, 0x8d, 0xa2, 0xa0, 0xcd, 0x85, 0x40, - 0x00, 0x23, 0x69, 0xe4, 0xc6, 0x69, 0xd3, 0xfc, 0xb9, 0xbb, 0x44, 0x91, 0x78, 0x67, 0xe7, 0x2c, - 0xd9, 0x59, 0xc9, 0x48, 0x52, 0x24, 0x61, 0x68, 0x72, 0x4f, 0x62, 0x2c, 0xfe, 0x09, 0x97, 0xf2, - 0xd9, 0x01, 0x0a, 0xe4, 0x21, 0x40, 0x81, 0xa2, 0x2d, 0x8a, 0xb4, 0xef, 0x41, 0xdf, 0x8b, 0xa2, - 0xaf, 0x7d, 0xea, 0x17, 0x08, 0xd0, 0x4f, 0xd0, 0x2f, 0x50, 0xa0, 0x7d, 0xe8, 0x27, 0x28, 0xf6, - 0x0f, 0x29, 0xca, 0xf2, 0x25, 0xa6, 0x74, 0x7d, 0x22, 0x77, 0x76, 0xf9, 0x9b, 0xe1, 0xcc, 0xec, - 0xcc, 0xec, 0x0e, 0xd4, 0xec, 0xd0, 0x39, 0x25, 0x61, 0x2b, 0x08, 0xfd, 0xc8, 0x47, 0x2f, 0x8c, - 0x4d, 0x3a, 0x76, 0x2c, 0x3f, 0x0c, 0x5a, 0x9e, 0xef, 0x9a, 0x76, 0x2b, 0x98, 0x4c, 0x47, 0x8e, - 0x47, 0x5b, 0x62, 0x15, 0x6d, 0x1d, 0x9b, 0x94, 0x88, 0xb5, 0xcd, 0x1f, 0x8d, 0x7c, 0x7f, 0x34, - 0x21, 0xdb, 0x7c, 0x74, 0x3c, 0x7d, 0xb8, 0x6d, 0x4f, 0x43, 0x33, 0x72, 0x7c, 0x4f, 0xce, 0x3f, - 0x7b, 0x71, 0x3e, 0x72, 0x5c, 0x42, 0x23, 0xd3, 0x0d, 0xe4, 0x82, 0xb7, 0x47, 0x4e, 0x34, 0x9e, - 0x1e, 0xb7, 0x2c, 0xdf, 0xdd, 0x4e, 0xf8, 0x6e, 0x73, 0xbe, 0xdb, 0x92, 0xef, 0x36, 0x1d, 0x9b, - 0x21, 0xb1, 0xb7, 0xc7, 0xd6, 0x84, 0x06, 0xc4, 0x62, 0x4f, 0x83, 0xbd, 0x08, 0x04, 0xed, 0x69, - 0x78, 0x6a, 0x68, 0xd2, 0x93, 0x8e, 0xef, 0x3d, 0x74, 0x46, 0x03, 0x6b, 0x4c, 0x5c, 0x13, 0x93, - 0xcf, 0xa7, 0x84, 0x46, 0xda, 0x47, 0xd0, 0x58, 0x9c, 0xa2, 0x81, 0xef, 0x51, 0x82, 0xde, 0x86, - 0x02, 0x03, 0x69, 0x28, 0xb7, 0x94, 0xad, 0xea, 0xce, 0x8f, 0x5b, 0x8f, 0xfb, 0x69, 0xc1, 0xbc, - 0x25, 0x99, 0xb7, 0x06, 0x01, 0xb1, 0x30, 0xff, 0x52, 0xbb, 0x09, 0xd7, 0x3b, 0x66, 0x60, 0x1e, - 0x3b, 0x13, 0x27, 0x72, 0x08, 0x8d, 0x99, 0x7e, 0x01, 0x37, 0xe6, 0xc9, 0x92, 0xe1, 0x31, 0xd4, - 0xac, 0x14, 0x5d, 0x32, 0xbe, 0xdb, 0xba, 0xba, 0xb6, 0x5b, 0x5d, 0x4e, 0x9a, 0x43, 0x9f, 0xc3, - 0xd4, 0x6e, 0x00, 0xba, 0xe7, 0x78, 0x23, 0x12, 0x06, 0xa1, 0xe3, 0x45, 0xb1, 0x44, 0x7f, 0xcc, - 0xc3, 0xf5, 0x39, 0xb2, 0x94, 0xc8, 0x07, 0x30, 0xa3, 0x28, 0x74, 0x8e, 0xa7, 0x11, 0x97, 0x27, - 0xbf, 0x55, 0xdd, 0x39, 0xc8, 0x22, 0xcf, 0x25, 0xa0, 0xad, 0x76, 0x82, 0xa8, 0x7b, 0x51, 0x78, - 0x8e, 0x53, 0x2c, 0x90, 0x05, 0xa5, 0x31, 0x31, 0x27, 0xd1, 0xb8, 0x91, 0xbb, 0xa5, 0x6c, 0xd5, - 0x77, 0x1e, 0xac, 0xca, 0x6c, 0x97, 0xa3, 0x0d, 0x22, 0x33, 0x22, 0x58, 0x42, 0xa3, 0x97, 0x00, - 0x89, 0x37, 0xc3, 0x26, 0xd4, 0x0a, 0x9d, 0x80, 0xb9, 0x63, 0x23, 0x7f, 0x4b, 0xd9, 0xaa, 0xe0, - 0x6b, 0x62, 0xa6, 0x3b, 0x9b, 0x68, 0xde, 0x81, 0xcd, 0x0b, 0x22, 0x23, 0x15, 0xf2, 0x27, 0xe4, - 0x9c, 0x1b, 0xa8, 0x82, 0xd9, 0x2b, 0xba, 0x01, 0xc5, 0x53, 0x73, 0x32, 0x25, 0x5c, 0xee, 0x0a, - 0x16, 0x83, 0x37, 0x72, 0xaf, 0x29, 0xda, 0xeb, 0x50, 0x4d, 0x09, 0x81, 0xea, 0x00, 0x47, 0xfd, - 0xae, 0x3e, 0xd4, 0x3b, 0x43, 0xbd, 0xab, 0xae, 0xa1, 0x0d, 0xa8, 0x1c, 0xf5, 0x77, 0xf5, 0xf6, - 0xfe, 0x70, 0xf7, 0x43, 0x55, 0x41, 0x55, 0x58, 0x8f, 0x07, 0x39, 0xed, 0x97, 0x80, 0x30, 0xb1, - 0xfc, 0x53, 0x12, 0x32, 0x27, 0x95, 0xc6, 0x42, 0x4f, 0xc1, 0x7a, 0x64, 0xd2, 0x13, 0xc3, 0xb1, - 0xa5, 0x00, 0x25, 0x36, 0xdc, 0xb3, 0x51, 0x1f, 0x4a, 0x63, 0xd3, 0xb3, 0x27, 0x42, 0x88, 0xea, - 0xce, 0xab, 0x59, 0x94, 0xc7, 0x38, 0xec, 0xf2, 0xaf, 0xb1, 0x44, 0x61, 0xee, 0x3b, 0xc7, 0x5e, - 0xa8, 0x54, 0xfb, 0x04, 0xd4, 0x41, 0x64, 0x86, 0x51, 0x5a, 0xa6, 0x77, 0xa1, 0xc0, 0x84, 0x90, - 0x2e, 0x9b, 0x99, 0xb1, 0xd8, 0x7f, 0x98, 0x63, 0x68, 0x5f, 0xe6, 0xe1, 0x5a, 0x8a, 0x81, 0x74, - 0xc5, 0x8f, 0xa0, 0x14, 0x12, 0x3a, 0x9d, 0x44, 0x9c, 0x47, 0x7d, 0xa7, 0x9b, 0x85, 0xc7, 0x02, - 0x5c, 0x0b, 0x73, 0x2c, 0x2c, 0x31, 0xd1, 0x16, 0xa8, 0xe2, 0x33, 0x83, 0x84, 0xa1, 0x1f, 0x1a, - 0x2e, 0x1d, 0x49, 0x4b, 0xd6, 0x05, 0x5d, 0x67, 0xe4, 0x1e, 0x1d, 0xa5, 0x94, 0x9c, 0x7f, 0x12, - 0x4a, 0x46, 0x0f, 0x41, 0xf5, 0x48, 0xf4, 0xc8, 0x0f, 0x4f, 0x0c, 0xa6, 0xe9, 0xd0, 0xb1, 0x49, - 0xa3, 0xc0, 0x91, 0xdf, 0xcc, 0x82, 0xdc, 0x17, 0x18, 0x07, 0x12, 0x02, 0x6f, 0x7a, 0xf3, 0x04, - 0xed, 0x45, 0x28, 0x89, 0x7f, 0x66, 0x2e, 0x36, 0x38, 0xea, 0x74, 0xf4, 0xc1, 0x40, 0x5d, 0x43, - 0x15, 0x28, 0x62, 0x7d, 0x88, 0x99, 0xeb, 0x55, 0xa0, 0x78, 0xaf, 0x3d, 0x6c, 0xef, 0xab, 0x39, - 0xed, 0x05, 0xd8, 0x7c, 0xdf, 0x74, 0xa2, 0xab, 0x78, 0x9d, 0x16, 0x81, 0x3a, 0x5b, 0x2b, 0x8d, - 0xd5, 0x9f, 0x33, 0x56, 0x46, 0x25, 0xe9, 0x67, 0x4e, 0x74, 0xc1, 0x3c, 0x2a, 0xe4, 0x49, 0x18, - 0x4a, 0x8b, 0xb0, 0x57, 0xed, 0x11, 0x6c, 0x0e, 0x22, 0x3f, 0xb8, 0xd2, 0xbe, 0x78, 0x05, 0xd6, - 0x59, 0x52, 0xf1, 0xa7, 0x91, 0xdc, 0x18, 0x4f, 0xb7, 0x44, 0xd2, 0x69, 0xc5, 0x49, 0xa7, 0xd5, - 0x95, 0x49, 0x09, 0xc7, 0x2b, 0xd1, 0x0f, 0xa0, 0x44, 0x9d, 0x91, 0x67, 0x4e, 0x64, 0x60, 0x90, - 0x23, 0x0d, 0x31, 0xef, 0x8f, 0x19, 0xcb, 0x1d, 0xf1, 0x12, 0xa0, 0x2e, 0xa1, 0x51, 0xe8, 0x9f, - 0x5f, 0x49, 0x63, 0x37, 0xe1, 0xfa, 0xdc, 0x72, 0x89, 0x82, 0x40, 0xdd, 0x77, 0x28, 0x57, 0x64, - 0x92, 0x2a, 0x4c, 0xb8, 0x96, 0xa2, 0x49, 0xed, 0xee, 0x43, 0x91, 0x21, 0xc5, 0x01, 0x39, 0xb3, - 0x07, 0xb2, 0x40, 0x34, 0xa5, 0x58, 0x80, 0x30, 0xe1, 0xf7, 0x3c, 0x96, 0xae, 0xae, 0x66, 0xee, - 0x6f, 0x72, 0x70, 0x7d, 0x6e, 0xbd, 0x14, 0x6a, 0xc5, 0x08, 0x20, 0x65, 0xe2, 0x18, 0x68, 0x08, - 0x25, 0xb1, 0x4c, 0xda, 0xeb, 0x76, 0x56, 0x34, 0x91, 0x06, 0x25, 0xa6, 0xc4, 0xba, 0x74, 0xa7, - 0xe5, 0xff, 0x2f, 0x3b, 0x4d, 0x8d, 0xff, 0x88, 0x7e, 0xaf, 0x3a, 0x3f, 0x85, 0x6b, 0xa9, 0xc5, - 0x52, 0x97, 0x0f, 0xa0, 0x48, 0x19, 0x41, 0x2a, 0xf3, 0x67, 0xcb, 0x28, 0x93, 0x62, 0x81, 0xa1, - 0x5d, 0x17, 0x1c, 0xf4, 0x53, 0xe2, 0x25, 0xf2, 0x68, 0x5d, 0xb8, 0x36, 0xe0, 0xfe, 0x7c, 0xa5, - 0x0d, 0x34, 0xdb, 0x0b, 0xb9, 0xb9, 0xbd, 0x70, 0x03, 0x50, 0x1a, 0x45, 0xfa, 0xf1, 0x39, 0x6c, - 0xea, 0x67, 0xc4, 0xba, 0x12, 0x72, 0x03, 0xd6, 0x2d, 0xdf, 0x75, 0x4d, 0xcf, 0x6e, 0xe4, 0x6e, - 0xe5, 0xb7, 0x2a, 0x38, 0x1e, 0xa6, 0x37, 0x6d, 0xfe, 0xaa, 0x9b, 0x56, 0xfb, 0x5a, 0x01, 0x75, - 0xc6, 0x5b, 0x6a, 0x93, 0x49, 0x1f, 0xd9, 0x0c, 0x88, 0xf1, 0xae, 0x61, 0x39, 0x92, 0xf4, 0x38, - 0xae, 0x08, 0x3a, 0x09, 0xc3, 0x54, 0xf0, 0xca, 0x3f, 0x89, 0xe0, 0xa5, 0xfd, 0x4b, 0x01, 0xb4, - 0x58, 0x97, 0xa1, 0xe7, 0xa0, 0x46, 0x89, 0x67, 0x1b, 0x42, 0x97, 0xc2, 0xd6, 0x65, 0x5c, 0x65, - 0x34, 0xa1, 0x54, 0x8a, 0x10, 0x14, 0xc8, 0x19, 0xb1, 0xb8, 0x7c, 0x65, 0xcc, 0xdf, 0x91, 0x0b, - 0xb5, 0x87, 0xd4, 0x70, 0xa8, 0x3f, 0x31, 0x93, 0xb2, 0xa5, 0xbe, 0xf3, 0xee, 0x6a, 0x45, 0x62, - 0xeb, 0xde, 0x60, 0x2f, 0x46, 0xc4, 0xd5, 0x87, 0x34, 0x19, 0x68, 0x2d, 0xa8, 0xa6, 0xe6, 0x50, - 0x19, 0x0a, 0xfd, 0x83, 0xbe, 0xae, 0xae, 0x21, 0x80, 0x52, 0x67, 0x17, 0x1f, 0x1c, 0x0c, 0x45, - 0xe6, 0xd8, 0xeb, 0xb5, 0xef, 0xeb, 0x6a, 0x4e, 0xfb, 0x5d, 0x01, 0x60, 0x96, 0xd1, 0x51, 0x1d, - 0x72, 0x89, 0xcd, 0x73, 0x8e, 0xcd, 0xfe, 0xc8, 0x33, 0xdd, 0xb8, 0x4a, 0xe2, 0xef, 0x68, 0x07, - 0x6e, 0xba, 0x74, 0x14, 0x98, 0xd6, 0x89, 0x21, 0x73, 0xb0, 0xc5, 0x3f, 0xe6, 0xbf, 0x56, 0xc3, - 0xd7, 0xe5, 0xa4, 0x94, 0x5a, 0xe0, 0xbe, 0x07, 0x79, 0xe2, 0x9d, 0x36, 0x0a, 0x3c, 0x00, 0xbe, - 0xb5, 0x5c, 0xb9, 0xd1, 0xd2, 0xbd, 0x53, 0x51, 0x81, 0x32, 0x2c, 0x34, 0x80, 0x4a, 0x48, 0xa8, - 0x3f, 0x0d, 0x2d, 0x42, 0x1b, 0xc5, 0xec, 0x1b, 0x0f, 0xc7, 0x1f, 0xe3, 0x19, 0x0e, 0xda, 0x83, - 0x92, 0xeb, 0x4f, 0xbd, 0x88, 0x36, 0x4a, 0x5c, 0xd4, 0x97, 0xb3, 0x20, 0xf6, 0xd8, 0x97, 0x58, - 0x02, 0xa0, 0x7d, 0x58, 0xb7, 0xc9, 0xa9, 0xc3, 0xa4, 0x5b, 0xe7, 0x58, 0x3b, 0x99, 0x6c, 0xce, - 0x3f, 0xc5, 0x31, 0x04, 0x33, 0xc4, 0x94, 0x92, 0xb0, 0x51, 0x16, 0x86, 0x60, 0xef, 0xe8, 0x19, - 0xa8, 0x98, 0x93, 0x89, 0x6f, 0x19, 0xb6, 0x13, 0x36, 0x2a, 0x7c, 0xa2, 0xcc, 0x09, 0x5d, 0x27, - 0x6c, 0xbe, 0x0a, 0xe5, 0x58, 0x5f, 0x99, 0xca, 0xdf, 0x7f, 0x28, 0x50, 0x49, 0x54, 0x83, 0x3e, - 0x86, 0x8d, 0xd0, 0x7c, 0x64, 0xcc, 0x14, 0x2d, 0x22, 0xdc, 0x6b, 0x99, 0x14, 0x6d, 0x3e, 0x9a, - 0xe9, 0xba, 0x16, 0xa6, 0x46, 0xc8, 0x82, 0xcd, 0x89, 0xe3, 0x4d, 0xcf, 0x52, 0x0c, 0x44, 0x06, - 0x79, 0x23, 0x0b, 0x83, 0x7d, 0x06, 0x31, 0x63, 0x51, 0x9f, 0xcc, 0x8d, 0xb5, 0xbf, 0x29, 0x50, - 0x4b, 0xcb, 0xc0, 0xd4, 0x61, 0x05, 0x53, 0xfe, 0x2b, 0x79, 0xcc, 0x5e, 0x59, 0x68, 0x71, 0x89, - 0xeb, 0x87, 0xe7, 0x9c, 0x7d, 0x1e, 0xcb, 0x11, 0xd3, 0xba, 0xed, 0xd0, 0x13, 0xee, 0xd9, 0x79, - 0xcc, 0xdf, 0x19, 0xcd, 0xf1, 0x03, 0xca, 0x8b, 0xbe, 0x3c, 0xe6, 0xef, 0xe8, 0x7d, 0x28, 0xcb, - 0xac, 0xc2, 0x5c, 0x31, 0xbf, 0x64, 0x8a, 0x8a, 0x25, 0xc4, 0x09, 0x98, 0xf6, 0xa7, 0x1c, 0x6c, - 0x5e, 0x98, 0x65, 0xc2, 0x0a, 0xaf, 0x88, 0x63, 0xb3, 0x18, 0x31, 0xc1, 0x2c, 0xc7, 0x8e, 0xab, - 0x2e, 0xfe, 0xce, 0xf7, 0x73, 0x20, 0x2b, 0xa2, 0x9c, 0x13, 0x30, 0xbb, 0xbb, 0xc7, 0x4e, 0x24, - 0xa4, 0x2f, 0x62, 0x31, 0x40, 0x9f, 0x40, 0x3d, 0x24, 0x94, 0x84, 0xa7, 0xc4, 0x36, 0x02, 0x3f, - 0x8c, 0xe2, 0x9f, 0xf8, 0xf9, 0x12, 0x3f, 0x71, 0xe8, 0x87, 0x11, 0xde, 0x88, 0xe1, 0xd8, 0x88, - 0xa2, 0x8f, 0x60, 0xc3, 0x3e, 0xf7, 0x4c, 0xd7, 0xb1, 0x24, 0x7c, 0x69, 0x35, 0xf8, 0x9a, 0x44, - 0xe3, 0xe8, 0xec, 0xc0, 0x96, 0x9a, 0x64, 0xbf, 0x38, 0x31, 0x8f, 0xc9, 0x44, 0x6a, 0x47, 0x0c, - 0xe6, 0x1d, 0xbe, 0x28, 0x1d, 0x5e, 0xfb, 0x2a, 0x07, 0xf5, 0x79, 0xef, 0x41, 0x3f, 0x04, 0xb0, - 0x82, 0xa9, 0x11, 0x90, 0xd0, 0xf1, 0x6d, 0xe9, 0x23, 0x15, 0x2b, 0x98, 0x1e, 0x72, 0x02, 0xdb, - 0x73, 0x6c, 0xfa, 0xf3, 0xa9, 0x1f, 0x99, 0xd2, 0x59, 0xca, 0x56, 0x30, 0x7d, 0x8f, 0x8d, 0xe3, - 0x6f, 0xf9, 0x05, 0x03, 0x95, 0x4e, 0xc3, 0x96, 0x0f, 0x38, 0x01, 0xbd, 0x0c, 0x37, 0x85, 0x5f, - 0x19, 0x13, 0xc7, 0x75, 0x22, 0xc3, 0xf1, 0x8c, 0xe3, 0x73, 0x76, 0x50, 0x17, 0xae, 0x84, 0xc4, - 0xe4, 0x3e, 0x9b, 0xdb, 0xf3, 0xde, 0x61, 0x33, 0x48, 0x83, 0x0d, 0xdf, 0x77, 0x0d, 0x6a, 0xf9, - 0x21, 0x31, 0x4c, 0xfb, 0x33, 0x1e, 0xe8, 0xf2, 0xb8, 0xea, 0xfb, 0xee, 0x80, 0xd1, 0xda, 0xf6, - 0x67, 0xe8, 0x59, 0xa8, 0x5a, 0xc1, 0x94, 0x92, 0xc8, 0x60, 0x8f, 0x46, 0x89, 0xff, 0x36, 0x08, - 0x52, 0x27, 0x98, 0xd2, 0xd4, 0x02, 0x97, 0xb8, 0x2c, 0x1a, 0xa5, 0x16, 0xf4, 0x88, 0x4b, 0xb5, - 0x8f, 0xa1, 0xc8, 0x63, 0x17, 0xfb, 0x3b, 0x9e, 0xf7, 0x03, 0x33, 0x1a, 0x4b, 0xfd, 0x95, 0x19, - 0xe1, 0xd0, 0x8c, 0xc6, 0x6c, 0x72, 0xec, 0xd3, 0x48, 0x4c, 0x0a, 0x27, 0x2b, 0x33, 0x02, 0x9f, - 0x6c, 0x42, 0x39, 0x24, 0xa6, 0xed, 0x7b, 0x93, 0x73, 0xfe, 0xe3, 0x65, 0x9c, 0x8c, 0xb5, 0xcf, - 0xa1, 0x24, 0xc2, 0xd9, 0x0a, 0xf8, 0x2f, 0x01, 0xb2, 0x46, 0xa1, 0x3f, 0x0d, 0x98, 0x65, 0x5c, - 0x87, 0x52, 0xc7, 0xf7, 0x68, 0x7c, 0x07, 0x20, 0x66, 0x0e, 0x67, 0x13, 0xda, 0xb7, 0x8a, 0x48, - 0x6b, 0xe2, 0xf0, 0xc6, 0x4a, 0x04, 0x99, 0xa3, 0x56, 0x3b, 0xf0, 0x4a, 0x94, 0xb8, 0xe0, 0x23, - 0xf2, 0xd6, 0x63, 0xa9, 0x82, 0x8f, 0x88, 0x82, 0x8f, 0xb0, 0xc2, 0x42, 0xe6, 0x51, 0x81, 0x29, - 0xd2, 0x68, 0xd5, 0x4e, 0x6a, 0x62, 0xa2, 0xfd, 0x5b, 0x49, 0xc2, 0x40, 0x5c, 0xb6, 0x22, 0x0b, - 0xca, 0x6c, 0x33, 0x19, 0xae, 0x19, 0xc8, 0x83, 0xc5, 0xee, 0x0a, 0x65, 0x71, 0x8b, 0xed, 0x9d, - 0x9e, 0x19, 0x88, 0x04, 0xbb, 0x1e, 0x88, 0x11, 0x8b, 0x29, 0xa6, 0x3d, 0x8b, 0x29, 0xec, 0x1d, - 0x3d, 0x0f, 0x75, 0x73, 0x1a, 0xf9, 0x86, 0x69, 0x9f, 0x92, 0x30, 0x72, 0x28, 0x91, 0x06, 0xdf, - 0x60, 0xd4, 0x76, 0x4c, 0x6c, 0xbe, 0x01, 0xb5, 0x34, 0xe6, 0xf7, 0x25, 0xa1, 0x62, 0x3a, 0x09, - 0x7d, 0x0a, 0x30, 0x2b, 0xcc, 0x98, 0x63, 0x90, 0x33, 0x27, 0x32, 0x2c, 0xdf, 0x16, 0x31, 0xaf, - 0x88, 0xcb, 0x8c, 0xd0, 0xf1, 0x6d, 0x72, 0xa1, 0xd6, 0x2d, 0xc6, 0xb5, 0x2e, 0xdb, 0x8b, 0x6c, - 0xe7, 0x9c, 0x38, 0x93, 0x09, 0xb1, 0xa5, 0x84, 0x15, 0xdf, 0x77, 0x1f, 0x70, 0x82, 0xf6, 0x9f, - 0x9c, 0x70, 0x10, 0x71, 0xe6, 0xb8, 0x52, 0xdd, 0x93, 0x18, 0x3d, 0xff, 0x04, 0x8c, 0xfe, 0x3c, - 0x6c, 0x52, 0xe7, 0x0b, 0x62, 0xf8, 0x9e, 0xc1, 0xb2, 0x8a, 0xe1, 0x1e, 0xcb, 0x28, 0x50, 0x63, - 0xe4, 0x03, 0xaf, 0xeb, 0xd0, 0x93, 0xde, 0x31, 0x7a, 0x1d, 0x80, 0x46, 0x66, 0x18, 0x11, 0xdb, - 0x30, 0x23, 0x59, 0xe5, 0x34, 0x17, 0x0a, 0xeb, 0x61, 0x7c, 0x05, 0x8b, 0x2b, 0x72, 0x75, 0x3b, - 0x42, 0x77, 0xa0, 0x66, 0xf9, 0x6e, 0x30, 0x21, 0xf2, 0xe3, 0xd2, 0xf7, 0x7e, 0x5c, 0x4d, 0xd6, - 0xb7, 0xa3, 0x54, 0x55, 0xbd, 0xfe, 0x44, 0xaa, 0xea, 0xbf, 0x2b, 0xe2, 0x98, 0x95, 0x3e, 0xea, - 0xa1, 0xc9, 0x25, 0xf7, 0x95, 0xfb, 0xab, 0x1c, 0x1e, 0xbf, 0xeb, 0xb2, 0x72, 0xd5, 0x8b, 0xc1, - 0x6f, 0xf3, 0x50, 0x49, 0x4e, 0x6b, 0x0b, 0x1e, 0xf3, 0x1a, 0x54, 0x92, 0x9b, 0x70, 0x59, 0xc4, - 0x7c, 0xa7, 0xa1, 0x92, 0xc5, 0xe8, 0x04, 0x90, 0x39, 0x1a, 0x25, 0x25, 0x90, 0x31, 0xa5, 0xe6, - 0x28, 0x3e, 0xe9, 0xde, 0xc9, 0xaa, 0x8c, 0x38, 0x91, 0x1d, 0x31, 0x10, 0xac, 0x9a, 0xa3, 0xd1, - 0x1c, 0x05, 0x7d, 0xa9, 0xc0, 0xcd, 0x79, 0x4e, 0xc6, 0xf1, 0xb9, 0x11, 0x38, 0xb6, 0xac, 0xcd, - 0x7b, 0x4b, 0x9d, 0x5d, 0x5b, 0x73, 0x4c, 0xde, 0x39, 0x3f, 0x74, 0x6c, 0xa1, 0x7e, 0x14, 0x2e, - 0x4c, 0x34, 0xbf, 0x52, 0xe0, 0xa9, 0xc7, 0xac, 0xbf, 0xc4, 0x1e, 0x83, 0xb4, 0x3d, 0x56, 0x56, - 0x48, 0xca, 0x9c, 0x7f, 0x56, 0xc4, 0x41, 0x7b, 0x5e, 0x3f, 0xf7, 0x66, 0xb5, 0x61, 0x75, 0xe7, - 0xa7, 0x59, 0x98, 0x75, 0x0e, 0x8f, 0x04, 0x0f, 0x5e, 0x51, 0x1e, 0xcc, 0x55, 0x94, 0x19, 0x6b, - 0x9d, 0x1e, 0xff, 0x52, 0xa0, 0x49, 0x18, 0xed, 0xaf, 0x79, 0x28, 0xc7, 0x2c, 0x58, 0x46, 0xa7, - 0xe7, 0x34, 0x22, 0xae, 0xe1, 0xc6, 0x31, 0x51, 0xc1, 0x20, 0x48, 0x3d, 0x16, 0x15, 0x9f, 0x81, - 0x0a, 0x3b, 0x22, 0x88, 0xe9, 0x1c, 0x9f, 0x2e, 0x33, 0x02, 0x9f, 0x7c, 0x16, 0xaa, 0x91, 0x1f, - 0x99, 0x13, 0x23, 0x72, 0xac, 0x13, 0x91, 0x44, 0x15, 0x0c, 0x9c, 0x34, 0x64, 0x14, 0xf4, 0x22, - 0x5c, 0x8b, 0xc6, 0xa1, 0x1f, 0x45, 0x13, 0x56, 0x10, 0xf2, 0xc2, 0x47, 0x14, 0x29, 0x05, 0xac, - 0x26, 0x13, 0xa2, 0x20, 0xa2, 0x2c, 0x1d, 0xcc, 0x16, 0x33, 0xaf, 0xe6, 0x61, 0xaa, 0x80, 0x37, - 0x12, 0x2a, 0xf3, 0x7a, 0xd4, 0x80, 0xf5, 0x80, 0x84, 0x16, 0xf1, 0x44, 0x24, 0x52, 0x70, 0x3c, - 0x44, 0x36, 0x6c, 0xba, 0xc4, 0xa4, 0xd3, 0x90, 0xd8, 0xc6, 0x43, 0x87, 0x4c, 0x6c, 0x71, 0x60, - 0xaa, 0x67, 0xab, 0xa1, 0x63, 0xdd, 0xb4, 0xee, 0x71, 0x08, 0x5c, 0x8f, 0x31, 0xc5, 0x98, 0x15, - 0x21, 0xe2, 0x0d, 0x6d, 0x42, 0x75, 0xf0, 0xe1, 0x60, 0xa8, 0xf7, 0x8c, 0xde, 0x41, 0x57, 0x97, - 0x57, 0xfa, 0x03, 0x1d, 0x8b, 0xa1, 0xc2, 0xe6, 0x87, 0x07, 0xc3, 0xf6, 0xbe, 0x31, 0xdc, 0xeb, - 0x3c, 0x18, 0xa8, 0x39, 0x74, 0x13, 0xae, 0x0d, 0x77, 0xf1, 0xc1, 0x70, 0xb8, 0xaf, 0x77, 0x8d, - 0x43, 0x1d, 0xef, 0x1d, 0x74, 0x07, 0x6a, 0x1e, 0x21, 0xa8, 0xcf, 0xc8, 0xc3, 0xbd, 0x9e, 0xae, - 0x16, 0x50, 0x15, 0xd6, 0x0f, 0x75, 0xdc, 0xd1, 0xfb, 0x43, 0xb5, 0xa8, 0xfd, 0x33, 0x07, 0xd5, - 0x94, 0x29, 0x99, 0x73, 0x87, 0x54, 0x1c, 0xa1, 0x0a, 0x98, 0xbd, 0xb2, 0x60, 0x63, 0x99, 0xd6, - 0x58, 0x98, 0xa8, 0x80, 0xc5, 0x80, 0x19, 0xcf, 0x35, 0xcf, 0x52, 0x71, 0xa0, 0x80, 0xcb, 0xae, - 0x79, 0x26, 0x40, 0x9e, 0x83, 0xda, 0x09, 0x09, 0x3d, 0x32, 0x91, 0xf3, 0xc2, 0x2c, 0x55, 0x41, - 0x13, 0x4b, 0xb6, 0x40, 0x95, 0x4b, 0x66, 0x30, 0xc2, 0x26, 0x75, 0x41, 0xef, 0xc5, 0x60, 0xa3, - 0x45, 0xd5, 0x97, 0xb8, 0xea, 0xef, 0x2e, 0xe9, 0xae, 0x8f, 0xd3, 0xfe, 0x20, 0xd1, 0xfe, 0x3a, - 0xe4, 0x71, 0x7c, 0x93, 0xdd, 0x69, 0x77, 0x76, 0x99, 0xc6, 0x37, 0xa0, 0xd2, 0x6b, 0x7f, 0x60, - 0x1c, 0xb5, 0x07, 0xf7, 0x75, 0x35, 0x87, 0x54, 0xa8, 0x3d, 0xd0, 0x71, 0x5f, 0xdf, 0x37, 0x8e, - 0x06, 0xed, 0xfb, 0xba, 0x9a, 0x47, 0x37, 0x40, 0x95, 0x14, 0xbe, 0x8e, 0x53, 0x0b, 0xda, 0x5f, - 0x72, 0xb0, 0x29, 0x82, 0x7f, 0x72, 0x61, 0xf6, 0xf8, 0x9b, 0xab, 0xe5, 0xe3, 0x73, 0x03, 0xd6, - 0x5d, 0x42, 0x13, 0x63, 0x54, 0x70, 0x3c, 0x44, 0x1e, 0x54, 0x4d, 0xcf, 0xf3, 0x23, 0x7e, 0xd7, - 0x42, 0x65, 0x04, 0xdd, 0xcf, 0x7e, 0xb5, 0x93, 0x88, 0xdf, 0x6a, 0xcf, 0xe0, 0x44, 0x00, 0x4d, - 0x33, 0x68, 0xde, 0x05, 0xf5, 0xe2, 0x82, 0x2c, 0x19, 0xec, 0x85, 0x97, 0x67, 0x09, 0x8c, 0x30, - 0x57, 0x3d, 0xea, 0x3f, 0xe8, 0x1f, 0xbc, 0xdf, 0x57, 0xd7, 0xd8, 0x00, 0x1f, 0xf5, 0xfb, 0x7b, - 0xfd, 0xfb, 0xaa, 0x82, 0x00, 0x4a, 0xfa, 0x07, 0x7b, 0x43, 0xbd, 0xab, 0xe6, 0x76, 0xfe, 0xbb, - 0x09, 0x25, 0x21, 0x24, 0xfa, 0x46, 0x66, 0xf0, 0x74, 0xf3, 0x15, 0x75, 0x96, 0xab, 0xa4, 0xe7, - 0xba, 0xba, 0xcd, 0xee, 0x6a, 0x20, 0xf2, 0x1e, 0x73, 0x0d, 0xfd, 0x41, 0x81, 0xda, 0xdc, 0x9d, - 0x5d, 0xa6, 0x9b, 0xa6, 0x4b, 0x5a, 0xbf, 0xcd, 0xb7, 0x97, 0x07, 0x48, 0xa4, 0xfa, 0x5a, 0x81, - 0x6a, 0xaa, 0xd5, 0x89, 0xee, 0x2e, 0xdd, 0x23, 0x15, 0x32, 0xbd, 0xb5, 0x62, 0x8f, 0x55, 0x5b, - 0xfb, 0x89, 0x82, 0x7e, 0xaf, 0x40, 0x35, 0xd5, 0x2c, 0xcc, 0x26, 0xd4, 0x62, 0x93, 0x33, 0x9b, - 0x50, 0x97, 0x75, 0x29, 0xd7, 0xd0, 0xaf, 0x15, 0xa8, 0x24, 0x8d, 0x3f, 0x74, 0x7b, 0xc9, 0x7e, - 0xa1, 0x10, 0xe7, 0xce, 0x4a, 0xdd, 0x46, 0x6d, 0x0d, 0xfd, 0x4a, 0x81, 0x72, 0xdc, 0x26, 0x43, - 0x99, 0x12, 0xd1, 0x85, 0x46, 0x5c, 0xf3, 0xf6, 0x72, 0x1f, 0xcf, 0x49, 0x12, 0x77, 0xb0, 0xb2, - 0x49, 0x72, 0xa1, 0xe1, 0xd6, 0xbc, 0xbd, 0xdc, 0xc7, 0x89, 0x24, 0xcc, 0x67, 0x52, 0x8d, 0xb0, - 0x6c, 0x3e, 0xb3, 0xd8, 0x70, 0xcb, 0xe6, 0x33, 0x97, 0x75, 0xe0, 0x84, 0xcf, 0x24, 0x0d, 0xb7, - 0x6c, 0x3e, 0x73, 0xb1, 0x77, 0x97, 0xcd, 0x67, 0x16, 0xba, 0x7c, 0x52, 0x3f, 0xa9, 0x56, 0x5b, - 0x36, 0xfd, 0x2c, 0xf6, 0xf4, 0xb2, 0xe9, 0xe7, 0x92, 0x1e, 0x9f, 0xd4, 0xcf, 0xec, 0xcc, 0x72, - 0x7b, 0xb9, 0xc6, 0xd4, 0x32, 0xfa, 0x59, 0x68, 0x92, 0x69, 0x6b, 0xe8, 0x37, 0xf2, 0x52, 0x46, - 0xb4, 0xb6, 0x50, 0x66, 0xbc, 0xb9, 0x96, 0x58, 0xf3, 0xcd, 0x15, 0xf2, 0x2c, 0x0f, 0x81, 0xbf, - 0x55, 0x00, 0x66, 0xed, 0xb0, 0x6c, 0xe2, 0x2c, 0x34, 0xe3, 0x9a, 0x77, 0x97, 0xfd, 0x7c, 0x6e, - 0xa3, 0xc7, 0xcd, 0xb0, 0x6c, 0x1b, 0xfd, 0x42, 0xfb, 0x2e, 0xdb, 0x46, 0xbf, 0xd8, 0x7f, 0xd3, - 0xd6, 0xde, 0x59, 0xff, 0x45, 0x51, 0x94, 0x44, 0x25, 0xfe, 0x78, 0xe5, 0x7f, 0x01, 0x00, 0x00, - 0xff, 0xff, 0xb0, 0x0e, 0x99, 0xe1, 0x4f, 0x26, 0x00, 0x00, +var fileDescriptor_driver_c5fa1f582257ecd9 = []byte{ + // 2732 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x59, 0xcb, 0x6f, 0x23, 0xc7, + 0xd1, 0x17, 0x49, 0x91, 0x22, 0x8b, 0x14, 0x35, 0xdb, 0xbb, 0xfb, 0x99, 0xa6, 0xf1, 0x7d, 0x5e, + 0x0f, 0x60, 0x40, 0xb0, 0xbd, 0x94, 0x2d, 0xe3, 0xf3, 0x3e, 0x02, 0x3f, 0x68, 0x6a, 0x76, 0x25, + 0xaf, 0x44, 0x29, 0x4d, 0x0a, 0xeb, 0x4d, 0x62, 0x4f, 0x46, 0x33, 0x2d, 0x72, 0x2c, 0xce, 0xc3, + 0xd3, 0x3d, 0x5a, 0x09, 0x41, 0x90, 0x20, 0x09, 0x82, 0x24, 0x40, 0x80, 0x5c, 0x82, 0xdc, 0x73, + 0x0b, 0x72, 0xcd, 0x29, 0x41, 0x2e, 0x01, 0xf2, 0x3f, 0xe4, 0x98, 0x4b, 0x80, 0x5c, 0x73, 0xcd, + 0x29, 0xe8, 0xc7, 0x0c, 0x87, 0x92, 0xd6, 0x1e, 0x52, 0x39, 0x4d, 0x77, 0x55, 0xf7, 0xaf, 0x6b, + 0xaa, 0xaa, 0xbb, 0xaa, 0xbb, 0xa0, 0xe1, 0x44, 0xee, 0x29, 0x89, 0x3a, 0x61, 0x14, 0xb0, 0x00, + 0xbd, 0x3e, 0xb6, 0xe8, 0xd8, 0xb5, 0x83, 0x28, 0xec, 0xf8, 0x81, 0x67, 0x39, 0x9d, 0x70, 0x12, + 0x8f, 0x5c, 0x9f, 0x76, 0xe4, 0x28, 0x2a, 0x87, 0xb5, 0xff, 0x6f, 0x14, 0x04, 0xa3, 0x09, 0xd9, + 0x10, 0xbd, 0xa3, 0xf8, 0x78, 0xc3, 0x89, 0x23, 0x8b, 0xb9, 0x81, 0xaf, 0xf8, 0xaf, 0x5e, 0xe4, + 0x33, 0xd7, 0x23, 0x94, 0x59, 0x5e, 0xa8, 0x06, 0x7c, 0x34, 0x72, 0xd9, 0x38, 0x3e, 0xea, 0xd8, + 0x81, 0xb7, 0x91, 0x2e, 0xb9, 0x21, 0x96, 0xdc, 0x50, 0x4b, 0x6e, 0xd0, 0xb1, 0x15, 0x11, 0x67, + 0x63, 0x6c, 0x4f, 0x68, 0x48, 0x6c, 0xfe, 0x35, 0x79, 0x43, 0x22, 0xe8, 0x2f, 0xc3, 0x4b, 0x43, + 0x8b, 0x9e, 0xf4, 0x02, 0xff, 0xd8, 0x1d, 0x0d, 0xec, 0x31, 0xf1, 0x2c, 0x4c, 0xbe, 0x8c, 0x09, + 0x65, 0xfa, 0x77, 0xa0, 0x75, 0x99, 0x45, 0xc3, 0xc0, 0xa7, 0x04, 0x7d, 0x04, 0xcb, 0x1c, 0xa4, + 0x55, 0xb8, 0x53, 0x58, 0xaf, 0x6f, 0xbe, 0xd5, 0x79, 0xd1, 0xff, 0xca, 0xc5, 0x3b, 0x6a, 0xf1, + 0xce, 0x20, 0x24, 0x36, 0x16, 0x33, 0xf5, 0xdb, 0x70, 0xb3, 0x67, 0x85, 0xd6, 0x91, 0x3b, 0x71, + 0x99, 0x4b, 0x68, 0xb2, 0x68, 0x0c, 0xb7, 0x66, 0xc9, 0x6a, 0xc1, 0xcf, 0xa0, 0x61, 0x67, 0xe8, + 0x6a, 0xe1, 0x07, 0x9d, 0x5c, 0x8a, 0xee, 0x6c, 0x89, 0xde, 0x0c, 0xf0, 0x0c, 0x9c, 0x7e, 0x0b, + 0xd0, 0x23, 0xd7, 0x1f, 0x91, 0x28, 0x8c, 0x5c, 0x9f, 0x25, 0xc2, 0xfc, 0xa2, 0x04, 0x37, 0x67, + 0xc8, 0x4a, 0x98, 0x2f, 0x00, 0x2c, 0xc6, 0x22, 0xf7, 0x28, 0x66, 0x42, 0x94, 0xd2, 0x7a, 0x7d, + 0xf3, 0x93, 0x9c, 0xa2, 0x5c, 0x81, 0xd7, 0xe9, 0xa6, 0x60, 0x86, 0xcf, 0xa2, 0x73, 0x9c, 0x41, + 0x47, 0x9f, 0x43, 0x65, 0x4c, 0xac, 0x09, 0x1b, 0xb7, 0x8a, 0x77, 0x0a, 0xeb, 0xcd, 0xcd, 0x47, + 0xd7, 0x58, 0x67, 0x5b, 0x00, 0x0d, 0x98, 0xc5, 0x08, 0x56, 0xa8, 0xe8, 0x2e, 0x20, 0xd9, 0x32, + 0x1d, 0x42, 0xed, 0xc8, 0x0d, 0xb9, 0xff, 0xb5, 0x4a, 0x77, 0x0a, 0xeb, 0x35, 0x7c, 0x43, 0x72, + 0xb6, 0xa6, 0x8c, 0xf6, 0xfb, 0xb0, 0x76, 0x41, 0x5a, 0xa4, 0x41, 0xe9, 0x84, 0x9c, 0x0b, 0x8b, + 0xd4, 0x30, 0x6f, 0xa2, 0x5b, 0x50, 0x3e, 0xb5, 0x26, 0x31, 0x11, 0x22, 0xd7, 0xb0, 0xec, 0x3c, + 0x2c, 0xde, 0x2f, 0xe8, 0x0f, 0xa0, 0x9e, 0x11, 0x02, 0x35, 0x01, 0x0e, 0xfb, 0x5b, 0xc6, 0xd0, + 0xe8, 0x0d, 0x8d, 0x2d, 0x6d, 0x09, 0xad, 0x42, 0xed, 0xb0, 0xbf, 0x6d, 0x74, 0x77, 0x87, 0xdb, + 0xcf, 0xb4, 0x02, 0xaa, 0xc3, 0x4a, 0xd2, 0x29, 0xea, 0x67, 0x80, 0x30, 0xb1, 0x83, 0x53, 0x12, + 0x71, 0xaf, 0x54, 0x26, 0x42, 0x2f, 0xc1, 0x0a, 0xb3, 0xe8, 0x89, 0xe9, 0x3a, 0x4a, 0x80, 0x0a, + 0xef, 0xee, 0x38, 0x68, 0x07, 0x2a, 0x63, 0xcb, 0x77, 0x26, 0x52, 0x88, 0xfa, 0xe6, 0x3b, 0x39, + 0xf5, 0xc6, 0xc1, 0xb7, 0xc5, 0x44, 0xac, 0x00, 0xb8, 0xab, 0xce, 0xac, 0x2c, 0xb5, 0xa9, 0x3f, + 0x03, 0x6d, 0xc0, 0xac, 0x88, 0x65, 0xc5, 0x31, 0x60, 0x99, 0xaf, 0xaf, 0xdc, 0x73, 0x9e, 0x35, + 0xe5, 0x36, 0xc3, 0x62, 0xba, 0xfe, 0xaf, 0x22, 0xdc, 0xc8, 0x60, 0x2b, 0xb7, 0x7b, 0x0a, 0x95, + 0x88, 0xd0, 0x78, 0xc2, 0x04, 0x7c, 0x73, 0xf3, 0xc3, 0x9c, 0xf0, 0x97, 0x90, 0x3a, 0x58, 0xc0, + 0x60, 0x05, 0x87, 0xd6, 0x41, 0x93, 0x33, 0x4c, 0x12, 0x45, 0x41, 0x64, 0x7a, 0x74, 0xa4, 0x4c, + 0xd7, 0x94, 0x74, 0x83, 0x93, 0xf7, 0xe8, 0x28, 0xa3, 0xd5, 0xd2, 0x35, 0xb5, 0x8a, 0x2c, 0xd0, + 0x7c, 0xc2, 0x9e, 0x07, 0xd1, 0x89, 0xc9, 0x55, 0x1b, 0xb9, 0x0e, 0x69, 0x2d, 0x0b, 0xd0, 0xf7, + 0x72, 0x82, 0xf6, 0xe5, 0xf4, 0x7d, 0x35, 0x1b, 0xaf, 0xf9, 0xb3, 0x04, 0xfd, 0x4d, 0xa8, 0xc8, + 0x3f, 0xe5, 0x9e, 0x34, 0x38, 0xec, 0xf5, 0x8c, 0xc1, 0x40, 0x5b, 0x42, 0x35, 0x28, 0x63, 0x63, + 0x88, 0xb9, 0x87, 0xd5, 0xa0, 0xfc, 0xa8, 0x3b, 0xec, 0xee, 0x6a, 0x45, 0xfd, 0x0d, 0x58, 0x7b, + 0x6a, 0xb9, 0x2c, 0x8f, 0x73, 0xe9, 0x01, 0x68, 0xd3, 0xb1, 0xca, 0x3a, 0x3b, 0x33, 0xd6, 0xc9, + 0xaf, 0x1a, 0xe3, 0xcc, 0x65, 0x17, 0xec, 0xa1, 0x41, 0x89, 0x44, 0x91, 0x32, 0x01, 0x6f, 0xea, + 0xcf, 0x61, 0x6d, 0xc0, 0x82, 0x30, 0x97, 0xe7, 0xbf, 0x0b, 0x2b, 0x3c, 0x4e, 0x04, 0x31, 0x53, + 0xae, 0xff, 0x72, 0x47, 0xc6, 0x91, 0x4e, 0x12, 0x47, 0x3a, 0x5b, 0x2a, 0xce, 0xe0, 0x64, 0x24, + 0xfa, 0x1f, 0xa8, 0x50, 0x77, 0xe4, 0x5b, 0x13, 0xb5, 0xf5, 0x55, 0x4f, 0x47, 0xdc, 0xc9, 0x93, + 0x85, 0x95, 0xe3, 0xf7, 0x00, 0x6d, 0x11, 0xca, 0xa2, 0xe0, 0x3c, 0x97, 0x3c, 0xb7, 0xa0, 0x7c, + 0x1c, 0x44, 0xb6, 0xdc, 0x88, 0x55, 0x2c, 0x3b, 0x7c, 0x53, 0xcd, 0x80, 0x28, 0xec, 0xbb, 0x80, + 0x76, 0x7c, 0x1e, 0x20, 0xf2, 0x19, 0xe2, 0x57, 0x45, 0xb8, 0x39, 0x33, 0x5e, 0x19, 0x63, 0xf1, + 0x7d, 0xc8, 0x0f, 0xa6, 0x98, 0xca, 0x7d, 0x88, 0xf6, 0xa1, 0x22, 0x47, 0x28, 0x4d, 0xde, 0x9b, + 0x03, 0x48, 0xc6, 0x1c, 0x05, 0xa7, 0x60, 0xae, 0x74, 0xfa, 0xd2, 0x7f, 0xdb, 0xe9, 0xb5, 0xe4, + 0x3f, 0xe8, 0xd7, 0xea, 0xef, 0xdb, 0x70, 0x23, 0x33, 0x58, 0x29, 0xef, 0x11, 0x94, 0x29, 0x27, + 0x28, 0xed, 0xbd, 0x3d, 0xa7, 0xf6, 0x28, 0x96, 0xd3, 0xf5, 0x9b, 0x12, 0xdc, 0x38, 0x25, 0x7e, + 0x2a, 0x8a, 0xbe, 0x05, 0x37, 0x06, 0xc2, 0xb5, 0x72, 0xf9, 0xce, 0xd4, 0x2d, 0x8b, 0x33, 0x6e, + 0x79, 0x0b, 0x50, 0x16, 0x45, 0x39, 0xcf, 0x39, 0xac, 0x19, 0x67, 0xc4, 0xce, 0x85, 0xdc, 0x82, + 0x15, 0x3b, 0xf0, 0x3c, 0xcb, 0x77, 0x5a, 0xc5, 0x3b, 0xa5, 0xf5, 0x1a, 0x4e, 0xba, 0xd9, 0xfd, + 0x53, 0xca, 0xbb, 0x7f, 0xf4, 0x5f, 0x16, 0x40, 0x9b, 0xae, 0xad, 0x14, 0xc9, 0xa5, 0x67, 0x0e, + 0x07, 0xe2, 0x6b, 0x37, 0xb0, 0xea, 0x29, 0x7a, 0xb2, 0xc5, 0x25, 0x9d, 0x44, 0x51, 0xe6, 0x08, + 0x29, 0x5d, 0xf3, 0x08, 0xd1, 0xff, 0x51, 0x00, 0x74, 0x39, 0xeb, 0x41, 0xaf, 0x41, 0x83, 0x12, + 0xdf, 0x31, 0xa5, 0x1a, 0xa5, 0x85, 0xab, 0xb8, 0xce, 0x69, 0x52, 0x9f, 0x14, 0x21, 0x58, 0x26, + 0x67, 0xc4, 0x56, 0xbb, 0x55, 0xb4, 0xd1, 0x18, 0x1a, 0xc7, 0xd4, 0x74, 0x69, 0x30, 0xb1, 0xd2, + 0xf4, 0xa0, 0xb9, 0x69, 0x2c, 0x9c, 0x7d, 0x75, 0x1e, 0x0d, 0x76, 0x12, 0x30, 0x5c, 0x3f, 0xa6, + 0x69, 0x47, 0xef, 0x40, 0x3d, 0xc3, 0x43, 0x55, 0x58, 0xee, 0xef, 0xf7, 0x0d, 0x6d, 0x09, 0x01, + 0x54, 0x7a, 0xdb, 0x78, 0x7f, 0x7f, 0x28, 0x4f, 0xed, 0x9d, 0xbd, 0xee, 0x63, 0x43, 0x2b, 0xea, + 0xff, 0x2e, 0x01, 0x4c, 0xc3, 0x27, 0x6a, 0x42, 0x31, 0xb5, 0x74, 0xd1, 0x75, 0xf8, 0xcf, 0xf8, + 0x96, 0x97, 0x24, 0x22, 0xa2, 0x8d, 0x36, 0xe1, 0xb6, 0x47, 0x47, 0xa1, 0x65, 0x9f, 0x98, 0x2a, + 0xea, 0xd9, 0x62, 0xb2, 0xf8, 0xab, 0x06, 0xbe, 0xa9, 0x98, 0x4a, 0x6a, 0x89, 0xbb, 0x0b, 0x25, + 0xe2, 0x9f, 0xb6, 0x96, 0x45, 0xaa, 0xf7, 0x70, 0xee, 0xb0, 0xde, 0x31, 0xfc, 0x53, 0x99, 0xda, + 0x71, 0x18, 0xd4, 0x87, 0x5a, 0x44, 0x68, 0x10, 0x47, 0x36, 0xa1, 0xad, 0xf2, 0x5c, 0x9b, 0x0c, + 0x27, 0xf3, 0xf0, 0x14, 0x02, 0x6d, 0x41, 0xc5, 0x0b, 0x62, 0x9f, 0xd1, 0x56, 0x45, 0x08, 0xf8, + 0x56, 0x4e, 0xb0, 0x3d, 0x3e, 0x09, 0xab, 0xb9, 0xe8, 0x31, 0xac, 0x38, 0xe4, 0xd4, 0xe5, 0x32, + 0xad, 0x08, 0x98, 0xbb, 0x79, 0xed, 0x2b, 0x66, 0xe1, 0x64, 0x36, 0x57, 0x7a, 0x4c, 0x49, 0xd4, + 0xaa, 0x4a, 0xa5, 0xf3, 0x36, 0x7a, 0x05, 0x6a, 0xd6, 0x64, 0x12, 0xd8, 0xa6, 0xe3, 0x46, 0xad, + 0x9a, 0x60, 0x54, 0x05, 0x61, 0xcb, 0x8d, 0xda, 0xef, 0x41, 0x35, 0x51, 0xd0, 0x5c, 0xd9, 0xe4, + 0x5f, 0x0b, 0x50, 0x4b, 0x15, 0x82, 0x3e, 0x85, 0xd5, 0xc8, 0x7a, 0x6e, 0x4e, 0x35, 0x2b, 0x8f, + 0xaf, 0x77, 0xf3, 0x6a, 0xd6, 0x7a, 0x3e, 0x55, 0x6e, 0x23, 0xca, 0xf4, 0xd0, 0xe7, 0xb0, 0x36, + 0x71, 0xfd, 0xf8, 0x2c, 0x83, 0x2d, 0xe3, 0xc1, 0xff, 0xe7, 0xc4, 0xde, 0xe5, 0xb3, 0xa7, 0xe8, + 0xcd, 0xc9, 0x4c, 0x5f, 0xff, 0x43, 0x01, 0x1a, 0xd9, 0xe5, 0xb9, 0x12, 0xec, 0x30, 0x16, 0x3f, + 0x50, 0xc2, 0xbc, 0xc9, 0x8f, 0x0c, 0x8f, 0x78, 0x41, 0x74, 0x2e, 0x56, 0x2e, 0x61, 0xd5, 0xe3, + 0xba, 0x76, 0x5c, 0x7a, 0x22, 0x7c, 0xb7, 0x84, 0x45, 0x9b, 0xd3, 0xdc, 0x20, 0xa4, 0x22, 0x9b, + 0x2a, 0x61, 0xd1, 0x46, 0x18, 0xaa, 0x2a, 0x50, 0x70, 0x8f, 0x2b, 0xcd, 0x1f, 0x70, 0x12, 0xe1, + 0x70, 0x8a, 0xa3, 0xff, 0xa6, 0x08, 0x6b, 0x17, 0xb8, 0x5c, 0x4e, 0xe9, 0x06, 0xc9, 0x71, 0x2b, + 0x7b, 0x5c, 0x26, 0xdb, 0x75, 0x92, 0x9c, 0x46, 0xb4, 0xc5, 0x66, 0x0d, 0x55, 0xbe, 0x51, 0x74, + 0x43, 0x6e, 0x68, 0xef, 0xc8, 0x65, 0x52, 0xf0, 0x32, 0x96, 0x1d, 0xf4, 0x0c, 0x9a, 0x11, 0xa1, + 0x24, 0x3a, 0x25, 0x8e, 0x19, 0x06, 0x11, 0x4b, 0xe4, 0xdf, 0x9c, 0x4f, 0xfe, 0x83, 0x20, 0x62, + 0x78, 0x35, 0x41, 0xe2, 0x3d, 0x8a, 0x9e, 0xc2, 0xaa, 0x73, 0xee, 0x5b, 0x9e, 0x6b, 0x2b, 0xe4, + 0xca, 0xc2, 0xc8, 0x0d, 0x05, 0x24, 0x80, 0xf9, 0x35, 0x27, 0xc3, 0xe4, 0x3f, 0x36, 0xb1, 0x8e, + 0xc8, 0x44, 0xe9, 0x44, 0x76, 0x66, 0xfd, 0xba, 0xac, 0xfc, 0x5a, 0xff, 0x49, 0x11, 0x9a, 0xb3, + 0xee, 0x82, 0xfe, 0x17, 0xc0, 0x0e, 0x63, 0x33, 0x24, 0x91, 0x1b, 0x38, 0xca, 0x29, 0x6a, 0x76, + 0x18, 0x1f, 0x08, 0x02, 0xdf, 0x5a, 0x9c, 0xfd, 0x65, 0x1c, 0x30, 0x4b, 0x79, 0x47, 0xd5, 0x0e, + 0xe3, 0x6f, 0xf2, 0x7e, 0x32, 0x57, 0xdc, 0xc3, 0xa9, 0xf2, 0x12, 0x3e, 0x7c, 0x20, 0x08, 0xe8, + 0x1d, 0xb8, 0x2d, 0x1d, 0xc9, 0x9c, 0xb8, 0x9e, 0xcb, 0x4c, 0xd7, 0x37, 0x8f, 0xce, 0xf9, 0xa5, + 0x56, 0xfa, 0x0e, 0x92, 0xcc, 0x5d, 0xce, 0xdb, 0xf1, 0x3f, 0xe6, 0x1c, 0xa4, 0xc3, 0x6a, 0x10, + 0x78, 0x26, 0xb5, 0x83, 0x88, 0x98, 0x96, 0xf3, 0x85, 0x38, 0xc0, 0x4a, 0xb8, 0x1e, 0x04, 0xde, + 0x80, 0xd3, 0xba, 0xce, 0x17, 0xe8, 0x55, 0xa8, 0xdb, 0x61, 0x4c, 0x09, 0x33, 0xf9, 0xa7, 0x55, + 0x11, 0xbf, 0x0d, 0x92, 0xd4, 0x0b, 0x63, 0x9a, 0x19, 0xe0, 0x11, 0x8f, 0x9f, 0x37, 0x99, 0x01, + 0x7b, 0xc4, 0xa3, 0xfa, 0x67, 0x50, 0x16, 0xa7, 0x13, 0xff, 0x3b, 0x11, 0xc0, 0x43, 0x8b, 0x8d, + 0x95, 0xfe, 0xaa, 0x9c, 0x70, 0x60, 0xb1, 0x31, 0x67, 0x8e, 0x03, 0xca, 0x24, 0x53, 0xba, 0x56, + 0x95, 0x13, 0x04, 0xb3, 0x0d, 0xd5, 0x88, 0x58, 0x4e, 0xe0, 0x4f, 0xce, 0xc5, 0x8f, 0x57, 0x71, + 0xda, 0xd7, 0xbf, 0x84, 0x8a, 0x3c, 0xb5, 0xae, 0x81, 0x7f, 0x17, 0x90, 0x3d, 0x8a, 0x82, 0x38, + 0xe4, 0x96, 0xf1, 0x5c, 0x4a, 0xdd, 0xc0, 0xa7, 0xc9, 0xcd, 0x59, 0x72, 0x0e, 0xa6, 0x0c, 0xfd, + 0x2f, 0x05, 0x19, 0xa9, 0xe4, 0x35, 0x88, 0xc7, 0x7a, 0x15, 0x76, 0x16, 0xbe, 0x2b, 0x2a, 0x80, + 0x24, 0x5f, 0x23, 0xea, 0x85, 0x60, 0xde, 0x7c, 0x8d, 0xc8, 0x7c, 0x8d, 0xf0, 0xe4, 0x40, 0x05, + 0x44, 0x09, 0x27, 0xe3, 0x61, 0xdd, 0x49, 0x13, 0x59, 0xa2, 0xff, 0xb3, 0x90, 0x6e, 0xf9, 0x24, + 0xe1, 0x44, 0x9f, 0x43, 0x95, 0xef, 0x1e, 0xd3, 0xb3, 0x42, 0xf5, 0x16, 0xd2, 0x5b, 0x2c, 0x97, + 0xed, 0xf0, 0xcd, 0xb2, 0x67, 0x85, 0x32, 0x52, 0xae, 0x84, 0xb2, 0xc7, 0x8f, 0x0e, 0xcb, 0x99, + 0x1e, 0x1d, 0xbc, 0x8d, 0x5e, 0x87, 0xa6, 0x15, 0xb3, 0xc0, 0xb4, 0x9c, 0x53, 0x12, 0x31, 0x97, + 0x12, 0x65, 0xe1, 0x55, 0x4e, 0xed, 0x26, 0xc4, 0xf6, 0x43, 0x68, 0x64, 0x31, 0xbf, 0x2e, 0xb8, + 0x94, 0xb3, 0xc1, 0xe5, 0xbb, 0x00, 0xd3, 0xbc, 0x8a, 0x7b, 0x02, 0x39, 0x73, 0x99, 0x69, 0x07, + 0x8e, 0x3c, 0xda, 0xca, 0xb8, 0xca, 0x09, 0xbd, 0xc0, 0x21, 0x17, 0xb2, 0xd4, 0x72, 0x92, 0xa5, + 0xf2, 0xcd, 0xc7, 0xb7, 0xca, 0x89, 0x3b, 0x99, 0x10, 0x47, 0x49, 0x58, 0x0b, 0x02, 0xef, 0x89, + 0x20, 0xe8, 0x7f, 0x2e, 0x4a, 0x8f, 0x90, 0x77, 0x84, 0x5c, 0xb9, 0x4b, 0x6a, 0xea, 0xd2, 0xf5, + 0x4c, 0xfd, 0x00, 0x80, 0x32, 0x2b, 0x62, 0xc4, 0x31, 0x2d, 0xa6, 0xae, 0xdd, 0xed, 0x4b, 0x69, + 0xee, 0x30, 0x79, 0x6e, 0xc4, 0x35, 0x35, 0xba, 0xcb, 0xd0, 0xfb, 0xd0, 0xb0, 0x03, 0x2f, 0x9c, + 0x10, 0x35, 0xb9, 0xfc, 0xb5, 0x93, 0xeb, 0xe9, 0xf8, 0x2e, 0xcb, 0xe4, 0xb8, 0x95, 0xeb, 0xe6, + 0xb8, 0x7f, 0x2c, 0xc8, 0xab, 0x4e, 0xf6, 0xa6, 0x85, 0x46, 0x57, 0xbc, 0xcd, 0x3d, 0x5e, 0xf0, + 0xda, 0xf6, 0x55, 0x0f, 0x73, 0xd7, 0x7d, 0x09, 0xfb, 0x53, 0x09, 0x6a, 0xe9, 0x8d, 0xe9, 0x92, + 0xed, 0xef, 0x43, 0x2d, 0x7d, 0xeb, 0x55, 0xb9, 0xc6, 0x57, 0x9a, 0x27, 0x1d, 0x8c, 0x8e, 0x01, + 0x59, 0xa3, 0x51, 0x9a, 0xa9, 0x98, 0x31, 0xb5, 0x46, 0xc9, 0x1d, 0xf3, 0xfe, 0x1c, 0x7a, 0x48, + 0xc2, 0xcf, 0x21, 0x9f, 0x8f, 0x35, 0x6b, 0x34, 0x9a, 0xa1, 0xa0, 0xef, 0xc1, 0xed, 0xd9, 0x35, + 0xcc, 0xa3, 0x73, 0x33, 0x74, 0x1d, 0x95, 0x23, 0x6f, 0xcf, 0x7b, 0x69, 0xec, 0xcc, 0xc0, 0x7f, + 0x7c, 0x7e, 0xe0, 0x3a, 0x52, 0xe7, 0x28, 0xba, 0xc4, 0x68, 0xff, 0x00, 0x5e, 0x7a, 0xc1, 0xf0, + 0x2b, 0x6c, 0xd0, 0xcf, 0xda, 0xe0, 0x3a, 0x4a, 0xc8, 0x58, 0xef, 0xb7, 0x05, 0x79, 0xb7, 0x9d, + 0xd5, 0x49, 0x77, 0x9a, 0xb6, 0xd5, 0x37, 0x37, 0x72, 0xae, 0xd3, 0x3b, 0x38, 0x94, 0xf0, 0x22, + 0xcf, 0xfb, 0x64, 0x26, 0xcf, 0xcb, 0x9f, 0x8b, 0xec, 0x89, 0x49, 0x12, 0x48, 0x21, 0xe8, 0xbf, + 0x2f, 0x41, 0x35, 0x41, 0xe7, 0x11, 0x97, 0x9e, 0x53, 0x46, 0x3c, 0xd3, 0x4b, 0x8e, 0xb0, 0x02, + 0x06, 0x49, 0xda, 0xe3, 0x87, 0xd8, 0x2b, 0x50, 0xe3, 0x99, 0xba, 0x64, 0x17, 0x05, 0xbb, 0xca, + 0x09, 0x82, 0xf9, 0x2a, 0xd4, 0x59, 0xc0, 0xac, 0x89, 0xc9, 0x5c, 0xfb, 0x44, 0x06, 0xb9, 0x02, + 0x06, 0x41, 0x1a, 0x72, 0x0a, 0x7a, 0x13, 0x6e, 0xb0, 0x71, 0x14, 0x30, 0x36, 0xe1, 0x69, 0x9a, + 0x48, 0x4c, 0x64, 0x12, 0xb1, 0x8c, 0xb5, 0x94, 0x21, 0x13, 0x16, 0xca, 0x4f, 0xef, 0xe9, 0x60, + 0xee, 0xba, 0xe2, 0x10, 0x59, 0xc6, 0xab, 0x29, 0x95, 0xbb, 0x36, 0xbf, 0xa2, 0x87, 0x24, 0xb2, + 0x89, 0x2f, 0xcf, 0x8a, 0x02, 0x4e, 0xba, 0xc8, 0x84, 0x35, 0x8f, 0x58, 0x34, 0x8e, 0x88, 0x63, + 0x1e, 0xbb, 0x64, 0xe2, 0xc8, 0x2b, 0x4b, 0x33, 0x77, 0x52, 0x9b, 0xa8, 0xa5, 0xf3, 0x48, 0xcc, + 0xc6, 0xcd, 0x04, 0x4e, 0xf6, 0x79, 0x7e, 0x20, 0x5b, 0x68, 0x0d, 0xea, 0x83, 0x67, 0x83, 0xa1, + 0xb1, 0x67, 0xee, 0xed, 0x6f, 0x19, 0xea, 0x8d, 0x7a, 0x60, 0x60, 0xd9, 0x2d, 0x70, 0xfe, 0x70, + 0x7f, 0xd8, 0xdd, 0x35, 0x87, 0x3b, 0xbd, 0x27, 0x03, 0xad, 0x88, 0x6e, 0xc3, 0x8d, 0xe1, 0x36, + 0xde, 0x1f, 0x0e, 0x77, 0x8d, 0x2d, 0xf3, 0xc0, 0xc0, 0x3b, 0xfb, 0x5b, 0x03, 0xad, 0x84, 0x10, + 0x34, 0xa7, 0xe4, 0xe1, 0xce, 0x9e, 0xa1, 0x2d, 0xa3, 0x3a, 0xac, 0x1c, 0x18, 0xb8, 0x67, 0xf4, + 0x87, 0x5a, 0x59, 0xff, 0x5b, 0x11, 0xea, 0x19, 0x2b, 0x72, 0x47, 0x8e, 0xa8, 0xbc, 0xc4, 0x2c, + 0x63, 0xde, 0xe4, 0x87, 0x89, 0x6d, 0xd9, 0x63, 0x69, 0x9d, 0x65, 0x2c, 0x3b, 0xdc, 0x6e, 0x9e, + 0x75, 0x96, 0xd9, 0xe7, 0xcb, 0xb8, 0xea, 0x59, 0x67, 0x12, 0xe4, 0x35, 0x68, 0x9c, 0x90, 0xc8, + 0x27, 0x13, 0xc5, 0x97, 0x16, 0xa9, 0x4b, 0x9a, 0x1c, 0xb2, 0x0e, 0x9a, 0x1a, 0x32, 0x85, 0x91, + 0xe6, 0x68, 0x4a, 0xfa, 0x5e, 0x02, 0x76, 0x74, 0x59, 0xeb, 0x15, 0xa1, 0xf5, 0x07, 0xf3, 0x3b, + 0xe9, 0x8b, 0x14, 0x3f, 0x48, 0x15, 0xbf, 0x02, 0x25, 0x9c, 0x3c, 0xd7, 0xf6, 0xba, 0xbd, 0x6d, + 0xae, 0xec, 0x55, 0xa8, 0xed, 0x75, 0x3f, 0x35, 0x0f, 0x07, 0xe2, 0xf2, 0x8f, 0x34, 0x68, 0x3c, + 0x31, 0x70, 0xdf, 0xd8, 0x55, 0x94, 0x12, 0xba, 0x05, 0x9a, 0xa2, 0x4c, 0xc7, 0x2d, 0xeb, 0xbf, + 0x2b, 0xc2, 0x9a, 0x3c, 0xd7, 0xd3, 0xf7, 0xa8, 0x17, 0x3f, 0x0c, 0x2d, 0x7e, 0xf4, 0xb6, 0x60, + 0xc5, 0x23, 0x34, 0xb5, 0x43, 0x0d, 0x27, 0x5d, 0xe4, 0x42, 0xdd, 0xf2, 0xfd, 0x80, 0x89, 0x47, + 0x0d, 0xaa, 0x8e, 0xc8, 0xc7, 0x73, 0x3d, 0x9f, 0xa4, 0x92, 0x77, 0xba, 0x53, 0x24, 0x79, 0x42, + 0x66, 0xb1, 0xdb, 0x1f, 0x80, 0x76, 0x71, 0xc0, 0x3c, 0x71, 0xe9, 0x8d, 0x77, 0xa6, 0x61, 0x89, + 0x70, 0x07, 0x3d, 0xec, 0x3f, 0xe9, 0xef, 0x3f, 0xed, 0x6b, 0x4b, 0xbc, 0x83, 0x0f, 0xfb, 0xfd, + 0x9d, 0xfe, 0x63, 0xad, 0x80, 0x00, 0x2a, 0xc6, 0xa7, 0x3b, 0x43, 0x63, 0x4b, 0x2b, 0x6e, 0xfe, + 0x7d, 0x15, 0x2a, 0x52, 0x48, 0xf4, 0x6b, 0x15, 0x92, 0xb3, 0x45, 0x43, 0xf4, 0xc1, 0xdc, 0xa9, + 0xed, 0x4c, 0x21, 0xb2, 0xfd, 0xe1, 0xc2, 0xf3, 0xd5, 0xbb, 0xe0, 0x12, 0xfa, 0x79, 0x01, 0x1a, + 0x33, 0x0f, 0x61, 0x79, 0xdf, 0x70, 0xae, 0xa8, 0x51, 0xb6, 0xbf, 0xb1, 0xd0, 0xdc, 0x54, 0x96, + 0x9f, 0x15, 0xa0, 0x9e, 0xa9, 0xce, 0xa1, 0x07, 0x8b, 0x54, 0xf4, 0xa4, 0x24, 0x0f, 0x17, 0x2f, + 0x06, 0xea, 0x4b, 0x6f, 0x17, 0xd0, 0x4f, 0x0b, 0x50, 0xcf, 0x94, 0xb6, 0x72, 0x8b, 0x72, 0xb9, + 0x10, 0x97, 0x5b, 0x94, 0xab, 0x2a, 0x69, 0x4b, 0xe8, 0x87, 0x05, 0xa8, 0xa5, 0x65, 0x2a, 0x74, + 0x6f, 0xfe, 0xc2, 0x96, 0x14, 0xe2, 0xfe, 0xa2, 0x15, 0x31, 0x7d, 0x09, 0x7d, 0x1f, 0xaa, 0x49, + 0x4d, 0x07, 0xe5, 0x0d, 0x23, 0x17, 0x0a, 0x46, 0xed, 0x7b, 0x73, 0xcf, 0xcb, 0x2e, 0x9f, 0x14, + 0x5a, 0x72, 0x2f, 0x7f, 0xa1, 0x24, 0xd4, 0xbe, 0x37, 0xf7, 0xbc, 0x74, 0x79, 0xee, 0x09, 0x99, + 0x7a, 0x4c, 0x6e, 0x4f, 0xb8, 0x5c, 0x08, 0xca, 0xed, 0x09, 0x57, 0x95, 0x7f, 0xa4, 0x20, 0x99, + 0x8a, 0x4e, 0x6e, 0x41, 0x2e, 0x57, 0x8d, 0x72, 0x0b, 0x72, 0x45, 0x01, 0x49, 0xb9, 0xe4, 0x34, + 0x41, 0xbf, 0x37, 0x77, 0x11, 0x64, 0x4e, 0x97, 0xbc, 0x54, 0x86, 0xd1, 0x97, 0xd0, 0x8f, 0xd4, + 0x93, 0x81, 0xac, 0xa0, 0xa0, 0x79, 0xa0, 0x66, 0x8a, 0x2e, 0xed, 0xf7, 0x16, 0x0b, 0x35, 0xe2, + 0x8c, 0xf8, 0x71, 0x01, 0x60, 0x5a, 0x6b, 0xc9, 0x2d, 0xc4, 0xa5, 0x22, 0x4f, 0xfb, 0xc1, 0x02, + 0x33, 0xb3, 0xdb, 0x23, 0x29, 0xaf, 0xe4, 0xde, 0x1e, 0x17, 0x6a, 0x41, 0xb9, 0xb7, 0xc7, 0xc5, + 0x3a, 0x8e, 0xbe, 0xf4, 0xf1, 0xca, 0xb7, 0xca, 0x32, 0xf6, 0x57, 0xc4, 0xe7, 0xdd, 0xff, 0x04, + 0x00, 0x00, 0xff, 0xff, 0x96, 0x01, 0xe4, 0xc4, 0xf0, 0x23, 0x00, 0x00, } diff --git a/plugins/drivers/base/proto/driver.proto b/plugins/drivers/proto/driver.proto similarity index 95% rename from plugins/drivers/base/proto/driver.proto rename to plugins/drivers/proto/driver.proto index 36c878144..9c870d27e 100644 --- a/plugins/drivers/base/proto/driver.proto +++ b/plugins/drivers/proto/driver.proto @@ -1,5 +1,5 @@ syntax = "proto3"; -package hashicorp.nomad.plugins.drivers.base.proto; +package hashicorp.nomad.plugins.drivers.proto; option go_package = "proto"; import "google/protobuf/duration.proto"; @@ -48,13 +48,9 @@ service Driver { // DestroyTask removes the task from the driver's internal state and cleans // up any additional resources created by the driver. It cannot be called - // on a running task. + // on a running task, unless force is set to true. rpc DestroyTask(DestroyTaskRequest) returns (DestroyTaskResponse) {} - // ListTasks returns a list of summary information of all the tasks the - // driver is tracking. - rpc ListTasks(ListTasksRequest) returns (ListTasksResponse) {} - // InspectTask returns detailed information for the given task rpc InspectTask(InspectTaskRequest) returns (InspectTaskResponse) {} @@ -205,18 +201,13 @@ message DestroyTaskRequest { // TaskId is the ID of the target task string task_id = 1; + + // Force destroys the task even if it is still in a running state + bool force = 2; } message DestroyTaskResponse {} -message ListTasksRequest {} - -message ListTasksResponse { - - // Tasks includes a list of summary information for each task - repeated TaskStatus tasks = 1; -} - message InspectTaskRequest { // TaskId is the ID of the target task @@ -471,19 +462,15 @@ message TaskStatus { // State is the state of the task's execution TaskState state = 3; - // SizeOnDiskMb is the disk space the driver reports the task is consuming - // in megabytes. - int64 size_on_disk_mb = 4; - // StartedAt is the timestamp when the task was started - google.protobuf.Timestamp started_at = 5; + google.protobuf.Timestamp started_at = 4; // CompletedAt is the timestamp when the task exited. // If the task is still running, CompletedAt will not be set - google.protobuf.Timestamp completed_at = 6; + google.protobuf.Timestamp completed_at = 5; // Result is set when CompletedAt is set. - ExitResult result = 7; + ExitResult result = 6; } message TaskDriverStatus { @@ -548,7 +535,7 @@ message MemoryUsage { enum Fields { RSS = 0; CACHE = 1; - MAX_UASGE = 2; + MAX_USAGE = 2; KERNEL_USAGE = 3; KERNEL_MAX_USAGE = 4; } diff --git a/plugins/drivers/server.go b/plugins/drivers/server.go new file mode 100644 index 000000000..4392626b6 --- /dev/null +++ b/plugins/drivers/server.go @@ -0,0 +1,254 @@ +package drivers + +import ( + "io" + + "golang.org/x/net/context" + + "github.com/golang/protobuf/ptypes" + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/plugins/drivers/proto" +) + +type driverPluginServer struct { + broker *plugin.GRPCBroker + impl DriverPlugin + logger hclog.Logger +} + +func (b *driverPluginServer) TaskConfigSchema(ctx context.Context, req *proto.TaskConfigSchemaRequest) (*proto.TaskConfigSchemaResponse, error) { + spec, err := b.impl.TaskConfigSchema() + if err != nil { + return nil, err + } + + resp := &proto.TaskConfigSchemaResponse{ + Spec: spec, + } + return resp, nil +} + +func (b *driverPluginServer) Capabilities(ctx context.Context, req *proto.CapabilitiesRequest) (*proto.CapabilitiesResponse, error) { + caps, err := b.impl.Capabilities() + if err != nil { + return nil, err + } + resp := &proto.CapabilitiesResponse{ + Capabilities: &proto.DriverCapabilities{ + SendSignals: caps.SendSignals, + Exec: caps.Exec, + }, + } + + switch caps.FSIsolation { + case FSIsolationNone: + resp.Capabilities.FsIsolation = proto.DriverCapabilities_NONE + case FSIsolationChroot: + resp.Capabilities.FsIsolation = proto.DriverCapabilities_CHROOT + case FSIsolationImage: + resp.Capabilities.FsIsolation = proto.DriverCapabilities_IMAGE + } + return resp, nil +} + +func (b *driverPluginServer) Fingerprint(req *proto.FingerprintRequest, srv proto.Driver_FingerprintServer) error { + ctx := srv.Context() + ch, err := b.impl.Fingerprint(ctx) + if err != nil { + return err + } + + for { + select { + case <-ctx.Done(): + return nil + case f, ok := <-ch: + + if !ok { + return nil + } + resp := &proto.FingerprintResponse{ + Attributes: f.Attributes, + Health: healthStateToProto(f.Health), + HealthDescription: f.HealthDescription, + } + + if err := srv.Send(resp); err != nil { + return err + } + } + } +} + +func (b *driverPluginServer) RecoverTask(ctx context.Context, req *proto.RecoverTaskRequest) (*proto.RecoverTaskResponse, error) { + err := b.impl.RecoverTask(taskHandleFromProto(req.Handle)) + if err != nil { + return nil, err + } + + return &proto.RecoverTaskResponse{}, nil +} + +func (b *driverPluginServer) StartTask(ctx context.Context, req *proto.StartTaskRequest) (*proto.StartTaskResponse, error) { + handle, err := b.impl.StartTask(taskConfigFromProto(req.Task)) + if err != nil { + return nil, err + } + + resp := &proto.StartTaskResponse{ + Handle: taskHandleToProto(handle), + } + + return resp, nil +} + +func (b *driverPluginServer) WaitTask(ctx context.Context, req *proto.WaitTaskRequest) (*proto.WaitTaskResponse, error) { + ch, err := b.impl.WaitTask(ctx, req.TaskId) + if err != nil { + return nil, err + } + + result := <-ch + var errStr string + if result.Err != nil { + errStr = result.Err.Error() + } + + resp := &proto.WaitTaskResponse{ + Err: errStr, + Result: &proto.ExitResult{ + ExitCode: int32(result.ExitCode), + Signal: int32(result.Signal), + OomKilled: result.OOMKilled, + }, + } + + return resp, nil +} + +func (b *driverPluginServer) StopTask(ctx context.Context, req *proto.StopTaskRequest) (*proto.StopTaskResponse, error) { + timeout, err := ptypes.Duration(req.Timeout) + if err != nil { + return nil, err + } + + err = b.impl.StopTask(req.TaskId, timeout, req.Signal) + if err != nil { + return nil, err + } + return &proto.StopTaskResponse{}, nil +} + +func (b *driverPluginServer) DestroyTask(ctx context.Context, req *proto.DestroyTaskRequest) (*proto.DestroyTaskResponse, error) { + err := b.impl.DestroyTask(req.TaskId, req.Force) + if err != nil { + return nil, err + } + return &proto.DestroyTaskResponse{}, nil +} + +func (b *driverPluginServer) InspectTask(ctx context.Context, req *proto.InspectTaskRequest) (*proto.InspectTaskResponse, error) { + status, err := b.impl.InspectTask(req.TaskId) + if err != nil { + return nil, err + } + + protoStatus, err := taskStatusToProto(status) + if err != nil { + return nil, err + } + + resp := &proto.InspectTaskResponse{ + Task: protoStatus, + Driver: &proto.TaskDriverStatus{ + Attributes: status.DriverAttributes, + }, + NetworkOverride: &proto.NetworkOverride{ + PortMap: status.NetworkOverride.PortMap, + Addr: status.NetworkOverride.Addr, + AutoAdvertise: status.NetworkOverride.AutoAdvertise, + }, + } + + return resp, nil +} + +func (b *driverPluginServer) TaskStats(ctx context.Context, req *proto.TaskStatsRequest) (*proto.TaskStatsResponse, error) { + stats, err := b.impl.TaskStats(req.TaskId) + if err != nil { + return nil, err + } + + pb, err := taskStatsToProto(stats) + if err != nil { + return nil, err + } + + resp := &proto.TaskStatsResponse{ + Stats: pb, + } + + return resp, nil +} + +func (b *driverPluginServer) ExecTask(ctx context.Context, req *proto.ExecTaskRequest) (*proto.ExecTaskResponse, error) { + timeout, err := ptypes.Duration(req.Timeout) + if err != nil { + return nil, err + } + + result, err := b.impl.ExecTask(req.TaskId, req.Command, timeout) + if err != nil { + return nil, err + } + resp := &proto.ExecTaskResponse{ + Stdout: result.Stdout, + Stderr: result.Stderr, + Result: exitResultToProto(result.ExitResult), + } + + return resp, nil +} + +func (b *driverPluginServer) SignalTask(ctx context.Context, req *proto.SignalTaskRequest) (*proto.SignalTaskResponse, error) { + err := b.impl.SignalTask(req.TaskId, req.Signal) + if err != nil { + return nil, err + } + + resp := &proto.SignalTaskResponse{} + return resp, nil +} + +func (b *driverPluginServer) TaskEvents(req *proto.TaskEventsRequest, srv proto.Driver_TaskEventsServer) error { + ch, err := b.impl.TaskEvents(srv.Context()) + if err != nil { + return err + } + + for { + event := <-ch + if event == nil { + break + } + pbTimestamp, err := ptypes.TimestampProto(event.Timestamp) + if err != nil { + return err + } + + pbEvent := &proto.DriverTaskEvent{ + TaskId: event.TaskID, + Timestamp: pbTimestamp, + Message: event.Message, + Annotations: event.Annotations, + } + + if err = srv.Send(pbEvent); err == io.EOF { + break + } else if err != nil { + return err + } + } + return nil +} diff --git a/plugins/drivers/task_handle.go b/plugins/drivers/task_handle.go new file mode 100644 index 000000000..9cf030c87 --- /dev/null +++ b/plugins/drivers/task_handle.go @@ -0,0 +1,30 @@ +package drivers + +import ( + "github.com/hashicorp/nomad/nomad/structs" + "github.com/ugorji/go/codec" +) + +// TaskHandle is the state shared between a driver and the client. +// It is returned to the client after starting the task and used +// for recovery of tasks during a driver restart. +type TaskHandle struct { + Driver string + Config *TaskConfig + State TaskState + driverState []byte +} + +func NewTaskHandle(driver string) *TaskHandle { + return &TaskHandle{Driver: driver} +} + +func (h *TaskHandle) SetDriverState(v interface{}) error { + h.driverState = []byte{} + return codec.NewEncoderBytes(&h.driverState, structs.MsgpackHandle).Encode(v) +} + +func (h *TaskHandle) GetDriverState(v interface{}) error { + return codec.NewDecoderBytes(h.driverState, structs.MsgpackHandle).Decode(v) + +} diff --git a/plugins/drivers/testing.go b/plugins/drivers/testing.go new file mode 100644 index 000000000..3cea5d457 --- /dev/null +++ b/plugins/drivers/testing.go @@ -0,0 +1,112 @@ +package drivers + +import ( + "io/ioutil" + "os" + "path/filepath" + "time" + + "github.com/mitchellh/go-testing-interface" + "github.com/stretchr/testify/require" + "golang.org/x/net/context" + + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/shared/hclspec" +) + +type DriverHarness struct { + DriverPlugin + client *plugin.GRPCClient + server *plugin.GRPCServer + t testing.T +} + +func NewDriverHarness(t testing.T, d DriverPlugin) *DriverHarness { + client, server := plugin.TestPluginGRPCConn(t, map[string]plugin.Plugin{ + base.PluginTypeDriver: &PluginDriver{ + impl: d, + logger: testlog.HCLogger(t), + }, + }) + + raw, err := client.Dispense(base.PluginTypeDriver) + if err != nil { + t.Fatalf("err dispensing plugin: %v", err) + } + + dClient := raw.(DriverPlugin) + h := &DriverHarness{ + client: client, + server: server, + DriverPlugin: dClient, + } + + return h +} + +func (h *DriverHarness) Kill() { + h.client.Close() + h.server.Stop() +} + +// MkAllocDir creates a tempory directory and allocdir structure. +// A cleanup func is returned and should be defered so as to not leak dirs +// between tests. +func (h *DriverHarness) MkAllocDir(t *TaskConfig) func() { + allocDir, err := ioutil.TempDir("", "nomad_driver_harness-") + require.NoError(h.t, err) + require.NoError(h.t, os.Mkdir(filepath.Join(allocDir, t.Name), os.ModePerm)) + require.NoError(h.t, os.MkdirAll(filepath.Join(allocDir, "alloc/logs"), os.ModePerm)) + t.AllocDir = allocDir + return func() { os.RemoveAll(allocDir) } +} + +// MockDriver is used for testing. +// Each function can be set as a closure to make assertions about how data +// is passed through the base plugin layer. +type MockDriver struct { + base.MockPlugin + TaskConfigSchemaF func() (*hclspec.Spec, error) + FingerprintF func(context.Context) (<-chan *Fingerprint, error) + CapabilitiesF func() (*Capabilities, error) + RecoverTaskF func(*TaskHandle) error + StartTaskF func(*TaskConfig) (*TaskHandle, error) + WaitTaskF func(context.Context, string) (<-chan *ExitResult, error) + StopTaskF func(string, time.Duration, string) error + DestroyTaskF func(string, bool) error + InspectTaskF func(string) (*TaskStatus, error) + TaskStatsF func(string) (*TaskStats, error) + TaskEventsF func(context.Context) (<-chan *TaskEvent, error) + SignalTaskF func(string, string) error + ExecTaskF func(string, []string, time.Duration) (*ExecTaskResult, error) +} + +func (d *MockDriver) TaskConfigSchema() (*hclspec.Spec, error) { return d.TaskConfigSchemaF() } +func (d *MockDriver) Fingerprint(ctx context.Context) (<-chan *Fingerprint, error) { + return d.FingerprintF(ctx) +} +func (d *MockDriver) Capabilities() (*Capabilities, error) { return d.CapabilitiesF() } +func (d *MockDriver) RecoverTask(h *TaskHandle) error { return d.RecoverTaskF(h) } +func (d *MockDriver) StartTask(c *TaskConfig) (*TaskHandle, error) { return d.StartTaskF(c) } +func (d *MockDriver) WaitTask(ctx context.Context, id string) (<-chan *ExitResult, error) { + return d.WaitTaskF(ctx, id) +} +func (d *MockDriver) StopTask(taskID string, timeout time.Duration, signal string) error { + return d.StopTaskF(taskID, timeout, signal) +} +func (d *MockDriver) DestroyTask(taskID string, force bool) error { + return d.DestroyTaskF(taskID, force) +} +func (d *MockDriver) InspectTask(taskID string) (*TaskStatus, error) { return d.InspectTaskF(taskID) } +func (d *MockDriver) TaskStats(taskID string) (*TaskStats, error) { return d.TaskStats(taskID) } +func (d *MockDriver) TaskEvents(ctx context.Context) (<-chan *TaskEvent, error) { + return d.TaskEventsF(ctx) +} +func (d *MockDriver) SignalTask(taskID string, signal string) error { + return d.SignalTask(taskID, signal) +} +func (d *MockDriver) ExecTask(taskID string, cmd []string, timeout time.Duration) (*ExecTaskResult, error) { + return d.ExecTaskF(taskID, cmd, timeout) +} diff --git a/plugins/drivers/testing_test.go b/plugins/drivers/testing_test.go new file mode 100644 index 000000000..ac4141f8d --- /dev/null +++ b/plugins/drivers/testing_test.go @@ -0,0 +1,24 @@ +package drivers + +import ( + "testing" + + "github.com/stretchr/testify/require" +) + +var _ DriverPlugin = (*MockDriver)(nil) + +// Very simple test to ensure the test harness works as expected +func TestDriverHarness(t *testing.T) { + handle := &TaskHandle{Config: &TaskConfig{Name: "mock"}} + d := &MockDriver{ + StartTaskF: func(task *TaskConfig) (*TaskHandle, error) { + return handle, nil + }, + } + harness := NewDriverHarness(t, d) + defer harness.Kill() + actual, err := harness.StartTask(&TaskConfig{}) + require.NoError(t, err) + require.Equal(t, handle.Config.Name, actual.Config.Name) +} diff --git a/plugins/drivers/utils.go b/plugins/drivers/utils.go new file mode 100644 index 000000000..7bf248491 --- /dev/null +++ b/plugins/drivers/utils.go @@ -0,0 +1,296 @@ +package drivers + +import ( + "strings" + "time" + + "github.com/golang/protobuf/ptypes" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/plugins/drivers/proto" +) + +var protoTaskStateMap = map[TaskState]proto.TaskState{ + TaskStateUnknown: proto.TaskState_UNKNOWN, + TaskStateRunning: proto.TaskState_RUNNING, + TaskStateExited: proto.TaskState_EXITED, +} + +func healthStateToProto(health HealthState) proto.FingerprintResponse_HealthState { + switch health { + case HealthStateUndetected: + return proto.FingerprintResponse_UNDETECTED + case HealthStateUnhealthy: + return proto.FingerprintResponse_UNHEALTHY + case HealthStateHealthy: + return proto.FingerprintResponse_HEALTHY + } + return proto.FingerprintResponse_UNDETECTED +} + +func healthStateFromProto(pb proto.FingerprintResponse_HealthState) HealthState { + switch pb { + case proto.FingerprintResponse_UNDETECTED: + return HealthStateUndetected + case proto.FingerprintResponse_UNHEALTHY: + return HealthStateUnhealthy + case proto.FingerprintResponse_HEALTHY: + return HealthStateHealthy + } + return HealthStateUndetected +} + +func taskConfigFromProto(pb *proto.TaskConfig) *TaskConfig { + if pb == nil { + return &TaskConfig{} + } + return &TaskConfig{ + ID: pb.Id, + Name: pb.Name, + Env: pb.Env, + rawDriverConfig: pb.MsgpackDriverConfig, + Resources: Resources{}, //TODO + Devices: []DeviceConfig{}, //TODO + Mounts: []MountConfig{}, //TODO + User: pb.User, + AllocDir: pb.AllocDir, + } +} + +func taskConfigToProto(cfg *TaskConfig) *proto.TaskConfig { + if cfg == nil { + return &proto.TaskConfig{} + } + pb := &proto.TaskConfig{ + Id: cfg.ID, + Name: cfg.Name, + Env: cfg.Env, + Resources: &proto.Resources{}, + Mounts: []*proto.Mount{}, + Devices: []*proto.Device{}, + User: cfg.User, + AllocDir: cfg.AllocDir, + MsgpackDriverConfig: cfg.rawDriverConfig, + } + return pb +} + +func taskHandleFromProto(pb *proto.TaskHandle) *TaskHandle { + if pb == nil { + return &TaskHandle{} + } + return &TaskHandle{ + Config: taskConfigFromProto(pb.Config), + State: TaskState(strings.ToLower(pb.State.String())), + driverState: pb.DriverState, + } +} + +func taskHandleToProto(handle *TaskHandle) *proto.TaskHandle { + return &proto.TaskHandle{ + Config: taskConfigToProto(handle.Config), + State: protoTaskStateMap[handle.State], + DriverState: handle.driverState, + } +} + +func exitResultToProto(result *ExitResult) *proto.ExitResult { + return &proto.ExitResult{ + ExitCode: int32(result.ExitCode), + Signal: int32(result.Signal), + OomKilled: result.OOMKilled, + } +} + +func exitResultFromProto(pb *proto.ExitResult) *ExitResult { + return &ExitResult{ + ExitCode: int(pb.ExitCode), + Signal: int(pb.Signal), + OOMKilled: pb.OomKilled, + } +} + +func taskStatusToProto(status *TaskStatus) (*proto.TaskStatus, error) { + started, err := ptypes.TimestampProto(status.StartedAt) + if err != nil { + return nil, err + } + completed, err := ptypes.TimestampProto(status.CompletedAt) + if err != nil { + return nil, err + } + return &proto.TaskStatus{ + Id: status.ID, + Name: status.Name, + StartedAt: started, + CompletedAt: completed, + Result: exitResultToProto(status.ExitResult), + }, nil +} + +func taskStatusFromProto(pb *proto.TaskStatus) (*TaskStatus, error) { + started, err := ptypes.Timestamp(pb.StartedAt) + if err != nil { + return nil, err + } + + completed, err := ptypes.Timestamp(pb.CompletedAt) + if err != nil { + return nil, err + } + + return &TaskStatus{ + ID: pb.Id, + Name: pb.Name, + StartedAt: started, + CompletedAt: completed, + ExitResult: exitResultFromProto(pb.Result), + }, nil +} + +func taskStatsToProto(stats *TaskStats) (*proto.TaskStats, error) { + timestamp, err := ptypes.TimestampProto(time.Unix(stats.Timestamp, 0)) + if err != nil { + return nil, err + } + + pids := map[string]*proto.TaskResourceUsage{} + for pid, ru := range stats.ResourceUsageByPid { + pids[pid] = resourceUsageToProto(ru) + } + + return &proto.TaskStats{ + Id: stats.ID, + Timestamp: timestamp, + AggResourceUsage: resourceUsageToProto(stats.AggResourceUsage), + ResourceUsageByPid: pids, + }, nil +} + +func taskStatsFromProto(pb *proto.TaskStats) (*TaskStats, error) { + timestamp, err := ptypes.Timestamp(pb.Timestamp) + if err != nil { + return nil, err + } + + pids := map[string]*cstructs.ResourceUsage{} + for pid, ru := range pb.ResourceUsageByPid { + pids[pid] = resourceUsageFromProto(ru) + } + + stats := &TaskStats{ + ID: pb.Id, + Timestamp: timestamp.Unix(), + AggResourceUsage: resourceUsageFromProto(pb.AggResourceUsage), + ResourceUsageByPid: pids, + } + + return stats, nil +} + +func resourceUsageToProto(ru *cstructs.ResourceUsage) *proto.TaskResourceUsage { + cpu := &proto.CPUUsage{} + for _, field := range ru.CpuStats.Measured { + switch field { + case "System Mode": + cpu.SystemMode = ru.CpuStats.SystemMode + cpu.MeasuredFields = append(cpu.MeasuredFields, proto.CPUUsage_SYSTEM_MODE) + case "User Mode": + cpu.UserMode = ru.CpuStats.UserMode + cpu.MeasuredFields = append(cpu.MeasuredFields, proto.CPUUsage_USER_MODE) + case "Total Ticks": + cpu.TotalTicks = ru.CpuStats.TotalTicks + cpu.MeasuredFields = append(cpu.MeasuredFields, proto.CPUUsage_TOTAL_TICKS) + case "Throttled Periods": + cpu.ThrottledPeriods = ru.CpuStats.ThrottledPeriods + cpu.MeasuredFields = append(cpu.MeasuredFields, proto.CPUUsage_THROTTLED_PERIODS) + case "Throttled Time": + cpu.ThrottledTime = ru.CpuStats.ThrottledTime + cpu.MeasuredFields = append(cpu.MeasuredFields, proto.CPUUsage_THROTTLED_TIME) + case "Percent": + cpu.Percent = ru.CpuStats.Percent + cpu.MeasuredFields = append(cpu.MeasuredFields, proto.CPUUsage_PERCENT) + } + } + + memory := &proto.MemoryUsage{} + for _, field := range ru.MemoryStats.Measured { + switch field { + case "RSS": + memory.Rss = ru.MemoryStats.RSS + memory.MeasuredFields = append(memory.MeasuredFields, proto.MemoryUsage_RSS) + case "Cache": + memory.Cache = ru.MemoryStats.Cache + memory.MeasuredFields = append(memory.MeasuredFields, proto.MemoryUsage_CACHE) + case "Max Usage": + memory.MaxUsage = ru.MemoryStats.MaxUsage + memory.MeasuredFields = append(memory.MeasuredFields, proto.MemoryUsage_MAX_USAGE) + case "Kernel Usage": + memory.KernelUsage = ru.MemoryStats.KernelUsage + memory.MeasuredFields = append(memory.MeasuredFields, proto.MemoryUsage_KERNEL_USAGE) + case "Kernel Max Usage": + memory.KernelMaxUsage = ru.MemoryStats.KernelMaxUsage + memory.MeasuredFields = append(memory.MeasuredFields, proto.MemoryUsage_KERNEL_MAX_USAGE) + } + } + + return &proto.TaskResourceUsage{ + Cpu: cpu, + Memory: memory, + } +} + +func resourceUsageFromProto(pb *proto.TaskResourceUsage) *cstructs.ResourceUsage { + cpu := cstructs.CpuStats{} + if pb.Cpu != nil { + for _, field := range pb.Cpu.MeasuredFields { + switch field { + case proto.CPUUsage_SYSTEM_MODE: + cpu.SystemMode = pb.Cpu.SystemMode + cpu.Measured = append(cpu.Measured, "System Mode") + case proto.CPUUsage_USER_MODE: + cpu.UserMode = pb.Cpu.UserMode + cpu.Measured = append(cpu.Measured, "User Mode") + case proto.CPUUsage_TOTAL_TICKS: + cpu.TotalTicks = pb.Cpu.TotalTicks + cpu.Measured = append(cpu.Measured, "Total Ticks") + case proto.CPUUsage_THROTTLED_PERIODS: + cpu.ThrottledPeriods = pb.Cpu.ThrottledPeriods + cpu.Measured = append(cpu.Measured, "Throttled Periods") + case proto.CPUUsage_THROTTLED_TIME: + cpu.ThrottledTime = pb.Cpu.ThrottledTime + cpu.Measured = append(cpu.Measured, "Throttled Time") + case proto.CPUUsage_PERCENT: + cpu.Percent = pb.Cpu.Percent + cpu.Measured = append(cpu.Measured, "Percent") + } + } + } + + memory := cstructs.MemoryStats{} + if pb.Memory != nil { + for _, field := range pb.Memory.MeasuredFields { + switch field { + case proto.MemoryUsage_RSS: + memory.RSS = pb.Memory.Rss + memory.Measured = append(memory.Measured, "RSS") + case proto.MemoryUsage_CACHE: + memory.Cache = pb.Memory.Cache + memory.Measured = append(memory.Measured, "Cache") + case proto.MemoryUsage_MAX_USAGE: + memory.MaxUsage = pb.Memory.MaxUsage + memory.Measured = append(memory.Measured, "Max Usage") + case proto.MemoryUsage_KERNEL_USAGE: + memory.KernelUsage = pb.Memory.KernelUsage + memory.Measured = append(memory.Measured, "Kernel Usage") + case proto.MemoryUsage_KERNEL_MAX_USAGE: + memory.KernelMaxUsage = pb.Memory.KernelMaxUsage + memory.Measured = append(memory.Measured, "Kernel Max Usage") + } + } + } + + return &cstructs.ResourceUsage{ + CpuStats: &cpu, + MemoryStats: &memory, + } +} diff --git a/plugins/drivers/utils/utils.go b/plugins/drivers/utils/utils.go new file mode 100644 index 000000000..4ace338ff --- /dev/null +++ b/plugins/drivers/utils/utils.go @@ -0,0 +1,99 @@ +package utils + +import ( + "encoding/json" + "fmt" + "io" + "os" + "os/exec" + + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/client/driver" + "github.com/hashicorp/nomad/client/driver/executor" + dstructs "github.com/hashicorp/nomad/client/driver/structs" + "github.com/hashicorp/nomad/helper/discover" + "github.com/hashicorp/nomad/nomad/structs" +) + +// CgroupsMounted returns true if the cgroups are mounted on a system otherwise +// returns false +func CgroupsMounted(node *structs.Node) bool { + _, ok := node.Attributes["unique.cgroup.mountpoint"] + return ok +} + +// CreateExecutor launches an executor plugin and returns an instance of the +// Executor interface +func CreateExecutor(w io.Writer, level hclog.Level, CMinPort, CMaxPort uint, + executorConfig *dstructs.ExecutorConfig) (executor.Executor, *plugin.Client, error) { + + c, err := json.Marshal(executorConfig) + if err != nil { + return nil, nil, fmt.Errorf("unable to create executor config: %v", err) + } + bin, err := discover.NomadExecutable() + if err != nil { + return nil, nil, fmt.Errorf("unable to find the nomad binary: %v", err) + } + + config := &plugin.ClientConfig{ + Cmd: exec.Command(bin, "executor", string(c)), + } + config.HandshakeConfig = driver.HandshakeConfig + config.Plugins = driver.GetPluginMap(w, level, executorConfig.FSIsolation) + config.MaxPort = CMaxPort + config.MinPort = CMinPort + + // setting the setsid of the plugin process so that it doesn't get signals sent to + // the nomad client. + if config.Cmd != nil { + isolateCommand(config.Cmd) + } + + executorClient := plugin.NewClient(config) + rpcClient, err := executorClient.Client() + if err != nil { + return nil, nil, fmt.Errorf("error creating rpc client for executor plugin: %v", err) + } + + raw, err := rpcClient.Dispense("executor") + if err != nil { + return nil, nil, fmt.Errorf("unable to dispense the executor plugin: %v", err) + } + executorPlugin := raw.(executor.Executor) + return executorPlugin, executorClient, nil +} + +func CreateExecutorWithConfig(config *plugin.ClientConfig, w io.Writer) (executor.Executor, *plugin.Client, error) { + config.HandshakeConfig = driver.HandshakeConfig + + // Setting this to DEBUG since the log level at the executor server process + // is already set, and this effects only the executor client. + config.Plugins = driver.GetPluginMap(w, hclog.Debug, false) + + executorClient := plugin.NewClient(config) + rpcClient, err := executorClient.Client() + if err != nil { + return nil, nil, fmt.Errorf("error creating rpc client for executor plugin: %v", err) + } + + raw, err := rpcClient.Dispense("executor") + if err != nil { + return nil, nil, fmt.Errorf("unable to dispense the executor plugin: %v", err) + } + executorPlugin, ok := raw.(*driver.ExecutorRPC) + if !ok { + return nil, nil, fmt.Errorf("unexpected executor rpc type: %T", raw) + } + return executorPlugin, executorClient, nil +} + +// KillProcess kills a process with the given pid +func KillProcess(pid int) error { + proc, err := os.FindProcess(pid) + if err != nil { + return err + } + return proc.Kill() +} diff --git a/plugins/drivers/utils/utils_unix.go b/plugins/drivers/utils/utils_unix.go new file mode 100644 index 000000000..f720a6e22 --- /dev/null +++ b/plugins/drivers/utils/utils_unix.go @@ -0,0 +1,18 @@ +// +build darwin dragonfly freebsd linux netbsd openbsd solaris + +package utils + +import ( + "os/exec" + "syscall" +) + +// isolateCommand sets the setsid flag in exec.Cmd to true so that the process +// becomes the process leader in a new session and doesn't receive signals that +// are sent to the parent process. +func isolateCommand(cmd *exec.Cmd) { + if cmd.SysProcAttr == nil { + cmd.SysProcAttr = &syscall.SysProcAttr{} + } + cmd.SysProcAttr.Setsid = true +} diff --git a/plugins/drivers/utils/utils_windows.go b/plugins/drivers/utils/utils_windows.go new file mode 100644 index 000000000..a12de4d9e --- /dev/null +++ b/plugins/drivers/utils/utils_windows.go @@ -0,0 +1,8 @@ +package utils + +import ( + "os/exec" +) + +// TODO Figure out if this is needed in Windows +func isolateCommand(cmd *exec.Cmd) {} From 5617f3615b7f56308830c4a0a342a10411bf932c Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 19 Sep 2018 15:56:50 -0400 Subject: [PATCH 074/152] driver/raw_exec: initial raw_exec implementation --- plugins/drivers/raw_exec/driver.go | 515 ++++++++++++++++++++++++ plugins/drivers/raw_exec/driver_test.go | 37 ++ plugins/drivers/utils/eventer.go | 101 +++++ plugins/drivers/utils/eventer_test.go | 72 ++++ 4 files changed, 725 insertions(+) create mode 100644 plugins/drivers/raw_exec/driver.go create mode 100644 plugins/drivers/raw_exec/driver_test.go create mode 100644 plugins/drivers/utils/eventer.go create mode 100644 plugins/drivers/utils/eventer_test.go diff --git a/plugins/drivers/raw_exec/driver.go b/plugins/drivers/raw_exec/driver.go new file mode 100644 index 000000000..dabf6f8cc --- /dev/null +++ b/plugins/drivers/raw_exec/driver.go @@ -0,0 +1,515 @@ +package raw_exec + +import ( + "fmt" + "os" + "path/filepath" + "strconv" + "sync" + "time" + + "github.com/hashicorp/consul-template/signals" + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/client/driver/executor" + dstructs "github.com/hashicorp/nomad/client/driver/structs" + bbase "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/drivers/base" + "github.com/hashicorp/nomad/plugins/drivers/utils" + "github.com/hashicorp/nomad/plugins/shared/hclspec" + "golang.org/x/net/context" +) + +const ( + // pluginName is the name of the plugin + pluginName = "raw_exec" + + // fingerprintPeriod is the interval at which the driver will send fingerprint responses + fingerprintPeriod = 30 * time.Second +) + +var ( + pluginInfo = &bbase.PluginInfoResponse{ + Type: bbase.PluginTypeDriver, + PluginApiVersion: "0.0.1", + PluginVersion: "0.1.0", + Name: pluginName, + } + + configSpec = hclspec.NewObject(map[string]*hclspec.Spec{ + "enabled": hclspec.NewDefault( + hclspec.NewAttr("enabled", "bool", false), + hclspec.NewLiteral("false"), + ), + "no_cgroups": hclspec.NewDefault( + hclspec.NewAttr("no_cgroups", "bool", false), + hclspec.NewLiteral("false"), + ), + }) + + taskConfigSpec = hclspec.NewObject(map[string]*hclspec.Spec{ + "command": hclspec.NewAttr("command", "string", true), + "args": hclspec.NewAttr("command", "list(string)", false), + }) + + capabilities = &base.Capabilities{ + SendSignals: true, + Exec: true, + FSIsolation: base.FSIsolationNone, + } +) + +// The RawExecDriver is a privileged version of the exec driver. It provides no +// resource isolation and just fork/execs. The Exec driver should be preferred +// and this should only be used when explicitly needed. +type RawExecDriver struct { + *utils.Eventer + config *Config + tasks *taskStore + + // fingerprintCh is a channel which other funcs can send fingerprints to + // that will immediatly be sent + fingerprintCh chan *base.Fingerprint + + stopCh chan struct{} + + logger hclog.Logger +} + +type Config struct { + // noCgroups tracks whether we should use a cgroup to manage the process + // tree + noCgroups bool `codec:"no_cgroups"` + + // enabled is set to true to enable the raw_exec driver + enabled bool `codec:"enabled"` +} + +type TaskConfig struct { + Command string `codec:"command"` + Args []string `codec:"args"` +} + +func NewRawExecDriver(logger hclog.Logger) base.DriverPlugin { + stopCh := make(chan struct{}) + return &RawExecDriver{ + Eventer: utils.NewEventer(stopCh), + config: &Config{}, + tasks: newTaskStore(), + fingerprintCh: make(chan *base.Fingerprint), + stopCh: stopCh, + logger: logger.Named(pluginName), + } +} + +func (r *RawExecDriver) PluginInfo() (*bbase.PluginInfoResponse, error) { + return pluginInfo, nil +} + +func (r *RawExecDriver) ConfigSchema() (*hclspec.Spec, error) { + return configSpec, nil +} + +func (r *RawExecDriver) SetConfig(data []byte) error { + var config Config + if err := bbase.MsgPackDecode(data, &config); err != nil { + return err + } + + r.config = &config + go r.fingerprintNow() + return nil +} + +func (r *RawExecDriver) Shutdown(ctx context.Context) error { + close(r.stopCh) + return nil +} + +func (r *RawExecDriver) TaskConfigSchema() (*hclspec.Spec, error) { + return taskConfigSpec, nil +} + +func (r *RawExecDriver) Capabilities() (*base.Capabilities, error) { + return capabilities, nil +} + +func (r *RawExecDriver) Fingerprint(ctx context.Context) (<-chan *base.Fingerprint, error) { + ch := make(chan *base.Fingerprint) + go r.fingerprint(ctx, ch) + return r.fingerprintCh, nil +} + +func (r *RawExecDriver) fingerprint(ctx context.Context, ch chan *base.Fingerprint) { + defer close(r.fingerprintCh) + + ticker := time.NewTimer(0) + for { + select { + case <-ctx.Done(): + return + case <-ticker.C: + ticker.Reset(fingerprintPeriod) + go r.fingerprintNow() + case f := <-r.fingerprintCh: + ch <- f + } + } +} + +func (r *RawExecDriver) fingerprintNow() { + if r.fingerprintCh == nil { + r.logger.Debug("fingerprint channel was nil, skipping fingerprint") + return + } + var health base.HealthState + var desc string + if r.config.enabled { + health = base.HealthStateHealthy + desc = "raw_exec enabled" + } else { + health = base.HealthStateUnhealthy + desc = "raw_exec not enabled" + } + r.fingerprintCh <- &base.Fingerprint{ + Attributes: map[string]string{}, + Health: health, + HealthDescription: desc, + } +} + +func (r *RawExecDriver) RecoverTask(handle *base.TaskHandle) error { + var taskState RawExecTaskState + + err := handle.GetDriverState(&taskState) + if err != nil { + return err + } + + pluginConfig := &plugin.ClientConfig{ + Reattach: taskState.ReattachConfig, + } + + exec, pluginClient, err := utils.CreateExecutorWithConfig(pluginConfig, os.Stderr) + if err != nil { + return err + } + + h := &rawExecTaskHandle{ + exec: exec, + pid: taskState.Pid, + pluginClient: pluginClient, + task: taskState.TaskConfig, + procState: base.TaskStateRunning, + startedAt: taskState.StartedAt, + } + + r.tasks.Set(taskState.TaskConfig.ID, h) + + go h.run() + return nil +} + +func (r *RawExecDriver) StartTask(cfg *base.TaskConfig) (*base.TaskHandle, error) { + if _, ok := r.tasks.Get(cfg.ID); ok { + return nil, fmt.Errorf("task with ID '%s' already started", cfg.ID) + } + + var driverConfig TaskConfig + if err := cfg.DecodeDriverConfig(&driverConfig); err != nil { + return nil, err + } + + handle := base.NewTaskHandle(pluginName) + handle.Config = cfg + + // Get the command to be ran + command := driverConfig.Command + if err := utils.ValidateCommand(command, "args"); err != nil { + return nil, err + } + + pluginLogFile := filepath.Join(cfg.TaskDir().Dir, "executor.out") + executorConfig := &dstructs.ExecutorConfig{ + LogFile: pluginLogFile, + LogLevel: "debug", + } + + exec, pluginClient, err := utils.CreateExecutor(os.Stderr, hclog.Debug, 14000, 14512, executorConfig) + if err != nil { + return nil, err + } + + execCmd := &executor.ExecCommand{ + Cmd: driverConfig.Command, + Args: driverConfig.Args, + Env: cfg.EnvList(), + User: cfg.User, + //TaskKillSignal: os.Interrupt, + BasicProcessCgroup: !r.config.noCgroups, + } + + ps, err := exec.Launch(execCmd) + if err != nil { + pluginClient.Kill() + return nil, err + } + + h := &rawExecTaskHandle{ + exec: exec, + pid: ps.Pid, + pluginClient: pluginClient, + task: cfg, + procState: base.TaskStateRunning, + startedAt: time.Now(), + } + + r.tasks.Set(cfg.ID, h) + + handle.SetDriverState(RawExecTaskState{ + ReattachConfig: pluginClient.ReattachConfig(), + Pid: ps.Pid, + TaskConfig: cfg, + StartedAt: h.startedAt, + }) + + go h.run() + return handle, nil +} + +func (r *RawExecDriver) WaitTask(ctx context.Context, taskID string) (<-chan *base.ExitResult, error) { + ch := make(chan *base.ExitResult) + handle, ok := r.tasks.Get(taskID) + if !ok { + return nil, base.ErrTaskNotFound + } + go r.handleWait(ctx, handle, ch) + + return ch, nil +} + +func (r *RawExecDriver) handleWait(ctx context.Context, handle *rawExecTaskHandle, ch chan *base.ExitResult) { + defer close(ch) + var result *base.ExitResult + ps, err := handle.exec.Wait() + if err != nil { + result = &base.ExitResult{ + Err: fmt.Errorf("executor: error waiting on process: %v", err), + } + } else { + result = &base.ExitResult{ + ExitCode: ps.ExitCode, + Signal: ps.Signal, + } + } + + select { + case <-ctx.Done(): + return + case ch <- result: + } +} + +func (r *RawExecDriver) StopTask(taskID string, timeout time.Duration, signal string) error { + /* + handle, ok := r.tasks.Get(taskID) + if !ok { + return base.ErrTaskNotFound + } + + //TODO executor only supports shutting down with the initial signal provided + if err := handle.exec.Shutdown(); err != nil { + if handle.pluginClient.Exited() { + return nil + } + return fmt.Errorf("executor Shutdown failed: %v", err) + } + + select { + case <-d.WaitTask(taskID): + return nil + case <-time.After(timeout): + if handle.pluginClient.Exited() { + return nil + } + if err := handle.exec.Exit(); err != nil { + return fmt.Errorf("executor Exit failed: %v", err) + } + return nil + }*/ + panic("not implemented") +} + +func (r *RawExecDriver) DestroyTask(taskID string, force bool) error { + handle, ok := r.tasks.Get(taskID) + if !ok { + return base.ErrTaskNotFound + } + + if !handle.pluginClient.Exited() { + if err := handle.exec.Destroy(); err != nil { + handle.logger.Error("destroying executor failed", "err", err) + } + } + + handle.pluginClient.Kill() + r.tasks.Delete(taskID) + return nil +} + +func (r *RawExecDriver) InspectTask(taskID string) (*base.TaskStatus, error) { + handle, ok := r.tasks.Get(taskID) + if !ok { + return nil, base.ErrTaskNotFound + } + + status := &base.TaskStatus{ + ID: handle.task.ID, + Name: handle.task.Name, + State: handle.procState, + SizeOnDiskMB: 0, + StartedAt: handle.startedAt, + CompletedAt: handle.completedAt, + ExitResult: handle.exitResult, + DriverAttributes: map[string]string{ + "pid": strconv.Itoa(handle.pid), + }, + NetworkOverride: &base.NetworkOverride{}, + } + + return status, nil +} + +func (r *RawExecDriver) TaskStats(taskID string) (*base.TaskStats, error) { + handle, ok := r.tasks.Get(taskID) + if !ok { + return nil, base.ErrTaskNotFound + } + + stats, err := handle.exec.Stats() + if err != nil { + return nil, err + } + + return &base.TaskStats{ + ID: handle.task.ID, + Timestamp: stats.Timestamp, + AggResourceUsage: stats.ResourceUsage, + ResourceUsageByPid: stats.Pids, + }, nil +} + +func (r *RawExecDriver) SignalTask(taskID string, signal string) error { + handle, ok := r.tasks.Get(taskID) + if !ok { + return base.ErrTaskNotFound + } + + sig := os.Interrupt + if s, ok := signals.SignalLookup[signal]; ok { + sig = s + } + return handle.exec.Signal(sig) +} + +func (r *RawExecDriver) ExecTask(taskID string, cmd []string, timeout time.Duration) (*base.ExecTaskResult, error) { + if len(cmd) == 0 { + return nil, fmt.Errorf("error cmd must have atleast one value") + } + handle, ok := r.tasks.Get(taskID) + if !ok { + return nil, base.ErrTaskNotFound + } + + args := []string{} + if len(cmd) > 1 { + args = cmd[1:] + } + + out, exitCode, err := handle.exec.Exec(time.Now().Add(timeout), cmd[0], args) + if err != nil { + return nil, err + } + + return &base.ExecTaskResult{ + Stdout: out, + ExitResult: &base.ExitResult{ + ExitCode: exitCode, + }, + }, nil +} + +type taskStore struct { + store map[string]*rawExecTaskHandle + lock sync.RWMutex +} + +func newTaskStore() *taskStore { + return &taskStore{store: map[string]*rawExecTaskHandle{}} +} + +func (ts *taskStore) Set(id string, handle *rawExecTaskHandle) { + ts.lock.Lock() + defer ts.lock.Unlock() + ts.store[id] = handle +} + +func (ts *taskStore) Get(id string) (*rawExecTaskHandle, bool) { + ts.lock.RLock() + defer ts.lock.RUnlock() + t, ok := ts.store[id] + return t, ok +} + +func (ts *taskStore) Delete(id string) { + ts.lock.Lock() + defer ts.lock.Unlock() + delete(ts.store, id) +} + +func (ts *taskStore) Range(f func(id string, handle *rawExecTaskHandle) bool) { + ts.lock.RLock() + defer ts.lock.RUnlock() + for k, v := range ts.store { + if f(k, v) { + break + } + } +} + +type RawExecTaskState struct { + ReattachConfig *plugin.ReattachConfig + TaskConfig *base.TaskConfig + Pid int + StartedAt time.Time +} + +type rawExecTaskHandle struct { + exec executor.Executor + pid int + pluginClient *plugin.Client + task *base.TaskConfig + procState base.TaskState + startedAt time.Time + completedAt time.Time + exitResult *base.ExitResult + logger hclog.Logger +} + +func (h *rawExecTaskHandle) run() { + if h.exitResult == nil { + h.exitResult = &base.ExitResult{} + } + + ps, err := h.exec.Wait() + if err != nil { + h.exitResult.Err = err + h.procState = base.TaskStateUnknown + h.completedAt = time.Now() + return + } + h.procState = base.TaskStateExited + h.exitResult.ExitCode = ps.ExitCode + h.exitResult.Signal = ps.Signal + h.completedAt = ps.Time +} diff --git a/plugins/drivers/raw_exec/driver_test.go b/plugins/drivers/raw_exec/driver_test.go new file mode 100644 index 000000000..7d1a026a4 --- /dev/null +++ b/plugins/drivers/raw_exec/driver_test.go @@ -0,0 +1,37 @@ +package raw_exec + +import ( + "context" + "testing" + + "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/helper/uuid" + "github.com/hashicorp/nomad/plugins/drivers/base" + "github.com/stretchr/testify/require" +) + +func TestDriverStartTask(t *testing.T) { + t.Parallel() + require := require.New(t) + + d := NewRawExecDriver(testlog.HCLogger(t)) + harness := base.NewDriverHarness(t, d) + task := &base.TaskConfig{ + ID: uuid.Generate(), + Name: "test", + } + task.EncodeDriverConfig(&TaskConfig{ + Command: "go", + Args: []string{"version"}, + }) + cleanup := harness.MkAllocDir(task) + defer cleanup() + + handle, err := harness.StartTask(task) + require.NoError(err) + + ch, err := harness.WaitTask(context.Background(), handle.Config.ID) + require.NoError(err) + result := <-ch + require.Zero(result.ExitCode) +} diff --git a/plugins/drivers/utils/eventer.go b/plugins/drivers/utils/eventer.go new file mode 100644 index 000000000..6de4209e5 --- /dev/null +++ b/plugins/drivers/utils/eventer.go @@ -0,0 +1,101 @@ +package utils + +import ( + "sync" + "time" + + "github.com/hashicorp/nomad/plugins/drivers/base" + "golang.org/x/net/context" +) + +var ( + //DefaultSendEventTimeout is the timeout used when publishing events to consumers + DefaultSendEventTimeout = 2 * time.Second +) + +// Eventer is a utility to control broadcast of TaskEvents to multiple consumers. +// It also implements the TaskStats func in the DriverPlugin interface so that +// it can be embedded in a implementing driver struct. +type Eventer struct { + sync.RWMutex + + // events is a channel were events to be broadcasted are sent + events chan *base.TaskEvent + + // streamers is a slice of consumers to broadcast events to + // access is gaurded by RWMutex + streamers []*eventStreamer + + // stop chan to allow control of event loop shutdown + stop chan struct{} +} + +// NewEventer returns an Eventer with a running event loop that can be stopped +// by closing the given stop channel +func NewEventer(stop chan struct{}) *Eventer { + e := &Eventer{ + events: make(chan *base.TaskEvent), + stop: stop, + } + go e.eventLoop() + return e +} + +// eventLoop is the main logic which pulls events from the channel and broadcasts +// them to all consumers +func (e *Eventer) eventLoop() { + for { + select { + case <-e.stop: + for _, stream := range e.streamers { + close(stream.ch) + } + return + case event := <-e.events: + e.RLock() + for _, stream := range e.streamers { + stream.send(event) + } + e.RUnlock() + } + } +} + +type eventStreamer struct { + timeout time.Duration + ctx context.Context + ch chan *base.TaskEvent +} + +func (s *eventStreamer) send(event *base.TaskEvent) { + select { + case <-time.After(s.timeout): + case <-s.ctx.Done(): + case s.ch <- event: + } +} + +func (e *Eventer) newStream(ctx context.Context) <-chan *base.TaskEvent { + e.Lock() + defer e.Unlock() + + stream := &eventStreamer{ + ch: make(chan *base.TaskEvent), + ctx: ctx, + timeout: DefaultSendEventTimeout, + } + e.streamers = append(e.streamers, stream) + + return stream.ch +} + +// TaskEvents is an implementation of the DriverPlugin.TaskEvents function +func (e *Eventer) TaskEvents(ctx context.Context) (<-chan *base.TaskEvent, error) { + stream := e.newStream(ctx) + return stream, nil +} + +// EmitEvent can be used to broadcast a new event +func (e *Eventer) EmitEvent(event *base.TaskEvent) { + e.events <- event +} diff --git a/plugins/drivers/utils/eventer_test.go b/plugins/drivers/utils/eventer_test.go new file mode 100644 index 000000000..93ab7f301 --- /dev/null +++ b/plugins/drivers/utils/eventer_test.go @@ -0,0 +1,72 @@ +package utils + +import ( + "context" + "sync" + "testing" + "time" + + "github.com/hashicorp/nomad/plugins/drivers/base" + "github.com/stretchr/testify/require" +) + +func TestEventer(t *testing.T) { + t.Parallel() + require := require.New(t) + + stop := make(chan struct{}) + e := NewEventer(stop) + + events := []*base.TaskEvent{ + { + TaskID: "a", + Timestamp: time.Now(), + }, + { + TaskID: "b", + Timestamp: time.Now(), + }, + { + TaskID: "c", + Timestamp: time.Now(), + }, + } + + consumer1, err := e.TaskEvents(context.Background()) + require.NoError(err) + consumer2, err := e.TaskEvents(context.Background()) + require.NoError(err) + + var buffer1, buffer2 []*base.TaskEvent + var wg sync.WaitGroup + wg.Add(2) + go func() { + defer wg.Done() + for { + event, ok := <-consumer1 + if !ok { + return + } + buffer1 = append(buffer1, event) + } + }() + go func() { + defer wg.Done() + for { + event, ok := <-consumer2 + if !ok { + return + } + buffer2 = append(buffer2, event) + } + }() + + for _, event := range events { + e.EmitEvent(event) + } + + close(stop) + wg.Wait() + require.Exactly(events, buffer1) + require.Exactly(events, buffer2) +} From d9628ff39415bbe80e4bb9ef3c39924d378bbe5d Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Tue, 25 Sep 2018 14:12:58 -0400 Subject: [PATCH 075/152] driver/raw_exec: more tests and bug fixes added wrapper struct for plugin.ReattachConfig to better handle serialization --- client/driver/executor/executor.go | 18 +- plugins/drivers/raw_exec/driver.go | 89 ++++++---- plugins/drivers/raw_exec/driver_test.go | 160 +++++++++++++++++- plugins/drivers/task_handle.go | 6 +- plugins/drivers/utils.go | 15 +- .../drivers/utils/plugin_reattach_config.go | 64 +++++++ 6 files changed, 305 insertions(+), 47 deletions(-) create mode 100644 plugins/drivers/utils/plugin_reattach_config.go diff --git a/client/driver/executor/executor.go b/client/driver/executor/executor.go index 62c3b71af..87cdfeb70 100644 --- a/client/driver/executor/executor.go +++ b/client/driver/executor/executor.go @@ -404,6 +404,7 @@ var ( // Exit cleans up the alloc directory, destroys resource container and kills the // user process func (e *UniversalExecutor) Shutdown(signal string, grace time.Duration) error { + e.logger.Info("shutdown requested", "signal", signal, "grace_period_ms", grace.Round(time.Millisecond)) var merr multierror.Error // If the executor did not launch a process, return. @@ -413,12 +414,15 @@ func (e *UniversalExecutor) Shutdown(signal string, grace time.Duration) error { // If there is no process we can't shutdown if e.childCmd.Process == nil { + e.logger.Warn("failed to shutdown", "error", "no process found") return fmt.Errorf("executor failed to shutdown error: no process found") } proc, err := os.FindProcess(e.childCmd.Process.Pid) if err != nil { - return fmt.Errorf("executor failed to find process: %v", err) + err = fmt.Errorf("executor failed to find process: %v", err) + e.logger.Warn("failed to shutdown", "error", err) + return err } // If grace is 0 then skip shutdown logic @@ -430,10 +434,13 @@ func (e *UniversalExecutor) Shutdown(signal string, grace time.Duration) error { sig, ok := signals.SignalLookup[signal] if !ok { - return fmt.Errorf("error unknown signal given for shutdown: %s", signal) + err = fmt.Errorf("error unknown signal given for shutdown: %s", signal) + e.logger.Warn("failed to shutdown", "error", err) + return err } if err := e.shutdownProcess(sig, proc); err != nil { + e.logger.Warn("failed to shutdown", "error", err) return err } @@ -460,7 +467,12 @@ func (e *UniversalExecutor) Shutdown(signal string, grace time.Duration) error { } } - return merr.ErrorOrNil() + if err := merr.ErrorOrNil(); err != nil { + e.logger.Warn("failed to shutdown", "error", err) + return err + } + + return nil } // lookupBin looks for path to the binary to run by looking for the binary in diff --git a/plugins/drivers/raw_exec/driver.go b/plugins/drivers/raw_exec/driver.go index dabf6f8cc..922c9d43f 100644 --- a/plugins/drivers/raw_exec/driver.go +++ b/plugins/drivers/raw_exec/driver.go @@ -183,15 +183,23 @@ func (r *RawExecDriver) RecoverTask(handle *base.TaskHandle) error { err := handle.GetDriverState(&taskState) if err != nil { + r.logger.Error("failed to recover task", "error", err, "task_id", handle.Config.ID) + return err + } + + plugRC, err := utils.ReattachConfigToGoPlugin(taskState.ReattachConfig) + if err != nil { + r.logger.Error("failed to recover task", "error", err, "task_id", handle.Config.ID) return err } pluginConfig := &plugin.ClientConfig{ - Reattach: taskState.ReattachConfig, + Reattach: plugRC, } exec, pluginClient, err := utils.CreateExecutorWithConfig(pluginConfig, os.Stderr) if err != nil { + r.logger.Error("failed to recover task", "error", err, "task_id", handle.Config.ID) return err } @@ -202,6 +210,7 @@ func (r *RawExecDriver) RecoverTask(handle *base.TaskHandle) error { task: taskState.TaskConfig, procState: base.TaskStateRunning, startedAt: taskState.StartedAt, + exitResult: &base.ExitResult{}, } r.tasks.Set(taskState.TaskConfig.ID, h) @@ -247,6 +256,7 @@ func (r *RawExecDriver) StartTask(cfg *base.TaskConfig) (*base.TaskHandle, error User: cfg.User, //TaskKillSignal: os.Interrupt, BasicProcessCgroup: !r.config.noCgroups, + TaskDir: cfg.TaskDir().Dir, } ps, err := exec.Launch(execCmd) @@ -261,17 +271,25 @@ func (r *RawExecDriver) StartTask(cfg *base.TaskConfig) (*base.TaskHandle, error pluginClient: pluginClient, task: cfg, procState: base.TaskStateRunning, - startedAt: time.Now(), + startedAt: time.Now().Round(time.Millisecond), + logger: r.logger, } r.tasks.Set(cfg.ID, h) - handle.SetDriverState(RawExecTaskState{ - ReattachConfig: pluginClient.ReattachConfig(), + driverState := RawExecTaskState{ + ReattachConfig: utils.ReattachConfigFromGoPlugin(pluginClient.ReattachConfig()), Pid: ps.Pid, TaskConfig: cfg, StartedAt: h.startedAt, - }) + } + + if err := handle.SetDriverState(&driverState); err != nil { + r.logger.Error("failed to start task, error setting driver state", "error", err) + exec.Shutdown("", 0) + pluginClient.Kill() + return nil, err + } go h.run() return handle, nil @@ -311,33 +329,19 @@ func (r *RawExecDriver) handleWait(ctx context.Context, handle *rawExecTaskHandl } func (r *RawExecDriver) StopTask(taskID string, timeout time.Duration, signal string) error { - /* - handle, ok := r.tasks.Get(taskID) - if !ok { - return base.ErrTaskNotFound + handle, ok := r.tasks.Get(taskID) + if !ok { + return base.ErrTaskNotFound + } + + if err := handle.exec.Shutdown(signal, timeout); err != nil { + if handle.pluginClient.Exited() { + return nil } + return fmt.Errorf("executor Shutdown failed: %v", err) + } - //TODO executor only supports shutting down with the initial signal provided - if err := handle.exec.Shutdown(); err != nil { - if handle.pluginClient.Exited() { - return nil - } - return fmt.Errorf("executor Shutdown failed: %v", err) - } - - select { - case <-d.WaitTask(taskID): - return nil - case <-time.After(timeout): - if handle.pluginClient.Exited() { - return nil - } - if err := handle.exec.Exit(); err != nil { - return fmt.Errorf("executor Exit failed: %v", err) - } - return nil - }*/ - panic("not implemented") + return nil } func (r *RawExecDriver) DestroyTask(taskID string, force bool) error { @@ -346,13 +350,20 @@ func (r *RawExecDriver) DestroyTask(taskID string, force bool) error { return base.ErrTaskNotFound } - if !handle.pluginClient.Exited() { - if err := handle.exec.Destroy(); err != nil { - handle.logger.Error("destroying executor failed", "err", err) - } + if handle.IsRunning() && !force { + return fmt.Errorf("cannot destroy running task") + } + + if !handle.pluginClient.Exited() { + if handle.IsRunning() { + if err := handle.exec.Shutdown("", 0); err != nil { + handle.logger.Error("destroying executor failed", "err", err) + } + } + + handle.pluginClient.Kill() } - handle.pluginClient.Kill() r.tasks.Delete(taskID) return nil } @@ -478,7 +489,7 @@ func (ts *taskStore) Range(f func(id string, handle *rawExecTaskHandle) bool) { } type RawExecTaskState struct { - ReattachConfig *plugin.ReattachConfig + ReattachConfig *utils.ReattachConfig TaskConfig *base.TaskConfig Pid int StartedAt time.Time @@ -496,6 +507,10 @@ type rawExecTaskHandle struct { logger hclog.Logger } +func (h *rawExecTaskHandle) IsRunning() bool { + return h.procState == base.TaskStateRunning +} + func (h *rawExecTaskHandle) run() { if h.exitResult == nil { h.exitResult = &base.ExitResult{} @@ -512,4 +527,6 @@ func (h *rawExecTaskHandle) run() { h.exitResult.ExitCode = ps.ExitCode h.exitResult.Signal = ps.Signal h.completedAt = ps.Time + + //h.pluginClient.Kill() } diff --git a/plugins/drivers/raw_exec/driver_test.go b/plugins/drivers/raw_exec/driver_test.go index 7d1a026a4..901982196 100644 --- a/plugins/drivers/raw_exec/driver_test.go +++ b/plugins/drivers/raw_exec/driver_test.go @@ -2,7 +2,11 @@ package raw_exec import ( "context" + "io/ioutil" + "path/filepath" + "sync" "testing" + "time" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/helper/uuid" @@ -10,7 +14,7 @@ import ( "github.com/stretchr/testify/require" ) -func TestDriverStartTask(t *testing.T) { +func TestRawExecDriver_StartWait(t *testing.T) { t.Parallel() require := require.New(t) @@ -34,4 +38,158 @@ func TestDriverStartTask(t *testing.T) { require.NoError(err) result := <-ch require.Zero(result.ExitCode) + require.NoError(harness.DestroyTask(task.ID, true)) +} + +func TestRawExecDriver_StartWaitStop(t *testing.T) { + t.Parallel() + require := require.New(t) + + d := NewRawExecDriver(testlog.HCLogger(t)) + harness := base.NewDriverHarness(t, d) + task := &base.TaskConfig{ + ID: uuid.Generate(), + Name: "test", + } + task.EncodeDriverConfig(&TaskConfig{ + Command: "/bin/bash", + Args: []string{"test.sh"}, + }) + cleanup := harness.MkAllocDir(task) + defer cleanup() + + testFile := filepath.Join(task.TaskDir().Dir, "test.sh") + testData := []byte(` +at_term() { + echo 'Terminated.' + exit 3 +} +trap at_term USR1 +while true; do + sleep 1 +done + `) + require.NoError(ioutil.WriteFile(testFile, testData, 0777)) + + handle, err := harness.StartTask(task) + require.NoError(err) + + ch, err := harness.WaitTask(context.Background(), handle.Config.ID) + require.NoError(err) + + var waitDone bool + var wg sync.WaitGroup + wg.Add(1) + go func() { + defer wg.Done() + result := <-ch + require.Equal(3, result.ExitCode) + waitDone = true + }() + + time.Sleep(100 * time.Millisecond) + + var stopDone bool + wg.Add(1) + go func() { + defer wg.Done() + err := harness.StopTask(task.ID, 1*time.Second, "SIGUSR1") + require.NoError(err) + stopDone = true + }() + + waitCh := make(chan struct{}) + go func() { + defer close(waitCh) + wg.Wait() + }() + + time.Sleep(100 * time.Millisecond) + + status, err := harness.InspectTask(task.ID) + require.NoError(err) + require.Equal(base.TaskStateRunning, status.State) + require.False(waitDone) + require.False(stopDone) + + select { + case <-waitCh: + status, err := harness.InspectTask(task.ID) + require.NoError(err) + require.Equal(base.TaskStateExited, status.State) + case <-time.After(1 * time.Second): + require.Fail("timeout waiting for task to shutdown") + } + + require.NoError(harness.DestroyTask(task.ID, true)) +} + +func TestRawExecDriver_StartWaitRecoverWaitStop(t *testing.T) { + t.Parallel() + require := require.New(t) + + d := NewRawExecDriver(testlog.HCLogger(t)) + harness := base.NewDriverHarness(t, d) + task := &base.TaskConfig{ + ID: uuid.Generate(), + Name: "test", + } + task.EncodeDriverConfig(&TaskConfig{ + Command: "sleep", + Args: []string{"1000"}, + }) + cleanup := harness.MkAllocDir(task) + defer cleanup() + + handle, err := harness.StartTask(task) + require.NoError(err) + + ch, err := harness.WaitTask(context.Background(), task.ID) + require.NoError(err) + + var waitDone bool + var wg sync.WaitGroup + wg.Add(1) + go func() { + defer wg.Done() + result := <-ch + require.Error(result.Err) + waitDone = true + }() + + originalStatus, err := d.InspectTask(task.ID) + require.NoError(err) + + d.(*RawExecDriver).tasks.Delete(task.ID) + + wg.Wait() + require.True(waitDone) + _, err = d.InspectTask(task.ID) + require.Equal(base.ErrTaskNotFound, err) + + err = d.RecoverTask(handle) + require.NoError(err) + + status, err := d.InspectTask(task.ID) + require.NoError(err) + require.Exactly(originalStatus, status) + + ch, err = harness.WaitTask(context.Background(), task.ID) + require.NoError(err) + + wg.Add(1) + waitDone = false + go func() { + defer wg.Done() + result := <-ch + require.NoError(result.Err) + require.Equal(9, result.Signal) + waitDone = true + }() + + require.NoError(d.StopTask(task.ID, 0, "")) + require.NoError(d.DestroyTask(task.ID, false)) + wg.Wait() + require.True(waitDone) + } diff --git a/plugins/drivers/task_handle.go b/plugins/drivers/task_handle.go index 9cf030c87..d6a7981a1 100644 --- a/plugins/drivers/task_handle.go +++ b/plugins/drivers/task_handle.go @@ -1,7 +1,7 @@ package drivers import ( - "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/base" "github.com/ugorji/go/codec" ) @@ -21,10 +21,10 @@ func NewTaskHandle(driver string) *TaskHandle { func (h *TaskHandle) SetDriverState(v interface{}) error { h.driverState = []byte{} - return codec.NewEncoderBytes(&h.driverState, structs.MsgpackHandle).Encode(v) + return codec.NewEncoderBytes(&h.driverState, base.MsgpackHandle).Encode(v) } func (h *TaskHandle) GetDriverState(v interface{}) error { - return codec.NewDecoderBytes(h.driverState, structs.MsgpackHandle).Decode(v) + return codec.NewDecoderBytes(h.driverState, base.MsgpackHandle).Decode(v) } diff --git a/plugins/drivers/utils.go b/plugins/drivers/utils.go index 7bf248491..41605aede 100644 --- a/plugins/drivers/utils.go +++ b/plugins/drivers/utils.go @@ -1,7 +1,6 @@ package drivers import ( - "strings" "time" "github.com/golang/protobuf/ptypes" @@ -9,12 +8,18 @@ import ( "github.com/hashicorp/nomad/plugins/drivers/proto" ) -var protoTaskStateMap = map[TaskState]proto.TaskState{ +var taskStateToProtoMap = map[TaskState]proto.TaskState{ TaskStateUnknown: proto.TaskState_UNKNOWN, TaskStateRunning: proto.TaskState_RUNNING, TaskStateExited: proto.TaskState_EXITED, } +var taskStateFromProtoMap = map[proto.TaskState]TaskState{ + proto.TaskState_UNKNOWN: TaskStateUnknown, + proto.TaskState_RUNNING: TaskStateRunning, + proto.TaskState_EXITED: TaskStateExited, +} + func healthStateToProto(health HealthState) proto.FingerprintResponse_HealthState { switch health { case HealthStateUndetected: @@ -80,7 +85,7 @@ func taskHandleFromProto(pb *proto.TaskHandle) *TaskHandle { } return &TaskHandle{ Config: taskConfigFromProto(pb.Config), - State: TaskState(strings.ToLower(pb.State.String())), + State: taskStateFromProtoMap[pb.State], driverState: pb.DriverState, } } @@ -88,7 +93,7 @@ func taskHandleFromProto(pb *proto.TaskHandle) *TaskHandle { func taskHandleToProto(handle *TaskHandle) *proto.TaskHandle { return &proto.TaskHandle{ Config: taskConfigToProto(handle.Config), - State: protoTaskStateMap[handle.State], + State: taskStateToProtoMap[handle.State], DriverState: handle.driverState, } } @@ -121,6 +126,7 @@ func taskStatusToProto(status *TaskStatus) (*proto.TaskStatus, error) { return &proto.TaskStatus{ Id: status.ID, Name: status.Name, + State: taskStateToProtoMap[status.State], StartedAt: started, CompletedAt: completed, Result: exitResultToProto(status.ExitResult), @@ -141,6 +147,7 @@ func taskStatusFromProto(pb *proto.TaskStatus) (*TaskStatus, error) { return &TaskStatus{ ID: pb.Id, Name: pb.Name, + State: taskStateFromProtoMap[pb.State], StartedAt: started, CompletedAt: completed, ExitResult: exitResultFromProto(pb.Result), diff --git a/plugins/drivers/utils/plugin_reattach_config.go b/plugins/drivers/utils/plugin_reattach_config.go new file mode 100644 index 000000000..3d5b80d33 --- /dev/null +++ b/plugins/drivers/utils/plugin_reattach_config.go @@ -0,0 +1,64 @@ +package utils + +import ( + "fmt" + "net" + + plugin "github.com/hashicorp/go-plugin" +) + +// ReattachConfig is a wrapper around plugin.ReattachConfig to better support +// serialization +type ReattachConfig struct { + Protocol string + Network string + Addr string + Pid int +} + +// ReattachConfigToGoPlugin converts a ReattachConfig wrapper struct into a go +// plugin ReattacConfig struct +func ReattachConfigToGoPlugin(rc *ReattachConfig) (*plugin.ReattachConfig, error) { + plug := &plugin.ReattachConfig{ + Protocol: plugin.Protocol(rc.Protocol), + Pid: rc.Pid, + } + + switch rc.Network { + case "tcp", "tcp4", "tcp6": + addr, err := net.ResolveTCPAddr(rc.Network, rc.Addr) + if err != nil { + return nil, err + } + plug.Addr = addr + case "udp", "udp4", "udp6": + addr, err := net.ResolveUDPAddr(rc.Network, rc.Addr) + if err != nil { + return nil, err + } + plug.Addr = addr + case "unix", "unixgram", "unixpacket": + addr, err := net.ResolveUnixAddr(rc.Network, rc.Addr) + if err != nil { + return nil, err + } + plug.Addr = addr + default: + return nil, fmt.Errorf("unknown network: %s", rc.Network) + } + + return plug, nil +} + +// ReattachConfigFromGoPlugin converts a go plugin ReattachConfig into a +// ReattachConfig wrapper struct +func ReattachConfigFromGoPlugin(plug *plugin.ReattachConfig) *ReattachConfig { + rc := &ReattachConfig{ + Protocol: string(plug.Protocol), + Network: plug.Addr.Network(), + Addr: plug.Addr.String(), + Pid: plug.Pid, + } + + return rc +} From 3c17f50b292bed0c88a0407d23dafa011a1f8fb4 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Tue, 25 Sep 2018 15:45:14 -0400 Subject: [PATCH 076/152] lint: remove unused code and fix spelling --- plugins/drivers/raw_exec/driver.go | 8 +------- 1 file changed, 1 insertion(+), 7 deletions(-) diff --git a/plugins/drivers/raw_exec/driver.go b/plugins/drivers/raw_exec/driver.go index 922c9d43f..32e25fbdd 100644 --- a/plugins/drivers/raw_exec/driver.go +++ b/plugins/drivers/raw_exec/driver.go @@ -68,7 +68,7 @@ type RawExecDriver struct { tasks *taskStore // fingerprintCh is a channel which other funcs can send fingerprints to - // that will immediatly be sent + // that will immediately be sent fingerprintCh chan *base.Fingerprint stopCh chan struct{} @@ -232,12 +232,6 @@ func (r *RawExecDriver) StartTask(cfg *base.TaskConfig) (*base.TaskHandle, error handle := base.NewTaskHandle(pluginName) handle.Config = cfg - // Get the command to be ran - command := driverConfig.Command - if err := utils.ValidateCommand(command, "args"); err != nil { - return nil, err - } - pluginLogFile := filepath.Join(cfg.TaskDir().Dir, "executor.out") executorConfig := &dstructs.ExecutorConfig{ LogFile: pluginLogFile, From 8644e8508cd39e13b9e055ce5f1f8aef331bf737 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 26 Sep 2018 01:12:58 -0400 Subject: [PATCH 077/152] driver/raw_exec: export driver config fields so they are encoded --- plugins/drivers/raw_exec/driver.go | 20 +++++++++++--------- 1 file changed, 11 insertions(+), 9 deletions(-) diff --git a/plugins/drivers/raw_exec/driver.go b/plugins/drivers/raw_exec/driver.go index 32e25fbdd..06ca75b76 100644 --- a/plugins/drivers/raw_exec/driver.go +++ b/plugins/drivers/raw_exec/driver.go @@ -77,12 +77,12 @@ type RawExecDriver struct { } type Config struct { - // noCgroups tracks whether we should use a cgroup to manage the process + // NoCgroups tracks whether we should use a cgroup to manage the process // tree - noCgroups bool `codec:"no_cgroups"` + NoCgroups bool `codec:"no_cgroups"` - // enabled is set to true to enable the raw_exec driver - enabled bool `codec:"enabled"` + // Enabled is set to true to enable the raw_exec driver + Enabled bool `codec:"enabled"` } type TaskConfig struct { @@ -137,7 +137,7 @@ func (r *RawExecDriver) Capabilities() (*base.Capabilities, error) { func (r *RawExecDriver) Fingerprint(ctx context.Context) (<-chan *base.Fingerprint, error) { ch := make(chan *base.Fingerprint) go r.fingerprint(ctx, ch) - return r.fingerprintCh, nil + return ch, nil } func (r *RawExecDriver) fingerprint(ctx context.Context, ch chan *base.Fingerprint) { @@ -164,12 +164,14 @@ func (r *RawExecDriver) fingerprintNow() { } var health base.HealthState var desc string - if r.config.enabled { + attrs := map[string]string{} + if r.config.Enabled { health = base.HealthStateHealthy desc = "raw_exec enabled" + attrs["driver.raw_exec"] = "1" } else { - health = base.HealthStateUnhealthy - desc = "raw_exec not enabled" + health = base.HealthStateUndetected + desc = "raw_exec disabled" } r.fingerprintCh <- &base.Fingerprint{ Attributes: map[string]string{}, @@ -249,7 +251,7 @@ func (r *RawExecDriver) StartTask(cfg *base.TaskConfig) (*base.TaskHandle, error Env: cfg.EnvList(), User: cfg.User, //TaskKillSignal: os.Interrupt, - BasicProcessCgroup: !r.config.noCgroups, + BasicProcessCgroup: !r.config.NoCgroups, TaskDir: cfg.TaskDir().Dir, } From 0e3f85222aff0eddf421e19ba6d3e6bf2c7890b7 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 26 Sep 2018 01:18:03 -0400 Subject: [PATCH 078/152] driver/raw_exec: port existing raw_exec tests and add some testing utilities --- client/driver/executor/executor.go | 5 +- helper/testtask/testtask.go | 10 + plugins/drivers/client.go | 2 +- plugins/drivers/driver.go | 4 + plugins/drivers/plugin.go | 7 +- plugins/drivers/raw_exec/driver_test.go | 322 +++++++++++++++++++++--- plugins/drivers/testing.go | 100 +++++++- 7 files changed, 400 insertions(+), 50 deletions(-) diff --git a/client/driver/executor/executor.go b/client/driver/executor/executor.go index 87cdfeb70..455ed109f 100644 --- a/client/driver/executor/executor.go +++ b/client/driver/executor/executor.go @@ -358,9 +358,10 @@ func (e *UniversalExecutor) UpdateResources(resources *Resources) error { func (e *UniversalExecutor) wait() { defer close(e.processExited) + pid := e.childCmd.Process.Pid err := e.childCmd.Wait() if err == nil { - e.exitState = &ProcessState{Pid: 0, ExitCode: 0, Time: time.Now()} + e.exitState = &ProcessState{Pid: pid, ExitCode: 0, Time: time.Now()} return } @@ -388,7 +389,7 @@ func (e *UniversalExecutor) wait() { e.logger.Warn("unexpected Cmd.Wait() error type", "error", err) } - e.exitState = &ProcessState{Pid: 0, ExitCode: exitCode, Signal: signal, Time: time.Now()} + e.exitState = &ProcessState{Pid: pid, ExitCode: exitCode, Signal: signal, Time: time.Now()} } var ( diff --git a/helper/testtask/testtask.go b/helper/testtask/testtask.go index db21e9eb4..7fddbd745 100644 --- a/helper/testtask/testtask.go +++ b/helper/testtask/testtask.go @@ -12,6 +12,7 @@ import ( "time" "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/drivers/base" ) // Path returns the path to the currently running executable. @@ -38,6 +39,15 @@ func SetTaskEnv(t *structs.Task) { t.Env["TEST_TASK"] = "execute" } +// SetTaskConfigEnv configures the environment of t so that Run executes a testtask +// script when called from within t. +func SetTaskConfigEnv(t *base.TaskConfig) { + if t.Env == nil { + t.Env = map[string]string{} + } + t.Env["TEST_TASK"] = "execute" +} + // Run interprets os.Args as a testtask script if the current program was // launched with an environment configured by SetCmdEnv or SetTaskEnv. It // returns false if the environment was not set by this package. diff --git a/plugins/drivers/client.go b/plugins/drivers/client.go index 1d1631caf..756bbaf05 100644 --- a/plugins/drivers/client.go +++ b/plugins/drivers/client.go @@ -17,7 +17,7 @@ import ( var _ DriverPlugin = &driverPluginClient{} type driverPluginClient struct { - base.BasePluginClient + *base.BasePluginClient client proto.DriverClient logger hclog.Logger diff --git a/plugins/drivers/driver.go b/plugins/drivers/driver.go index 94c14851f..77aff9ada 100644 --- a/plugins/drivers/driver.go +++ b/plugins/drivers/driver.go @@ -177,6 +177,10 @@ type ExitResult struct { Err error } +func (r *ExitResult) Successful() bool { + return r.ExitCode == 0 && r.Signal == 0 && r.Err == nil +} + type TaskStatus struct { ID string Name string diff --git a/plugins/drivers/plugin.go b/plugins/drivers/plugin.go index 10af11368..3b0362c08 100644 --- a/plugins/drivers/plugin.go +++ b/plugins/drivers/plugin.go @@ -5,7 +5,9 @@ import ( hclog "github.com/hashicorp/go-hclog" plugin "github.com/hashicorp/go-plugin" - "github.com/hashicorp/nomad/plugins/drivers/proto" + "github.com/hashicorp/nomad/plugins/base" + baseproto "github.com/hashicorp/nomad/plugins/base/proto" + "github.com/hashicorp/nomad/plugins/drivers/base/proto" "google.golang.org/grpc" ) @@ -35,6 +37,9 @@ func (p *PluginDriver) GRPCServer(broker *plugin.GRPCBroker, s *grpc.Server) err func (p *PluginDriver) GRPCClient(ctx context.Context, broker *plugin.GRPCBroker, c *grpc.ClientConn) (interface{}, error) { return &driverPluginClient{ + BasePluginClient: &base.BasePluginClient{ + Client: baseproto.NewBasePluginClient(c), + }, client: proto.NewDriverClient(c), logger: p.logger, }, nil diff --git a/plugins/drivers/raw_exec/driver_test.go b/plugins/drivers/raw_exec/driver_test.go index 901982196..8c50e737a 100644 --- a/plugins/drivers/raw_exec/driver_test.go +++ b/plugins/drivers/raw_exec/driver_test.go @@ -2,18 +2,106 @@ package raw_exec import ( "context" + "fmt" "io/ioutil" + "os" "path/filepath" + "strconv" "sync" + "syscall" "testing" "time" + "github.com/davecgh/go-spew/spew" + ctestutil "github.com/hashicorp/nomad/client/testutil" "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/helper/testtask" "github.com/hashicorp/nomad/helper/uuid" + basePlug "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/drivers/base" + "github.com/hashicorp/nomad/testutil" "github.com/stretchr/testify/require" ) +func TestMain(m *testing.M) { + if !testtask.Run() { + os.Exit(m.Run()) + } +} + +func TestRawExecDriver_SetConfig(t *testing.T) { + t.Parallel() + require := require.New(t) + + d := NewRawExecDriver(testlog.HCLogger(t)) + harness := base.NewDriverHarness(t, d) + + // Disable raw exec. + config := &Config{} + + var data []byte + require.NoError(basePlug.MsgPackEncode(&data, config)) + require.NoError(harness.SetConfig(data)) + require.Exactly(config, d.(*RawExecDriver).config) + + config.Enabled = true + config.NoCgroups = true + data = []byte{} + require.NoError(basePlug.MsgPackEncode(&data, config)) + require.NoError(harness.SetConfig(data)) + require.Exactly(config, d.(*RawExecDriver).config) + + config.NoCgroups = false + data = []byte{} + require.NoError(basePlug.MsgPackEncode(&data, config)) + require.NoError(harness.SetConfig(data)) + require.Exactly(config, d.(*RawExecDriver).config) +} + +func TestRawExecDriver_Fingerprint(t *testing.T) { + t.Parallel() + require := require.New(t) + + d := NewRawExecDriver(testlog.HCLogger(t)) + harness := base.NewDriverHarness(t, d) + + // Disable raw exec. + config := &Config{} + + var data []byte + require.NoError(basePlug.MsgPackEncode(&data, config)) + require.NoError(harness.SetConfig(data)) + + fingerCh, err := harness.Fingerprint(context.Background()) + require.NoError(err) + select { + case finger := <-fingerCh: + require.Equal(base.HealthStateUndetected, finger.Health) + require.Empty(finger.Attributes["driver.raw_exec"]) + case <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second): + require.Fail("timeout receiving fingerprint") + } + + // Enable raw exec + config.Enabled = true + data = []byte{} + require.NoError(basePlug.MsgPackEncode(&data, config)) + require.NoError(harness.SetConfig(data)) + +FINGER_LOOP: + for { + select { + case finger := <-fingerCh: + if finger.Health == base.HealthStateHealthy { + break FINGER_LOOP + } + case <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second): + require.Fail("timeout receiving fingerprint") + break FINGER_LOOP + } + } +} + func TestRawExecDriver_StartWait(t *testing.T) { t.Parallel() require := require.New(t) @@ -28,7 +116,7 @@ func TestRawExecDriver_StartWait(t *testing.T) { Command: "go", Args: []string{"version"}, }) - cleanup := harness.MkAllocDir(task) + cleanup := harness.MkAllocDir(task, false) defer cleanup() handle, err := harness.StartTask(task) @@ -52,25 +140,12 @@ func TestRawExecDriver_StartWaitStop(t *testing.T) { Name: "test", } task.EncodeDriverConfig(&TaskConfig{ - Command: "/bin/bash", - Args: []string{"test.sh"}, + Command: testtask.Path(), + Args: []string{"sleep", "100s"}, }) - cleanup := harness.MkAllocDir(task) + cleanup := harness.MkAllocDir(task, false) defer cleanup() - testFile := filepath.Join(task.TaskDir().Dir, "test.sh") - testData := []byte(` -at_term() { - echo 'Terminated.' - exit 3 -} -trap at_term USR1 -while true; do - sleep 1 -done - `) - require.NoError(ioutil.WriteFile(testFile, testData, 0777)) - handle, err := harness.StartTask(task) require.NoError(err) @@ -83,17 +158,18 @@ done go func() { defer wg.Done() result := <-ch - require.Equal(3, result.ExitCode) + spew.Dump(result) + require.Equal(2, result.Signal) waitDone = true }() - time.Sleep(100 * time.Millisecond) + require.NoError(harness.WaitUntilStarted(task.ID, 1*time.Second)) var stopDone bool wg.Add(1) go func() { defer wg.Done() - err := harness.StopTask(task.ID, 1*time.Second, "SIGUSR1") + err := harness.StopTask(task.ID, 2*time.Second, "SIGINT") require.NoError(err) stopDone = true }() @@ -104,14 +180,6 @@ done wg.Wait() }() - time.Sleep(100 * time.Millisecond) - - status, err := harness.InspectTask(task.ID) - require.NoError(err) - require.Equal(base.TaskStateRunning, status.State) - require.False(waitDone) - require.False(stopDone) - select { case <-waitCh: status, err := harness.InspectTask(task.ID) @@ -132,13 +200,14 @@ func TestRawExecDriver_StartWaitRecoverWaitStop(t *testing.T) { harness := base.NewDriverHarness(t, d) task := &base.TaskConfig{ ID: uuid.Generate(), - Name: "test", + Name: "sleep", } task.EncodeDriverConfig(&TaskConfig{ - Command: "sleep", - Args: []string{"1000"}, + Command: testtask.Path(), + Args: []string{"sleep", "100s"}, }) - cleanup := harness.MkAllocDir(task) + testtask.SetTaskConfigEnv(task) + cleanup := harness.MkAllocDir(task, false) defer cleanup() handle, err := harness.StartTask(task) @@ -183,13 +252,198 @@ func TestRawExecDriver_StartWaitRecoverWaitStop(t *testing.T) { defer wg.Done() result := <-ch require.NoError(result.Err) + require.NotZero(result.ExitCode) require.Equal(9, result.Signal) waitDone = true }() - require.NoError(d.StopTask(task.ID, 0, "")) - require.NoError(d.DestroyTask(task.ID, false)) + time.Sleep(300 * time.Millisecond) + require.NoError(d.StopTask(task.ID, 0, "SIGKILL")) wg.Wait() + require.NoError(d.DestroyTask(task.ID, false)) require.True(waitDone) } + +func TestRawExecDriver_Start_Wait_AllocDir(t *testing.T) { + t.Parallel() + require := require.New(t) + + d := NewRawExecDriver(testlog.HCLogger(t)) + harness := base.NewDriverHarness(t, d) + + task := &base.TaskConfig{ + ID: uuid.Generate(), + Name: "sleep", + } + + cleanup := harness.MkAllocDir(task, false) + defer cleanup() + + exp := []byte("win") + file := "output.txt" + outPath := fmt.Sprintf(`%s/%s`, task.TaskDir().SharedAllocDir, file) + task.EncodeDriverConfig(&TaskConfig{ + Command: testtask.Path(), + Args: []string{ + "sleep", "1s", "write", + string(exp), outPath, + }, + }) + + testtask.SetTaskConfigEnv(task) + + _, err := harness.StartTask(task) + require.NoError(err) + + // Task should terminate quickly + waitCh, err := harness.WaitTask(context.Background(), task.ID) + select { + case res := <-waitCh: + require.NoError(res.Err) + require.True(res.Successful()) + case <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second): + require.Fail("WaitTask timeout") + } + + // Check that data was written to the shared alloc directory. + outputFile := filepath.Join(task.TaskDir().SharedAllocDir, file) + act, err := ioutil.ReadFile(outputFile) + require.NoError(err) + require.Exactly(exp, act) + require.NoError(harness.DestroyTask(task.ID, true)) +} + +// This test creates a process tree such that without cgroups tracking the +// processes cleanup of the children would not be possible. Thus the test +// asserts that the processes get killed properly when using cgroups. +func TestRawExecDriver_Start_Kill_Wait_Cgroup(t *testing.T) { + ctestutil.ExecCompatible(t) + t.Parallel() + require := require.New(t) + pidFile := "pid" + + d := NewRawExecDriver(testlog.HCLogger(t)) + harness := base.NewDriverHarness(t, d) + + task := &base.TaskConfig{ + ID: uuid.Generate(), + Name: "sleep", + User: "root", + } + + cleanup := harness.MkAllocDir(task, false) + defer cleanup() + + task.EncodeDriverConfig(&TaskConfig{ + Command: testtask.Path(), + Args: []string{"fork/exec", pidFile, "pgrp", "0", "sleep", "20s"}, + }) + + testtask.SetTaskConfigEnv(task) + + _, err := harness.StartTask(task) + require.NoError(err) + + // Find the process + var pidData []byte + testutil.WaitForResult(func() (bool, error) { + var err error + pidData, err = ioutil.ReadFile(filepath.Join(task.TaskDir().Dir, pidFile)) + if err != nil { + return false, err + } + + if len(pidData) == 0 { + return false, fmt.Errorf("pidFile empty") + } + + return true, nil + }, func(err error) { + require.NoError(err) + }) + + pid, err := strconv.Atoi(string(pidData)) + require.NoError(err, "failed to read pidData: %s", string(pidData)) + + // Check the pid is up + process, err := os.FindProcess(pid) + require.NoError(err) + require.NoError(process.Signal(syscall.Signal(0))) + + var wg sync.WaitGroup + wg.Add(1) + go func() { + defer wg.Done() + time.Sleep(1 * time.Second) + err := harness.StopTask(task.ID, 0, "") + + // Can't rely on the ordering between wait and kill on travis... + if !testutil.IsTravis() { + require.NoError(err) + } + }() + + // Task should terminate quickly + waitCh, err := harness.WaitTask(context.Background(), task.ID) + require.NoError(err) + select { + case res := <-waitCh: + require.False(res.Successful()) + case <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second): + require.Fail("WaitTask timeout") + } + + testutil.WaitForResult(func() (bool, error) { + if err := process.Signal(syscall.Signal(0)); err == nil { + return false, fmt.Errorf("process should not exist: %v", pid) + } + + return true, nil + }, func(err error) { + require.NoError(err) + }) + + wg.Wait() + require.NoError(harness.DestroyTask(task.ID, true)) +} + +func TestRawExecDriver_Exec(t *testing.T) { + t.Parallel() + require := require.New(t) + + d := NewRawExecDriver(testlog.HCLogger(t)) + harness := base.NewDriverHarness(t, d) + + task := &base.TaskConfig{ + ID: uuid.Generate(), + Name: "sleep", + } + + cleanup := harness.MkAllocDir(task, false) + defer cleanup() + + task.EncodeDriverConfig(&TaskConfig{ + Command: testtask.Path(), + Args: []string{"sleep", "9000s"}, + }) + + testtask.SetTaskConfigEnv(task) + + _, err := harness.StartTask(task) + require.NoError(err) + + // Exec a command that should work + res, err := harness.ExecTask(task.ID, []string{"/usr/bin/stat", "/tmp"}, 1*time.Second) + require.NoError(err) + require.True(res.ExitResult.Successful()) + require.True(len(res.Stdout) > 100) + + // Exec a command that should fail + res, err = harness.ExecTask(task.ID, []string{"/usr/bin/stat", "notarealfile123abc"}, 1*time.Second) + require.NoError(err) + require.False(res.ExitResult.Successful()) + require.Contains(string(res.Stdout), "No such file or directory") + + require.NoError(harness.DestroyTask(task.ID, true)) +} diff --git a/plugins/drivers/testing.go b/plugins/drivers/testing.go index 3cea5d457..48dd00dae 100644 --- a/plugins/drivers/testing.go +++ b/plugins/drivers/testing.go @@ -1,17 +1,22 @@ package drivers import ( + "fmt" "io/ioutil" - "os" "path/filepath" + "runtime" "time" "github.com/mitchellh/go-testing-interface" "github.com/stretchr/testify/require" "golang.org/x/net/context" + hclog "github.com/hashicorp/go-hclog" plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/client/allocdir" + "github.com/hashicorp/nomad/client/logmon" "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/helper/uuid" "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/shared/hclspec" ) @@ -21,15 +26,22 @@ type DriverHarness struct { client *plugin.GRPCClient server *plugin.GRPCServer t testing.T + lm logmon.LogMon + logger hclog.Logger } func NewDriverHarness(t testing.T, d DriverPlugin) *DriverHarness { - client, server := plugin.TestPluginGRPCConn(t, map[string]plugin.Plugin{ - base.PluginTypeDriver: &PluginDriver{ - impl: d, - logger: testlog.HCLogger(t), + logger := testlog.HCLogger(t).Named("driver_harness") + client, server := plugin.TestPluginGRPCConn(t, + map[string]plugin.Plugin{ + base.PluginTypeDriver: &PluginDriver{ + impl: d, + logger: logger.Named("driver_plugin"), + }, + base.PluginTypeBase: &base.PluginBase{Impl: d}, + "logmon": logmon.NewPlugin(logmon.NewLogMon(logger.Named("logmon"))), }, - }) + ) raw, err := client.Dispense(base.PluginTypeDriver) if err != nil { @@ -41,8 +53,15 @@ func NewDriverHarness(t testing.T, d DriverPlugin) *DriverHarness { client: client, server: server, DriverPlugin: dClient, + logger: logger, } + raw, err = client.Dispense("logmon") + if err != nil { + t.Fatalf("err dispensing plugin: %v", err) + } + + h.lm = raw.(logmon.LogMon) return h } @@ -52,15 +71,72 @@ func (h *DriverHarness) Kill() { } // MkAllocDir creates a tempory directory and allocdir structure. +// If enableLogs is set to true a logmon instance will be started to write logs +// to the LogDir of the task // A cleanup func is returned and should be defered so as to not leak dirs // between tests. -func (h *DriverHarness) MkAllocDir(t *TaskConfig) func() { - allocDir, err := ioutil.TempDir("", "nomad_driver_harness-") +func (h *DriverHarness) MkAllocDir(t *TaskConfig, enableLogs bool) func() { + dir, err := ioutil.TempDir("", "nomad_driver_harness-") require.NoError(h.t, err) - require.NoError(h.t, os.Mkdir(filepath.Join(allocDir, t.Name), os.ModePerm)) - require.NoError(h.t, os.MkdirAll(filepath.Join(allocDir, "alloc/logs"), os.ModePerm)) - t.AllocDir = allocDir - return func() { os.RemoveAll(allocDir) } + t.AllocDir = dir + + allocDir := allocdir.NewAllocDir(h.logger, dir) + require.NoError(h.t, allocDir.Build()) + taskDir := allocDir.NewTaskDir(t.Name) + require.NoError(h.t, taskDir.Build(false, nil, 0)) + + //logmon + if enableLogs { + var stdoutFifo, stderrFifo string + if runtime.GOOS == "windows" { + id := uuid.Generate()[:8] + stdoutFifo = fmt.Sprintf("//./pipe/%s-%s.stdout", t.Name, id) + stderrFifo = fmt.Sprintf("//./pipe/%s-%s.stderr", t.Name, id) + } else { + stdoutFifo = filepath.Join(taskDir.LogDir, fmt.Sprintf(".%s.stdout.fifo", t.Name)) + stderrFifo = filepath.Join(taskDir.LogDir, fmt.Sprintf(".%s.stderr.fifo", t.Name)) + } + err = h.lm.Start(&logmon.LogConfig{ + LogDir: taskDir.LogDir, + StdoutLogFile: fmt.Sprintf("%s.stdout", t.Name), + StderrLogFile: fmt.Sprintf("%s.stderr", t.Name), + StdoutFifo: stdoutFifo, + StderrFifo: stderrFifo, + MaxFiles: 10, + MaxFileSizeMB: 10, + }) + require.NoError(h.t, err) + + return func() { + h.lm.Stop() + allocDir.Destroy() + } + } + + return func() { + allocDir.Destroy() + } +} + +// WaitUntilStarted will block until the task for the given ID is in the running +// state or the timeout is reached +func (h *DriverHarness) WaitUntilStarted(taskID string, timeout time.Duration) error { + deadline := time.Now().Add(timeout) + var lastState TaskState + for { + status, err := h.InspectTask(taskID) + if err != nil { + return err + } + if status.State == TaskStateRunning { + return nil + } + lastState = status.State + if time.Now().After(deadline) { + return fmt.Errorf("task never transitioned to running, currently '%s'", lastState) + } + time.Sleep(100 * time.Millisecond) + } } // MockDriver is used for testing. From 8b876e1cce76417a2384b760c3a81a38fd16086d Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 26 Sep 2018 10:42:02 -0400 Subject: [PATCH 079/152] fix package references after drivers/base subpackage removed --- helper/testtask/testtask.go | 4 +- plugins/drivers/plugin.go | 2 +- plugins/drivers/raw_exec/driver.go | 108 ++++++++++++------------ plugins/drivers/raw_exec/driver_test.go | 44 +++++----- plugins/drivers/utils/eventer.go | 18 ++-- plugins/drivers/utils/eventer_test.go | 6 +- 6 files changed, 88 insertions(+), 94 deletions(-) diff --git a/helper/testtask/testtask.go b/helper/testtask/testtask.go index 7fddbd745..0152e3be2 100644 --- a/helper/testtask/testtask.go +++ b/helper/testtask/testtask.go @@ -12,7 +12,7 @@ import ( "time" "github.com/hashicorp/nomad/nomad/structs" - "github.com/hashicorp/nomad/plugins/drivers/base" + "github.com/hashicorp/nomad/plugins/drivers" ) // Path returns the path to the currently running executable. @@ -41,7 +41,7 @@ func SetTaskEnv(t *structs.Task) { // SetTaskConfigEnv configures the environment of t so that Run executes a testtask // script when called from within t. -func SetTaskConfigEnv(t *base.TaskConfig) { +func SetTaskConfigEnv(t *drivers.TaskConfig) { if t.Env == nil { t.Env = map[string]string{} } diff --git a/plugins/drivers/plugin.go b/plugins/drivers/plugin.go index 3b0362c08..b485c8836 100644 --- a/plugins/drivers/plugin.go +++ b/plugins/drivers/plugin.go @@ -7,7 +7,7 @@ import ( plugin "github.com/hashicorp/go-plugin" "github.com/hashicorp/nomad/plugins/base" baseproto "github.com/hashicorp/nomad/plugins/base/proto" - "github.com/hashicorp/nomad/plugins/drivers/base/proto" + "github.com/hashicorp/nomad/plugins/drivers/proto" "google.golang.org/grpc" ) diff --git a/plugins/drivers/raw_exec/driver.go b/plugins/drivers/raw_exec/driver.go index 06ca75b76..615230734 100644 --- a/plugins/drivers/raw_exec/driver.go +++ b/plugins/drivers/raw_exec/driver.go @@ -13,8 +13,8 @@ import ( plugin "github.com/hashicorp/go-plugin" "github.com/hashicorp/nomad/client/driver/executor" dstructs "github.com/hashicorp/nomad/client/driver/structs" - bbase "github.com/hashicorp/nomad/plugins/base" - "github.com/hashicorp/nomad/plugins/drivers/base" + "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/drivers" "github.com/hashicorp/nomad/plugins/drivers/utils" "github.com/hashicorp/nomad/plugins/shared/hclspec" "golang.org/x/net/context" @@ -29,8 +29,8 @@ const ( ) var ( - pluginInfo = &bbase.PluginInfoResponse{ - Type: bbase.PluginTypeDriver, + pluginInfo = &base.PluginInfoResponse{ + Type: base.PluginTypeDriver, PluginApiVersion: "0.0.1", PluginVersion: "0.1.0", Name: pluginName, @@ -52,10 +52,10 @@ var ( "args": hclspec.NewAttr("command", "list(string)", false), }) - capabilities = &base.Capabilities{ + capabilities = &drivers.Capabilities{ SendSignals: true, Exec: true, - FSIsolation: base.FSIsolationNone, + FSIsolation: drivers.FSIsolationNone, } ) @@ -69,7 +69,7 @@ type RawExecDriver struct { // fingerprintCh is a channel which other funcs can send fingerprints to // that will immediately be sent - fingerprintCh chan *base.Fingerprint + fingerprintCh chan *drivers.Fingerprint stopCh chan struct{} @@ -90,19 +90,19 @@ type TaskConfig struct { Args []string `codec:"args"` } -func NewRawExecDriver(logger hclog.Logger) base.DriverPlugin { +func NewRawExecDriver(logger hclog.Logger) drivers.DriverPlugin { stopCh := make(chan struct{}) return &RawExecDriver{ Eventer: utils.NewEventer(stopCh), config: &Config{}, tasks: newTaskStore(), - fingerprintCh: make(chan *base.Fingerprint), + fingerprintCh: make(chan *drivers.Fingerprint), stopCh: stopCh, logger: logger.Named(pluginName), } } -func (r *RawExecDriver) PluginInfo() (*bbase.PluginInfoResponse, error) { +func (r *RawExecDriver) PluginInfo() (*base.PluginInfoResponse, error) { return pluginInfo, nil } @@ -112,7 +112,7 @@ func (r *RawExecDriver) ConfigSchema() (*hclspec.Spec, error) { func (r *RawExecDriver) SetConfig(data []byte) error { var config Config - if err := bbase.MsgPackDecode(data, &config); err != nil { + if err := base.MsgPackDecode(data, &config); err != nil { return err } @@ -130,17 +130,17 @@ func (r *RawExecDriver) TaskConfigSchema() (*hclspec.Spec, error) { return taskConfigSpec, nil } -func (r *RawExecDriver) Capabilities() (*base.Capabilities, error) { +func (r *RawExecDriver) Capabilities() (*drivers.Capabilities, error) { return capabilities, nil } -func (r *RawExecDriver) Fingerprint(ctx context.Context) (<-chan *base.Fingerprint, error) { - ch := make(chan *base.Fingerprint) +func (r *RawExecDriver) Fingerprint(ctx context.Context) (<-chan *drivers.Fingerprint, error) { + ch := make(chan *drivers.Fingerprint) go r.fingerprint(ctx, ch) return ch, nil } -func (r *RawExecDriver) fingerprint(ctx context.Context, ch chan *base.Fingerprint) { +func (r *RawExecDriver) fingerprint(ctx context.Context, ch chan *drivers.Fingerprint) { defer close(r.fingerprintCh) ticker := time.NewTimer(0) @@ -162,25 +162,25 @@ func (r *RawExecDriver) fingerprintNow() { r.logger.Debug("fingerprint channel was nil, skipping fingerprint") return } - var health base.HealthState + var health drivers.HealthState var desc string attrs := map[string]string{} if r.config.Enabled { - health = base.HealthStateHealthy + health = drivers.HealthStateHealthy desc = "raw_exec enabled" attrs["driver.raw_exec"] = "1" } else { - health = base.HealthStateUndetected + health = drivers.HealthStateUndetected desc = "raw_exec disabled" } - r.fingerprintCh <- &base.Fingerprint{ + r.fingerprintCh <- &drivers.Fingerprint{ Attributes: map[string]string{}, Health: health, HealthDescription: desc, } } -func (r *RawExecDriver) RecoverTask(handle *base.TaskHandle) error { +func (r *RawExecDriver) RecoverTask(handle *drivers.TaskHandle) error { var taskState RawExecTaskState err := handle.GetDriverState(&taskState) @@ -210,9 +210,9 @@ func (r *RawExecDriver) RecoverTask(handle *base.TaskHandle) error { pid: taskState.Pid, pluginClient: pluginClient, task: taskState.TaskConfig, - procState: base.TaskStateRunning, + procState: drivers.TaskStateRunning, startedAt: taskState.StartedAt, - exitResult: &base.ExitResult{}, + exitResult: &drivers.ExitResult{}, } r.tasks.Set(taskState.TaskConfig.ID, h) @@ -221,7 +221,7 @@ func (r *RawExecDriver) RecoverTask(handle *base.TaskHandle) error { return nil } -func (r *RawExecDriver) StartTask(cfg *base.TaskConfig) (*base.TaskHandle, error) { +func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) { if _, ok := r.tasks.Get(cfg.ID); ok { return nil, fmt.Errorf("task with ID '%s' already started", cfg.ID) } @@ -231,7 +231,7 @@ func (r *RawExecDriver) StartTask(cfg *base.TaskConfig) (*base.TaskHandle, error return nil, err } - handle := base.NewTaskHandle(pluginName) + handle := drivers.NewTaskHandle(pluginName) handle.Config = cfg pluginLogFile := filepath.Join(cfg.TaskDir().Dir, "executor.out") @@ -266,7 +266,7 @@ func (r *RawExecDriver) StartTask(cfg *base.TaskConfig) (*base.TaskHandle, error pid: ps.Pid, pluginClient: pluginClient, task: cfg, - procState: base.TaskStateRunning, + procState: drivers.TaskStateRunning, startedAt: time.Now().Round(time.Millisecond), logger: r.logger, } @@ -291,27 +291,27 @@ func (r *RawExecDriver) StartTask(cfg *base.TaskConfig) (*base.TaskHandle, error return handle, nil } -func (r *RawExecDriver) WaitTask(ctx context.Context, taskID string) (<-chan *base.ExitResult, error) { - ch := make(chan *base.ExitResult) +func (r *RawExecDriver) WaitTask(ctx context.Context, taskID string) (<-chan *drivers.ExitResult, error) { + ch := make(chan *drivers.ExitResult) handle, ok := r.tasks.Get(taskID) if !ok { - return nil, base.ErrTaskNotFound + return nil, drivers.ErrTaskNotFound } go r.handleWait(ctx, handle, ch) return ch, nil } -func (r *RawExecDriver) handleWait(ctx context.Context, handle *rawExecTaskHandle, ch chan *base.ExitResult) { +func (r *RawExecDriver) handleWait(ctx context.Context, handle *rawExecTaskHandle, ch chan *drivers.ExitResult) { defer close(ch) - var result *base.ExitResult + var result *drivers.ExitResult ps, err := handle.exec.Wait() if err != nil { - result = &base.ExitResult{ + result = &drivers.ExitResult{ Err: fmt.Errorf("executor: error waiting on process: %v", err), } } else { - result = &base.ExitResult{ + result = &drivers.ExitResult{ ExitCode: ps.ExitCode, Signal: ps.Signal, } @@ -327,7 +327,7 @@ func (r *RawExecDriver) handleWait(ctx context.Context, handle *rawExecTaskHandl func (r *RawExecDriver) StopTask(taskID string, timeout time.Duration, signal string) error { handle, ok := r.tasks.Get(taskID) if !ok { - return base.ErrTaskNotFound + return drivers.ErrTaskNotFound } if err := handle.exec.Shutdown(signal, timeout); err != nil { @@ -343,7 +343,7 @@ func (r *RawExecDriver) StopTask(taskID string, timeout time.Duration, signal st func (r *RawExecDriver) DestroyTask(taskID string, force bool) error { handle, ok := r.tasks.Get(taskID) if !ok { - return base.ErrTaskNotFound + return drivers.ErrTaskNotFound } if handle.IsRunning() && !force { @@ -364,13 +364,13 @@ func (r *RawExecDriver) DestroyTask(taskID string, force bool) error { return nil } -func (r *RawExecDriver) InspectTask(taskID string) (*base.TaskStatus, error) { +func (r *RawExecDriver) InspectTask(taskID string) (*drivers.TaskStatus, error) { handle, ok := r.tasks.Get(taskID) if !ok { - return nil, base.ErrTaskNotFound + return nil, drivers.ErrTaskNotFound } - status := &base.TaskStatus{ + status := &drivers.TaskStatus{ ID: handle.task.ID, Name: handle.task.Name, State: handle.procState, @@ -381,16 +381,16 @@ func (r *RawExecDriver) InspectTask(taskID string) (*base.TaskStatus, error) { DriverAttributes: map[string]string{ "pid": strconv.Itoa(handle.pid), }, - NetworkOverride: &base.NetworkOverride{}, + NetworkOverride: &drivers.NetworkOverride{}, } return status, nil } -func (r *RawExecDriver) TaskStats(taskID string) (*base.TaskStats, error) { +func (r *RawExecDriver) TaskStats(taskID string) (*drivers.TaskStats, error) { handle, ok := r.tasks.Get(taskID) if !ok { - return nil, base.ErrTaskNotFound + return nil, drivers.ErrTaskNotFound } stats, err := handle.exec.Stats() @@ -398,7 +398,7 @@ func (r *RawExecDriver) TaskStats(taskID string) (*base.TaskStats, error) { return nil, err } - return &base.TaskStats{ + return &drivers.TaskStats{ ID: handle.task.ID, Timestamp: stats.Timestamp, AggResourceUsage: stats.ResourceUsage, @@ -409,7 +409,7 @@ func (r *RawExecDriver) TaskStats(taskID string) (*base.TaskStats, error) { func (r *RawExecDriver) SignalTask(taskID string, signal string) error { handle, ok := r.tasks.Get(taskID) if !ok { - return base.ErrTaskNotFound + return drivers.ErrTaskNotFound } sig := os.Interrupt @@ -419,13 +419,13 @@ func (r *RawExecDriver) SignalTask(taskID string, signal string) error { return handle.exec.Signal(sig) } -func (r *RawExecDriver) ExecTask(taskID string, cmd []string, timeout time.Duration) (*base.ExecTaskResult, error) { +func (r *RawExecDriver) ExecTask(taskID string, cmd []string, timeout time.Duration) (*drivers.ExecTaskResult, error) { if len(cmd) == 0 { return nil, fmt.Errorf("error cmd must have atleast one value") } handle, ok := r.tasks.Get(taskID) if !ok { - return nil, base.ErrTaskNotFound + return nil, drivers.ErrTaskNotFound } args := []string{} @@ -438,9 +438,9 @@ func (r *RawExecDriver) ExecTask(taskID string, cmd []string, timeout time.Durat return nil, err } - return &base.ExecTaskResult{ + return &drivers.ExecTaskResult{ Stdout: out, - ExitResult: &base.ExitResult{ + ExitResult: &drivers.ExitResult{ ExitCode: exitCode, }, }, nil @@ -486,7 +486,7 @@ func (ts *taskStore) Range(f func(id string, handle *rawExecTaskHandle) bool) { type RawExecTaskState struct { ReattachConfig *utils.ReattachConfig - TaskConfig *base.TaskConfig + TaskConfig *drivers.TaskConfig Pid int StartedAt time.Time } @@ -495,31 +495,31 @@ type rawExecTaskHandle struct { exec executor.Executor pid int pluginClient *plugin.Client - task *base.TaskConfig - procState base.TaskState + task *drivers.TaskConfig + procState drivers.TaskState startedAt time.Time completedAt time.Time - exitResult *base.ExitResult + exitResult *drivers.ExitResult logger hclog.Logger } func (h *rawExecTaskHandle) IsRunning() bool { - return h.procState == base.TaskStateRunning + return h.procState == drivers.TaskStateRunning } func (h *rawExecTaskHandle) run() { if h.exitResult == nil { - h.exitResult = &base.ExitResult{} + h.exitResult = &drivers.ExitResult{} } ps, err := h.exec.Wait() if err != nil { h.exitResult.Err = err - h.procState = base.TaskStateUnknown + h.procState = drivers.TaskStateUnknown h.completedAt = time.Now() return } - h.procState = base.TaskStateExited + h.procState = drivers.TaskStateExited h.exitResult.ExitCode = ps.ExitCode h.exitResult.Signal = ps.Signal h.completedAt = ps.Time diff --git a/plugins/drivers/raw_exec/driver_test.go b/plugins/drivers/raw_exec/driver_test.go index 8c50e737a..0ca3caa17 100644 --- a/plugins/drivers/raw_exec/driver_test.go +++ b/plugins/drivers/raw_exec/driver_test.go @@ -12,13 +12,12 @@ import ( "testing" "time" - "github.com/davecgh/go-spew/spew" ctestutil "github.com/hashicorp/nomad/client/testutil" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/helper/testtask" "github.com/hashicorp/nomad/helper/uuid" basePlug "github.com/hashicorp/nomad/plugins/base" - "github.com/hashicorp/nomad/plugins/drivers/base" + "github.com/hashicorp/nomad/plugins/drivers" "github.com/hashicorp/nomad/testutil" "github.com/stretchr/testify/require" ) @@ -34,7 +33,7 @@ func TestRawExecDriver_SetConfig(t *testing.T) { require := require.New(t) d := NewRawExecDriver(testlog.HCLogger(t)) - harness := base.NewDriverHarness(t, d) + harness := drivers.NewDriverHarness(t, d) // Disable raw exec. config := &Config{} @@ -63,7 +62,7 @@ func TestRawExecDriver_Fingerprint(t *testing.T) { require := require.New(t) d := NewRawExecDriver(testlog.HCLogger(t)) - harness := base.NewDriverHarness(t, d) + harness := drivers.NewDriverHarness(t, d) // Disable raw exec. config := &Config{} @@ -76,7 +75,7 @@ func TestRawExecDriver_Fingerprint(t *testing.T) { require.NoError(err) select { case finger := <-fingerCh: - require.Equal(base.HealthStateUndetected, finger.Health) + require.Equal(drivers.HealthStateUndetected, finger.Health) require.Empty(finger.Attributes["driver.raw_exec"]) case <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second): require.Fail("timeout receiving fingerprint") @@ -92,7 +91,7 @@ FINGER_LOOP: for { select { case finger := <-fingerCh: - if finger.Health == base.HealthStateHealthy { + if finger.Health == drivers.HealthStateHealthy { break FINGER_LOOP } case <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second): @@ -107,8 +106,8 @@ func TestRawExecDriver_StartWait(t *testing.T) { require := require.New(t) d := NewRawExecDriver(testlog.HCLogger(t)) - harness := base.NewDriverHarness(t, d) - task := &base.TaskConfig{ + harness := drivers.NewDriverHarness(t, d) + task := &drivers.TaskConfig{ ID: uuid.Generate(), Name: "test", } @@ -134,8 +133,8 @@ func TestRawExecDriver_StartWaitStop(t *testing.T) { require := require.New(t) d := NewRawExecDriver(testlog.HCLogger(t)) - harness := base.NewDriverHarness(t, d) - task := &base.TaskConfig{ + harness := drivers.NewDriverHarness(t, d) + task := &drivers.TaskConfig{ ID: uuid.Generate(), Name: "test", } @@ -152,26 +151,21 @@ func TestRawExecDriver_StartWaitStop(t *testing.T) { ch, err := harness.WaitTask(context.Background(), handle.Config.ID) require.NoError(err) - var waitDone bool var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() result := <-ch - spew.Dump(result) require.Equal(2, result.Signal) - waitDone = true }() require.NoError(harness.WaitUntilStarted(task.ID, 1*time.Second)) - var stopDone bool wg.Add(1) go func() { defer wg.Done() err := harness.StopTask(task.ID, 2*time.Second, "SIGINT") require.NoError(err) - stopDone = true }() waitCh := make(chan struct{}) @@ -184,7 +178,7 @@ func TestRawExecDriver_StartWaitStop(t *testing.T) { case <-waitCh: status, err := harness.InspectTask(task.ID) require.NoError(err) - require.Equal(base.TaskStateExited, status.State) + require.Equal(drivers.TaskStateExited, status.State) case <-time.After(1 * time.Second): require.Fail("timeout waiting for task to shutdown") } @@ -197,8 +191,8 @@ func TestRawExecDriver_StartWaitRecoverWaitStop(t *testing.T) { require := require.New(t) d := NewRawExecDriver(testlog.HCLogger(t)) - harness := base.NewDriverHarness(t, d) - task := &base.TaskConfig{ + harness := drivers.NewDriverHarness(t, d) + task := &drivers.TaskConfig{ ID: uuid.Generate(), Name: "sleep", } @@ -234,7 +228,7 @@ func TestRawExecDriver_StartWaitRecoverWaitStop(t *testing.T) { wg.Wait() require.True(waitDone) _, err = d.InspectTask(task.ID) - require.Equal(base.ErrTaskNotFound, err) + require.Equal(drivers.ErrTaskNotFound, err) err = d.RecoverTask(handle) require.NoError(err) @@ -270,9 +264,9 @@ func TestRawExecDriver_Start_Wait_AllocDir(t *testing.T) { require := require.New(t) d := NewRawExecDriver(testlog.HCLogger(t)) - harness := base.NewDriverHarness(t, d) + harness := drivers.NewDriverHarness(t, d) - task := &base.TaskConfig{ + task := &drivers.TaskConfig{ ID: uuid.Generate(), Name: "sleep", } @@ -324,9 +318,9 @@ func TestRawExecDriver_Start_Kill_Wait_Cgroup(t *testing.T) { pidFile := "pid" d := NewRawExecDriver(testlog.HCLogger(t)) - harness := base.NewDriverHarness(t, d) + harness := drivers.NewDriverHarness(t, d) - task := &base.TaskConfig{ + task := &drivers.TaskConfig{ ID: uuid.Generate(), Name: "sleep", User: "root", @@ -413,9 +407,9 @@ func TestRawExecDriver_Exec(t *testing.T) { require := require.New(t) d := NewRawExecDriver(testlog.HCLogger(t)) - harness := base.NewDriverHarness(t, d) + harness := drivers.NewDriverHarness(t, d) - task := &base.TaskConfig{ + task := &drivers.TaskConfig{ ID: uuid.Generate(), Name: "sleep", } diff --git a/plugins/drivers/utils/eventer.go b/plugins/drivers/utils/eventer.go index 6de4209e5..da6d0c500 100644 --- a/plugins/drivers/utils/eventer.go +++ b/plugins/drivers/utils/eventer.go @@ -4,7 +4,7 @@ import ( "sync" "time" - "github.com/hashicorp/nomad/plugins/drivers/base" + "github.com/hashicorp/nomad/plugins/drivers" "golang.org/x/net/context" ) @@ -20,7 +20,7 @@ type Eventer struct { sync.RWMutex // events is a channel were events to be broadcasted are sent - events chan *base.TaskEvent + events chan *drivers.TaskEvent // streamers is a slice of consumers to broadcast events to // access is gaurded by RWMutex @@ -34,7 +34,7 @@ type Eventer struct { // by closing the given stop channel func NewEventer(stop chan struct{}) *Eventer { e := &Eventer{ - events: make(chan *base.TaskEvent), + events: make(chan *drivers.TaskEvent), stop: stop, } go e.eventLoop() @@ -64,10 +64,10 @@ func (e *Eventer) eventLoop() { type eventStreamer struct { timeout time.Duration ctx context.Context - ch chan *base.TaskEvent + ch chan *drivers.TaskEvent } -func (s *eventStreamer) send(event *base.TaskEvent) { +func (s *eventStreamer) send(event *drivers.TaskEvent) { select { case <-time.After(s.timeout): case <-s.ctx.Done(): @@ -75,12 +75,12 @@ func (s *eventStreamer) send(event *base.TaskEvent) { } } -func (e *Eventer) newStream(ctx context.Context) <-chan *base.TaskEvent { +func (e *Eventer) newStream(ctx context.Context) <-chan *drivers.TaskEvent { e.Lock() defer e.Unlock() stream := &eventStreamer{ - ch: make(chan *base.TaskEvent), + ch: make(chan *drivers.TaskEvent), ctx: ctx, timeout: DefaultSendEventTimeout, } @@ -90,12 +90,12 @@ func (e *Eventer) newStream(ctx context.Context) <-chan *base.TaskEvent { } // TaskEvents is an implementation of the DriverPlugin.TaskEvents function -func (e *Eventer) TaskEvents(ctx context.Context) (<-chan *base.TaskEvent, error) { +func (e *Eventer) TaskEvents(ctx context.Context) (<-chan *drivers.TaskEvent, error) { stream := e.newStream(ctx) return stream, nil } // EmitEvent can be used to broadcast a new event -func (e *Eventer) EmitEvent(event *base.TaskEvent) { +func (e *Eventer) EmitEvent(event *drivers.TaskEvent) { e.events <- event } diff --git a/plugins/drivers/utils/eventer_test.go b/plugins/drivers/utils/eventer_test.go index 93ab7f301..ad14dd283 100644 --- a/plugins/drivers/utils/eventer_test.go +++ b/plugins/drivers/utils/eventer_test.go @@ -6,7 +6,7 @@ import ( "testing" "time" - "github.com/hashicorp/nomad/plugins/drivers/base" + "github.com/hashicorp/nomad/plugins/drivers" "github.com/stretchr/testify/require" ) @@ -17,7 +17,7 @@ func TestEventer(t *testing.T) { stop := make(chan struct{}) e := NewEventer(stop) - events := []*base.TaskEvent{ + events := []*drivers.TaskEvent{ { TaskID: "a", Timestamp: time.Now(), @@ -37,7 +37,7 @@ func TestEventer(t *testing.T) { consumer2, err := e.TaskEvents(context.Background()) require.NoError(err) - var buffer1, buffer2 []*base.TaskEvent + var buffer1, buffer2 []*drivers.TaskEvent var wg sync.WaitGroup wg.Add(2) go func() { From 1f6873806e99a14aca2f83811536497522f42f25 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 26 Sep 2018 15:04:21 -0400 Subject: [PATCH 080/152] raw_exec: move package outside of plugins dir --- {plugins/drivers => drivers}/raw_exec/driver.go | 1 - {plugins/drivers => drivers}/raw_exec/driver_test.go | 0 2 files changed, 1 deletion(-) rename {plugins/drivers => drivers}/raw_exec/driver.go (99%) rename {plugins/drivers => drivers}/raw_exec/driver_test.go (100%) diff --git a/plugins/drivers/raw_exec/driver.go b/drivers/raw_exec/driver.go similarity index 99% rename from plugins/drivers/raw_exec/driver.go rename to drivers/raw_exec/driver.go index 615230734..e0bcc6e6d 100644 --- a/plugins/drivers/raw_exec/driver.go +++ b/drivers/raw_exec/driver.go @@ -374,7 +374,6 @@ func (r *RawExecDriver) InspectTask(taskID string) (*drivers.TaskStatus, error) ID: handle.task.ID, Name: handle.task.Name, State: handle.procState, - SizeOnDiskMB: 0, StartedAt: handle.startedAt, CompletedAt: handle.completedAt, ExitResult: handle.exitResult, diff --git a/plugins/drivers/raw_exec/driver_test.go b/drivers/raw_exec/driver_test.go similarity index 100% rename from plugins/drivers/raw_exec/driver_test.go rename to drivers/raw_exec/driver_test.go From 352c05cdf4b7a78bfbdc69068c57bb2ef7dd79b0 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 26 Sep 2018 15:32:26 -0400 Subject: [PATCH 081/152] plugin/drivers: plumb in stdout/stderr paths --- client/logmon/logmon.go | 4 +- plugins/drivers/driver.go | 2 + plugins/drivers/proto/driver.pb.go | 468 +++++++++++++++-------------- plugins/drivers/proto/driver.proto | 6 + plugins/drivers/testing.go | 16 +- plugins/drivers/utils.go | 4 + 6 files changed, 268 insertions(+), 232 deletions(-) diff --git a/client/logmon/logmon.go b/client/logmon/logmon.go index 2c7609eae..480af0053 100644 --- a/client/logmon/logmon.go +++ b/client/logmon/logmon.go @@ -67,7 +67,9 @@ func (l *logmonImpl) Start(cfg *LogConfig) error { } func (l *logmonImpl) Stop() error { - l.tl.Close() + if l.tl != nil { + l.tl.Close() + } return nil } diff --git a/plugins/drivers/driver.go b/plugins/drivers/driver.go index 77aff9ada..1f47a7165 100644 --- a/plugins/drivers/driver.go +++ b/plugins/drivers/driver.go @@ -102,6 +102,8 @@ type TaskConfig struct { User string AllocDir string rawDriverConfig []byte + StdoutPath string + StderrPath string } func (tc *TaskConfig) EnvList() []string { diff --git a/plugins/drivers/proto/driver.pb.go b/plugins/drivers/proto/driver.pb.go index 0c9542e5e..ee4667d47 100644 --- a/plugins/drivers/proto/driver.pb.go +++ b/plugins/drivers/proto/driver.pb.go @@ -49,7 +49,7 @@ func (x TaskState) String() string { return proto.EnumName(TaskState_name, int32(x)) } func (TaskState) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{0} + return fileDescriptor_driver_16b159caa26c45f9, []int{0} } type FingerprintResponse_HealthState int32 @@ -75,7 +75,7 @@ func (x FingerprintResponse_HealthState) String() string { return proto.EnumName(FingerprintResponse_HealthState_name, int32(x)) } func (FingerprintResponse_HealthState) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{5, 0} + return fileDescriptor_driver_16b159caa26c45f9, []int{5, 0} } type StartTaskResponse_Result int32 @@ -101,7 +101,7 @@ func (x StartTaskResponse_Result) String() string { return proto.EnumName(StartTaskResponse_Result_name, int32(x)) } func (StartTaskResponse_Result) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{9, 0} + return fileDescriptor_driver_16b159caa26c45f9, []int{9, 0} } type DriverCapabilities_FSIsolation int32 @@ -127,7 +127,7 @@ func (x DriverCapabilities_FSIsolation) String() string { return proto.EnumName(DriverCapabilities_FSIsolation_name, int32(x)) } func (DriverCapabilities_FSIsolation) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{25, 0} + return fileDescriptor_driver_16b159caa26c45f9, []int{25, 0} } type CPUUsage_Fields int32 @@ -162,7 +162,7 @@ func (x CPUUsage_Fields) String() string { return proto.EnumName(CPUUsage_Fields_name, int32(x)) } func (CPUUsage_Fields) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{41, 0} + return fileDescriptor_driver_16b159caa26c45f9, []int{41, 0} } type MemoryUsage_Fields int32 @@ -194,7 +194,7 @@ func (x MemoryUsage_Fields) String() string { return proto.EnumName(MemoryUsage_Fields_name, int32(x)) } func (MemoryUsage_Fields) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{42, 0} + return fileDescriptor_driver_16b159caa26c45f9, []int{42, 0} } type TaskConfigSchemaRequest struct { @@ -207,7 +207,7 @@ func (m *TaskConfigSchemaRequest) Reset() { *m = TaskConfigSchemaRequest func (m *TaskConfigSchemaRequest) String() string { return proto.CompactTextString(m) } func (*TaskConfigSchemaRequest) ProtoMessage() {} func (*TaskConfigSchemaRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{0} + return fileDescriptor_driver_16b159caa26c45f9, []int{0} } func (m *TaskConfigSchemaRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskConfigSchemaRequest.Unmarshal(m, b) @@ -239,7 +239,7 @@ func (m *TaskConfigSchemaResponse) Reset() { *m = TaskConfigSchemaRespon func (m *TaskConfigSchemaResponse) String() string { return proto.CompactTextString(m) } func (*TaskConfigSchemaResponse) ProtoMessage() {} func (*TaskConfigSchemaResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{1} + return fileDescriptor_driver_16b159caa26c45f9, []int{1} } func (m *TaskConfigSchemaResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskConfigSchemaResponse.Unmarshal(m, b) @@ -276,7 +276,7 @@ func (m *CapabilitiesRequest) Reset() { *m = CapabilitiesRequest{} } func (m *CapabilitiesRequest) String() string { return proto.CompactTextString(m) } func (*CapabilitiesRequest) ProtoMessage() {} func (*CapabilitiesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{2} + return fileDescriptor_driver_16b159caa26c45f9, []int{2} } func (m *CapabilitiesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CapabilitiesRequest.Unmarshal(m, b) @@ -311,7 +311,7 @@ func (m *CapabilitiesResponse) Reset() { *m = CapabilitiesResponse{} } func (m *CapabilitiesResponse) String() string { return proto.CompactTextString(m) } func (*CapabilitiesResponse) ProtoMessage() {} func (*CapabilitiesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{3} + return fileDescriptor_driver_16b159caa26c45f9, []int{3} } func (m *CapabilitiesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CapabilitiesResponse.Unmarshal(m, b) @@ -348,7 +348,7 @@ func (m *FingerprintRequest) Reset() { *m = FingerprintRequest{} } func (m *FingerprintRequest) String() string { return proto.CompactTextString(m) } func (*FingerprintRequest) ProtoMessage() {} func (*FingerprintRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{4} + return fileDescriptor_driver_16b159caa26c45f9, []int{4} } func (m *FingerprintRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_FingerprintRequest.Unmarshal(m, b) @@ -391,7 +391,7 @@ func (m *FingerprintResponse) Reset() { *m = FingerprintResponse{} } func (m *FingerprintResponse) String() string { return proto.CompactTextString(m) } func (*FingerprintResponse) ProtoMessage() {} func (*FingerprintResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{5} + return fileDescriptor_driver_16b159caa26c45f9, []int{5} } func (m *FingerprintResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_FingerprintResponse.Unmarshal(m, b) @@ -446,7 +446,7 @@ func (m *RecoverTaskRequest) Reset() { *m = RecoverTaskRequest{} } func (m *RecoverTaskRequest) String() string { return proto.CompactTextString(m) } func (*RecoverTaskRequest) ProtoMessage() {} func (*RecoverTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{6} + return fileDescriptor_driver_16b159caa26c45f9, []int{6} } func (m *RecoverTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecoverTaskRequest.Unmarshal(m, b) @@ -490,7 +490,7 @@ func (m *RecoverTaskResponse) Reset() { *m = RecoverTaskResponse{} } func (m *RecoverTaskResponse) String() string { return proto.CompactTextString(m) } func (*RecoverTaskResponse) ProtoMessage() {} func (*RecoverTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{7} + return fileDescriptor_driver_16b159caa26c45f9, []int{7} } func (m *RecoverTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecoverTaskResponse.Unmarshal(m, b) @@ -522,7 +522,7 @@ func (m *StartTaskRequest) Reset() { *m = StartTaskRequest{} } func (m *StartTaskRequest) String() string { return proto.CompactTextString(m) } func (*StartTaskRequest) ProtoMessage() {} func (*StartTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{8} + return fileDescriptor_driver_16b159caa26c45f9, []int{8} } func (m *StartTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StartTaskRequest.Unmarshal(m, b) @@ -576,7 +576,7 @@ func (m *StartTaskResponse) Reset() { *m = StartTaskResponse{} } func (m *StartTaskResponse) String() string { return proto.CompactTextString(m) } func (*StartTaskResponse) ProtoMessage() {} func (*StartTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{9} + return fileDescriptor_driver_16b159caa26c45f9, []int{9} } func (m *StartTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StartTaskResponse.Unmarshal(m, b) @@ -636,7 +636,7 @@ func (m *WaitTaskRequest) Reset() { *m = WaitTaskRequest{} } func (m *WaitTaskRequest) String() string { return proto.CompactTextString(m) } func (*WaitTaskRequest) ProtoMessage() {} func (*WaitTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{10} + return fileDescriptor_driver_16b159caa26c45f9, []int{10} } func (m *WaitTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WaitTaskRequest.Unmarshal(m, b) @@ -677,7 +677,7 @@ func (m *WaitTaskResponse) Reset() { *m = WaitTaskResponse{} } func (m *WaitTaskResponse) String() string { return proto.CompactTextString(m) } func (*WaitTaskResponse) ProtoMessage() {} func (*WaitTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{11} + return fileDescriptor_driver_16b159caa26c45f9, []int{11} } func (m *WaitTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WaitTaskResponse.Unmarshal(m, b) @@ -729,7 +729,7 @@ func (m *StopTaskRequest) Reset() { *m = StopTaskRequest{} } func (m *StopTaskRequest) String() string { return proto.CompactTextString(m) } func (*StopTaskRequest) ProtoMessage() {} func (*StopTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{12} + return fileDescriptor_driver_16b159caa26c45f9, []int{12} } func (m *StopTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StopTaskRequest.Unmarshal(m, b) @@ -780,7 +780,7 @@ func (m *StopTaskResponse) Reset() { *m = StopTaskResponse{} } func (m *StopTaskResponse) String() string { return proto.CompactTextString(m) } func (*StopTaskResponse) ProtoMessage() {} func (*StopTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{13} + return fileDescriptor_driver_16b159caa26c45f9, []int{13} } func (m *StopTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StopTaskResponse.Unmarshal(m, b) @@ -814,7 +814,7 @@ func (m *DestroyTaskRequest) Reset() { *m = DestroyTaskRequest{} } func (m *DestroyTaskRequest) String() string { return proto.CompactTextString(m) } func (*DestroyTaskRequest) ProtoMessage() {} func (*DestroyTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{14} + return fileDescriptor_driver_16b159caa26c45f9, []int{14} } func (m *DestroyTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DestroyTaskRequest.Unmarshal(m, b) @@ -858,7 +858,7 @@ func (m *DestroyTaskResponse) Reset() { *m = DestroyTaskResponse{} } func (m *DestroyTaskResponse) String() string { return proto.CompactTextString(m) } func (*DestroyTaskResponse) ProtoMessage() {} func (*DestroyTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{15} + return fileDescriptor_driver_16b159caa26c45f9, []int{15} } func (m *DestroyTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DestroyTaskResponse.Unmarshal(m, b) @@ -890,7 +890,7 @@ func (m *InspectTaskRequest) Reset() { *m = InspectTaskRequest{} } func (m *InspectTaskRequest) String() string { return proto.CompactTextString(m) } func (*InspectTaskRequest) ProtoMessage() {} func (*InspectTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{16} + return fileDescriptor_driver_16b159caa26c45f9, []int{16} } func (m *InspectTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectTaskRequest.Unmarshal(m, b) @@ -933,7 +933,7 @@ func (m *InspectTaskResponse) Reset() { *m = InspectTaskResponse{} } func (m *InspectTaskResponse) String() string { return proto.CompactTextString(m) } func (*InspectTaskResponse) ProtoMessage() {} func (*InspectTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{17} + return fileDescriptor_driver_16b159caa26c45f9, []int{17} } func (m *InspectTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectTaskResponse.Unmarshal(m, b) @@ -986,7 +986,7 @@ func (m *TaskStatsRequest) Reset() { *m = TaskStatsRequest{} } func (m *TaskStatsRequest) String() string { return proto.CompactTextString(m) } func (*TaskStatsRequest) ProtoMessage() {} func (*TaskStatsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{18} + return fileDescriptor_driver_16b159caa26c45f9, []int{18} } func (m *TaskStatsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStatsRequest.Unmarshal(m, b) @@ -1025,7 +1025,7 @@ func (m *TaskStatsResponse) Reset() { *m = TaskStatsResponse{} } func (m *TaskStatsResponse) String() string { return proto.CompactTextString(m) } func (*TaskStatsResponse) ProtoMessage() {} func (*TaskStatsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{19} + return fileDescriptor_driver_16b159caa26c45f9, []int{19} } func (m *TaskStatsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStatsResponse.Unmarshal(m, b) @@ -1062,7 +1062,7 @@ func (m *TaskEventsRequest) Reset() { *m = TaskEventsRequest{} } func (m *TaskEventsRequest) String() string { return proto.CompactTextString(m) } func (*TaskEventsRequest) ProtoMessage() {} func (*TaskEventsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{20} + return fileDescriptor_driver_16b159caa26c45f9, []int{20} } func (m *TaskEventsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskEventsRequest.Unmarshal(m, b) @@ -1096,7 +1096,7 @@ func (m *SignalTaskRequest) Reset() { *m = SignalTaskRequest{} } func (m *SignalTaskRequest) String() string { return proto.CompactTextString(m) } func (*SignalTaskRequest) ProtoMessage() {} func (*SignalTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{21} + return fileDescriptor_driver_16b159caa26c45f9, []int{21} } func (m *SignalTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SignalTaskRequest.Unmarshal(m, b) @@ -1140,7 +1140,7 @@ func (m *SignalTaskResponse) Reset() { *m = SignalTaskResponse{} } func (m *SignalTaskResponse) String() string { return proto.CompactTextString(m) } func (*SignalTaskResponse) ProtoMessage() {} func (*SignalTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{22} + return fileDescriptor_driver_16b159caa26c45f9, []int{22} } func (m *SignalTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SignalTaskResponse.Unmarshal(m, b) @@ -1177,7 +1177,7 @@ func (m *ExecTaskRequest) Reset() { *m = ExecTaskRequest{} } func (m *ExecTaskRequest) String() string { return proto.CompactTextString(m) } func (*ExecTaskRequest) ProtoMessage() {} func (*ExecTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{23} + return fileDescriptor_driver_16b159caa26c45f9, []int{23} } func (m *ExecTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExecTaskRequest.Unmarshal(m, b) @@ -1234,7 +1234,7 @@ func (m *ExecTaskResponse) Reset() { *m = ExecTaskResponse{} } func (m *ExecTaskResponse) String() string { return proto.CompactTextString(m) } func (*ExecTaskResponse) ProtoMessage() {} func (*ExecTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{24} + return fileDescriptor_driver_16b159caa26c45f9, []int{24} } func (m *ExecTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExecTaskResponse.Unmarshal(m, b) @@ -1293,7 +1293,7 @@ func (m *DriverCapabilities) Reset() { *m = DriverCapabilities{} } func (m *DriverCapabilities) String() string { return proto.CompactTextString(m) } func (*DriverCapabilities) ProtoMessage() {} func (*DriverCapabilities) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{25} + return fileDescriptor_driver_16b159caa26c45f9, []int{25} } func (m *DriverCapabilities) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DriverCapabilities.Unmarshal(m, b) @@ -1354,7 +1354,11 @@ type TaskConfig struct { User string `protobuf:"bytes,8,opt,name=user,proto3" json:"user,omitempty"` // AllocDir is the directory on the host where the allocation directory // exists. - AllocDir string `protobuf:"bytes,9,opt,name=alloc_dir,json=allocDir,proto3" json:"alloc_dir,omitempty"` + AllocDir string `protobuf:"bytes,9,opt,name=alloc_dir,json=allocDir,proto3" json:"alloc_dir,omitempty"` + // StdoutPath is the path to the file to open and write task stdout to + StdoutPath string `protobuf:"bytes,10,opt,name=stdout_path,json=stdoutPath,proto3" json:"stdout_path,omitempty"` + // StderrPath is the path to the file to open and write task stderr to + StderrPath string `protobuf:"bytes,11,opt,name=stderr_path,json=stderrPath,proto3" json:"stderr_path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1364,7 +1368,7 @@ func (m *TaskConfig) Reset() { *m = TaskConfig{} } func (m *TaskConfig) String() string { return proto.CompactTextString(m) } func (*TaskConfig) ProtoMessage() {} func (*TaskConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{26} + return fileDescriptor_driver_16b159caa26c45f9, []int{26} } func (m *TaskConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskConfig.Unmarshal(m, b) @@ -1447,6 +1451,20 @@ func (m *TaskConfig) GetAllocDir() string { return "" } +func (m *TaskConfig) GetStdoutPath() string { + if m != nil { + return m.StdoutPath + } + return "" +} + +func (m *TaskConfig) GetStderrPath() string { + if m != nil { + return m.StderrPath + } + return "" +} + type Resources struct { // RawResources are the resources set for the task RawResources *RawResources `protobuf:"bytes,1,opt,name=raw_resources,json=rawResources,proto3" json:"raw_resources,omitempty"` @@ -1461,7 +1479,7 @@ func (m *Resources) Reset() { *m = Resources{} } func (m *Resources) String() string { return proto.CompactTextString(m) } func (*Resources) ProtoMessage() {} func (*Resources) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{27} + return fileDescriptor_driver_16b159caa26c45f9, []int{27} } func (m *Resources) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Resources.Unmarshal(m, b) @@ -1510,7 +1528,7 @@ func (m *RawResources) Reset() { *m = RawResources{} } func (m *RawResources) String() string { return proto.CompactTextString(m) } func (*RawResources) ProtoMessage() {} func (*RawResources) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{28} + return fileDescriptor_driver_16b159caa26c45f9, []int{28} } func (m *RawResources) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RawResources.Unmarshal(m, b) @@ -1581,7 +1599,7 @@ func (m *NetworkResource) Reset() { *m = NetworkResource{} } func (m *NetworkResource) String() string { return proto.CompactTextString(m) } func (*NetworkResource) ProtoMessage() {} func (*NetworkResource) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{29} + return fileDescriptor_driver_16b159caa26c45f9, []int{29} } func (m *NetworkResource) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NetworkResource.Unmarshal(m, b) @@ -1655,7 +1673,7 @@ func (m *NetworkPort) Reset() { *m = NetworkPort{} } func (m *NetworkPort) String() string { return proto.CompactTextString(m) } func (*NetworkPort) ProtoMessage() {} func (*NetworkPort) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{30} + return fileDescriptor_driver_16b159caa26c45f9, []int{30} } func (m *NetworkPort) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NetworkPort.Unmarshal(m, b) @@ -1713,7 +1731,7 @@ func (m *LinuxResources) Reset() { *m = LinuxResources{} } func (m *LinuxResources) String() string { return proto.CompactTextString(m) } func (*LinuxResources) ProtoMessage() {} func (*LinuxResources) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{31} + return fileDescriptor_driver_16b159caa26c45f9, []int{31} } func (m *LinuxResources) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LinuxResources.Unmarshal(m, b) @@ -1798,7 +1816,7 @@ func (m *Mount) Reset() { *m = Mount{} } func (m *Mount) String() string { return proto.CompactTextString(m) } func (*Mount) ProtoMessage() {} func (*Mount) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{32} + return fileDescriptor_driver_16b159caa26c45f9, []int{32} } func (m *Mount) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mount.Unmarshal(m, b) @@ -1861,7 +1879,7 @@ func (m *Device) Reset() { *m = Device{} } func (m *Device) String() string { return proto.CompactTextString(m) } func (*Device) ProtoMessage() {} func (*Device) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{33} + return fileDescriptor_driver_16b159caa26c45f9, []int{33} } func (m *Device) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Device.Unmarshal(m, b) @@ -1919,7 +1937,7 @@ func (m *TaskHandle) Reset() { *m = TaskHandle{} } func (m *TaskHandle) String() string { return proto.CompactTextString(m) } func (*TaskHandle) ProtoMessage() {} func (*TaskHandle) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{34} + return fileDescriptor_driver_16b159caa26c45f9, []int{34} } func (m *TaskHandle) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskHandle.Unmarshal(m, b) @@ -1979,7 +1997,7 @@ func (m *NetworkOverride) Reset() { *m = NetworkOverride{} } func (m *NetworkOverride) String() string { return proto.CompactTextString(m) } func (*NetworkOverride) ProtoMessage() {} func (*NetworkOverride) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{35} + return fileDescriptor_driver_16b159caa26c45f9, []int{35} } func (m *NetworkOverride) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NetworkOverride.Unmarshal(m, b) @@ -2037,7 +2055,7 @@ func (m *ExitResult) Reset() { *m = ExitResult{} } func (m *ExitResult) String() string { return proto.CompactTextString(m) } func (*ExitResult) ProtoMessage() {} func (*ExitResult) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{36} + return fileDescriptor_driver_16b159caa26c45f9, []int{36} } func (m *ExitResult) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExitResult.Unmarshal(m, b) @@ -2100,7 +2118,7 @@ func (m *TaskStatus) Reset() { *m = TaskStatus{} } func (m *TaskStatus) String() string { return proto.CompactTextString(m) } func (*TaskStatus) ProtoMessage() {} func (*TaskStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{37} + return fileDescriptor_driver_16b159caa26c45f9, []int{37} } func (m *TaskStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStatus.Unmarshal(m, b) @@ -2175,7 +2193,7 @@ func (m *TaskDriverStatus) Reset() { *m = TaskDriverStatus{} } func (m *TaskDriverStatus) String() string { return proto.CompactTextString(m) } func (*TaskDriverStatus) ProtoMessage() {} func (*TaskDriverStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{38} + return fileDescriptor_driver_16b159caa26c45f9, []int{38} } func (m *TaskDriverStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskDriverStatus.Unmarshal(m, b) @@ -2220,7 +2238,7 @@ func (m *TaskStats) Reset() { *m = TaskStats{} } func (m *TaskStats) String() string { return proto.CompactTextString(m) } func (*TaskStats) ProtoMessage() {} func (*TaskStats) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{39} + return fileDescriptor_driver_16b159caa26c45f9, []int{39} } func (m *TaskStats) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStats.Unmarshal(m, b) @@ -2282,7 +2300,7 @@ func (m *TaskResourceUsage) Reset() { *m = TaskResourceUsage{} } func (m *TaskResourceUsage) String() string { return proto.CompactTextString(m) } func (*TaskResourceUsage) ProtoMessage() {} func (*TaskResourceUsage) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{40} + return fileDescriptor_driver_16b159caa26c45f9, []int{40} } func (m *TaskResourceUsage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskResourceUsage.Unmarshal(m, b) @@ -2334,7 +2352,7 @@ func (m *CPUUsage) Reset() { *m = CPUUsage{} } func (m *CPUUsage) String() string { return proto.CompactTextString(m) } func (*CPUUsage) ProtoMessage() {} func (*CPUUsage) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{41} + return fileDescriptor_driver_16b159caa26c45f9, []int{41} } func (m *CPUUsage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CPUUsage.Unmarshal(m, b) @@ -2420,7 +2438,7 @@ func (m *MemoryUsage) Reset() { *m = MemoryUsage{} } func (m *MemoryUsage) String() string { return proto.CompactTextString(m) } func (*MemoryUsage) ProtoMessage() {} func (*MemoryUsage) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{42} + return fileDescriptor_driver_16b159caa26c45f9, []int{42} } func (m *MemoryUsage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MemoryUsage.Unmarshal(m, b) @@ -2500,7 +2518,7 @@ func (m *DriverTaskEvent) Reset() { *m = DriverTaskEvent{} } func (m *DriverTaskEvent) String() string { return proto.CompactTextString(m) } func (*DriverTaskEvent) ProtoMessage() {} func (*DriverTaskEvent) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_c5fa1f582257ecd9, []int{43} + return fileDescriptor_driver_16b159caa26c45f9, []int{43} } func (m *DriverTaskEvent) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DriverTaskEvent.Unmarshal(m, b) @@ -3190,179 +3208,181 @@ var _Driver_serviceDesc = grpc.ServiceDesc{ Metadata: "driver.proto", } -func init() { proto.RegisterFile("driver.proto", fileDescriptor_driver_c5fa1f582257ecd9) } +func init() { proto.RegisterFile("driver.proto", fileDescriptor_driver_16b159caa26c45f9) } -var fileDescriptor_driver_c5fa1f582257ecd9 = []byte{ - // 2732 bytes of a gzipped FileDescriptorProto +var fileDescriptor_driver_16b159caa26c45f9 = []byte{ + // 2757 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x59, 0xcb, 0x6f, 0x23, 0xc7, - 0xd1, 0x17, 0x49, 0x91, 0x22, 0x8b, 0x14, 0x35, 0xdb, 0xbb, 0xfb, 0x99, 0xa6, 0xf1, 0x7d, 0x5e, - 0x0f, 0x60, 0x40, 0xb0, 0xbd, 0x94, 0x2d, 0xe3, 0xf3, 0x3e, 0x02, 0x3f, 0x68, 0x6a, 0x76, 0x25, - 0xaf, 0x44, 0x29, 0x4d, 0x0a, 0xeb, 0x4d, 0x62, 0x4f, 0x46, 0x33, 0x2d, 0x72, 0x2c, 0xce, 0xc3, - 0xd3, 0x3d, 0x5a, 0x09, 0x41, 0x90, 0x20, 0x09, 0x82, 0x24, 0x40, 0x80, 0x5c, 0x82, 0xdc, 0x73, - 0x0b, 0x72, 0xcd, 0x29, 0x41, 0x2e, 0x01, 0xf2, 0x3f, 0xe4, 0x98, 0x4b, 0x80, 0x5c, 0x73, 0xcd, - 0x29, 0xe8, 0xc7, 0x0c, 0x87, 0x92, 0xd6, 0x1e, 0x52, 0x39, 0x4d, 0x77, 0x55, 0xf7, 0xaf, 0x6b, - 0xaa, 0xaa, 0xbb, 0xaa, 0xbb, 0xa0, 0xe1, 0x44, 0xee, 0x29, 0x89, 0x3a, 0x61, 0x14, 0xb0, 0x00, - 0xbd, 0x3e, 0xb6, 0xe8, 0xd8, 0xb5, 0x83, 0x28, 0xec, 0xf8, 0x81, 0x67, 0x39, 0x9d, 0x70, 0x12, - 0x8f, 0x5c, 0x9f, 0x76, 0xe4, 0x28, 0x2a, 0x87, 0xb5, 0xff, 0x6f, 0x14, 0x04, 0xa3, 0x09, 0xd9, - 0x10, 0xbd, 0xa3, 0xf8, 0x78, 0xc3, 0x89, 0x23, 0x8b, 0xb9, 0x81, 0xaf, 0xf8, 0xaf, 0x5e, 0xe4, - 0x33, 0xd7, 0x23, 0x94, 0x59, 0x5e, 0xa8, 0x06, 0x7c, 0x34, 0x72, 0xd9, 0x38, 0x3e, 0xea, 0xd8, - 0x81, 0xb7, 0x91, 0x2e, 0xb9, 0x21, 0x96, 0xdc, 0x50, 0x4b, 0x6e, 0xd0, 0xb1, 0x15, 0x11, 0x67, - 0x63, 0x6c, 0x4f, 0x68, 0x48, 0x6c, 0xfe, 0x35, 0x79, 0x43, 0x22, 0xe8, 0x2f, 0xc3, 0x4b, 0x43, - 0x8b, 0x9e, 0xf4, 0x02, 0xff, 0xd8, 0x1d, 0x0d, 0xec, 0x31, 0xf1, 0x2c, 0x4c, 0xbe, 0x8c, 0x09, - 0x65, 0xfa, 0x77, 0xa0, 0x75, 0x99, 0x45, 0xc3, 0xc0, 0xa7, 0x04, 0x7d, 0x04, 0xcb, 0x1c, 0xa4, - 0x55, 0xb8, 0x53, 0x58, 0xaf, 0x6f, 0xbe, 0xd5, 0x79, 0xd1, 0xff, 0xca, 0xc5, 0x3b, 0x6a, 0xf1, - 0xce, 0x20, 0x24, 0x36, 0x16, 0x33, 0xf5, 0xdb, 0x70, 0xb3, 0x67, 0x85, 0xd6, 0x91, 0x3b, 0x71, - 0x99, 0x4b, 0x68, 0xb2, 0x68, 0x0c, 0xb7, 0x66, 0xc9, 0x6a, 0xc1, 0xcf, 0xa0, 0x61, 0x67, 0xe8, - 0x6a, 0xe1, 0x07, 0x9d, 0x5c, 0x8a, 0xee, 0x6c, 0x89, 0xde, 0x0c, 0xf0, 0x0c, 0x9c, 0x7e, 0x0b, - 0xd0, 0x23, 0xd7, 0x1f, 0x91, 0x28, 0x8c, 0x5c, 0x9f, 0x25, 0xc2, 0xfc, 0xa2, 0x04, 0x37, 0x67, - 0xc8, 0x4a, 0x98, 0x2f, 0x00, 0x2c, 0xc6, 0x22, 0xf7, 0x28, 0x66, 0x42, 0x94, 0xd2, 0x7a, 0x7d, - 0xf3, 0x93, 0x9c, 0xa2, 0x5c, 0x81, 0xd7, 0xe9, 0xa6, 0x60, 0x86, 0xcf, 0xa2, 0x73, 0x9c, 0x41, - 0x47, 0x9f, 0x43, 0x65, 0x4c, 0xac, 0x09, 0x1b, 0xb7, 0x8a, 0x77, 0x0a, 0xeb, 0xcd, 0xcd, 0x47, - 0xd7, 0x58, 0x67, 0x5b, 0x00, 0x0d, 0x98, 0xc5, 0x08, 0x56, 0xa8, 0xe8, 0x2e, 0x20, 0xd9, 0x32, - 0x1d, 0x42, 0xed, 0xc8, 0x0d, 0xb9, 0xff, 0xb5, 0x4a, 0x77, 0x0a, 0xeb, 0x35, 0x7c, 0x43, 0x72, - 0xb6, 0xa6, 0x8c, 0xf6, 0xfb, 0xb0, 0x76, 0x41, 0x5a, 0xa4, 0x41, 0xe9, 0x84, 0x9c, 0x0b, 0x8b, - 0xd4, 0x30, 0x6f, 0xa2, 0x5b, 0x50, 0x3e, 0xb5, 0x26, 0x31, 0x11, 0x22, 0xd7, 0xb0, 0xec, 0x3c, - 0x2c, 0xde, 0x2f, 0xe8, 0x0f, 0xa0, 0x9e, 0x11, 0x02, 0x35, 0x01, 0x0e, 0xfb, 0x5b, 0xc6, 0xd0, - 0xe8, 0x0d, 0x8d, 0x2d, 0x6d, 0x09, 0xad, 0x42, 0xed, 0xb0, 0xbf, 0x6d, 0x74, 0x77, 0x87, 0xdb, - 0xcf, 0xb4, 0x02, 0xaa, 0xc3, 0x4a, 0xd2, 0x29, 0xea, 0x67, 0x80, 0x30, 0xb1, 0x83, 0x53, 0x12, - 0x71, 0xaf, 0x54, 0x26, 0x42, 0x2f, 0xc1, 0x0a, 0xb3, 0xe8, 0x89, 0xe9, 0x3a, 0x4a, 0x80, 0x0a, - 0xef, 0xee, 0x38, 0x68, 0x07, 0x2a, 0x63, 0xcb, 0x77, 0x26, 0x52, 0x88, 0xfa, 0xe6, 0x3b, 0x39, - 0xf5, 0xc6, 0xc1, 0xb7, 0xc5, 0x44, 0xac, 0x00, 0xb8, 0xab, 0xce, 0xac, 0x2c, 0xb5, 0xa9, 0x3f, - 0x03, 0x6d, 0xc0, 0xac, 0x88, 0x65, 0xc5, 0x31, 0x60, 0x99, 0xaf, 0xaf, 0xdc, 0x73, 0x9e, 0x35, - 0xe5, 0x36, 0xc3, 0x62, 0xba, 0xfe, 0xaf, 0x22, 0xdc, 0xc8, 0x60, 0x2b, 0xb7, 0x7b, 0x0a, 0x95, - 0x88, 0xd0, 0x78, 0xc2, 0x04, 0x7c, 0x73, 0xf3, 0xc3, 0x9c, 0xf0, 0x97, 0x90, 0x3a, 0x58, 0xc0, - 0x60, 0x05, 0x87, 0xd6, 0x41, 0x93, 0x33, 0x4c, 0x12, 0x45, 0x41, 0x64, 0x7a, 0x74, 0xa4, 0x4c, - 0xd7, 0x94, 0x74, 0x83, 0x93, 0xf7, 0xe8, 0x28, 0xa3, 0xd5, 0xd2, 0x35, 0xb5, 0x8a, 0x2c, 0xd0, - 0x7c, 0xc2, 0x9e, 0x07, 0xd1, 0x89, 0xc9, 0x55, 0x1b, 0xb9, 0x0e, 0x69, 0x2d, 0x0b, 0xd0, 0xf7, - 0x72, 0x82, 0xf6, 0xe5, 0xf4, 0x7d, 0x35, 0x1b, 0xaf, 0xf9, 0xb3, 0x04, 0xfd, 0x4d, 0xa8, 0xc8, - 0x3f, 0xe5, 0x9e, 0x34, 0x38, 0xec, 0xf5, 0x8c, 0xc1, 0x40, 0x5b, 0x42, 0x35, 0x28, 0x63, 0x63, - 0x88, 0xb9, 0x87, 0xd5, 0xa0, 0xfc, 0xa8, 0x3b, 0xec, 0xee, 0x6a, 0x45, 0xfd, 0x0d, 0x58, 0x7b, - 0x6a, 0xb9, 0x2c, 0x8f, 0x73, 0xe9, 0x01, 0x68, 0xd3, 0xb1, 0xca, 0x3a, 0x3b, 0x33, 0xd6, 0xc9, - 0xaf, 0x1a, 0xe3, 0xcc, 0x65, 0x17, 0xec, 0xa1, 0x41, 0x89, 0x44, 0x91, 0x32, 0x01, 0x6f, 0xea, - 0xcf, 0x61, 0x6d, 0xc0, 0x82, 0x30, 0x97, 0xe7, 0xbf, 0x0b, 0x2b, 0x3c, 0x4e, 0x04, 0x31, 0x53, - 0xae, 0xff, 0x72, 0x47, 0xc6, 0x91, 0x4e, 0x12, 0x47, 0x3a, 0x5b, 0x2a, 0xce, 0xe0, 0x64, 0x24, - 0xfa, 0x1f, 0xa8, 0x50, 0x77, 0xe4, 0x5b, 0x13, 0xb5, 0xf5, 0x55, 0x4f, 0x47, 0xdc, 0xc9, 0x93, - 0x85, 0x95, 0xe3, 0xf7, 0x00, 0x6d, 0x11, 0xca, 0xa2, 0xe0, 0x3c, 0x97, 0x3c, 0xb7, 0xa0, 0x7c, - 0x1c, 0x44, 0xb6, 0xdc, 0x88, 0x55, 0x2c, 0x3b, 0x7c, 0x53, 0xcd, 0x80, 0x28, 0xec, 0xbb, 0x80, - 0x76, 0x7c, 0x1e, 0x20, 0xf2, 0x19, 0xe2, 0x57, 0x45, 0xb8, 0x39, 0x33, 0x5e, 0x19, 0x63, 0xf1, - 0x7d, 0xc8, 0x0f, 0xa6, 0x98, 0xca, 0x7d, 0x88, 0xf6, 0xa1, 0x22, 0x47, 0x28, 0x4d, 0xde, 0x9b, - 0x03, 0x48, 0xc6, 0x1c, 0x05, 0xa7, 0x60, 0xae, 0x74, 0xfa, 0xd2, 0x7f, 0xdb, 0xe9, 0xb5, 0xe4, - 0x3f, 0xe8, 0xd7, 0xea, 0xef, 0xdb, 0x70, 0x23, 0x33, 0x58, 0x29, 0xef, 0x11, 0x94, 0x29, 0x27, - 0x28, 0xed, 0xbd, 0x3d, 0xa7, 0xf6, 0x28, 0x96, 0xd3, 0xf5, 0x9b, 0x12, 0xdc, 0x38, 0x25, 0x7e, - 0x2a, 0x8a, 0xbe, 0x05, 0x37, 0x06, 0xc2, 0xb5, 0x72, 0xf9, 0xce, 0xd4, 0x2d, 0x8b, 0x33, 0x6e, - 0x79, 0x0b, 0x50, 0x16, 0x45, 0x39, 0xcf, 0x39, 0xac, 0x19, 0x67, 0xc4, 0xce, 0x85, 0xdc, 0x82, - 0x15, 0x3b, 0xf0, 0x3c, 0xcb, 0x77, 0x5a, 0xc5, 0x3b, 0xa5, 0xf5, 0x1a, 0x4e, 0xba, 0xd9, 0xfd, - 0x53, 0xca, 0xbb, 0x7f, 0xf4, 0x5f, 0x16, 0x40, 0x9b, 0xae, 0xad, 0x14, 0xc9, 0xa5, 0x67, 0x0e, - 0x07, 0xe2, 0x6b, 0x37, 0xb0, 0xea, 0x29, 0x7a, 0xb2, 0xc5, 0x25, 0x9d, 0x44, 0x51, 0xe6, 0x08, - 0x29, 0x5d, 0xf3, 0x08, 0xd1, 0xff, 0x51, 0x00, 0x74, 0x39, 0xeb, 0x41, 0xaf, 0x41, 0x83, 0x12, - 0xdf, 0x31, 0xa5, 0x1a, 0xa5, 0x85, 0xab, 0xb8, 0xce, 0x69, 0x52, 0x9f, 0x14, 0x21, 0x58, 0x26, - 0x67, 0xc4, 0x56, 0xbb, 0x55, 0xb4, 0xd1, 0x18, 0x1a, 0xc7, 0xd4, 0x74, 0x69, 0x30, 0xb1, 0xd2, - 0xf4, 0xa0, 0xb9, 0x69, 0x2c, 0x9c, 0x7d, 0x75, 0x1e, 0x0d, 0x76, 0x12, 0x30, 0x5c, 0x3f, 0xa6, - 0x69, 0x47, 0xef, 0x40, 0x3d, 0xc3, 0x43, 0x55, 0x58, 0xee, 0xef, 0xf7, 0x0d, 0x6d, 0x09, 0x01, - 0x54, 0x7a, 0xdb, 0x78, 0x7f, 0x7f, 0x28, 0x4f, 0xed, 0x9d, 0xbd, 0xee, 0x63, 0x43, 0x2b, 0xea, - 0xff, 0x2e, 0x01, 0x4c, 0xc3, 0x27, 0x6a, 0x42, 0x31, 0xb5, 0x74, 0xd1, 0x75, 0xf8, 0xcf, 0xf8, - 0x96, 0x97, 0x24, 0x22, 0xa2, 0x8d, 0x36, 0xe1, 0xb6, 0x47, 0x47, 0xa1, 0x65, 0x9f, 0x98, 0x2a, - 0xea, 0xd9, 0x62, 0xb2, 0xf8, 0xab, 0x06, 0xbe, 0xa9, 0x98, 0x4a, 0x6a, 0x89, 0xbb, 0x0b, 0x25, - 0xe2, 0x9f, 0xb6, 0x96, 0x45, 0xaa, 0xf7, 0x70, 0xee, 0xb0, 0xde, 0x31, 0xfc, 0x53, 0x99, 0xda, - 0x71, 0x18, 0xd4, 0x87, 0x5a, 0x44, 0x68, 0x10, 0x47, 0x36, 0xa1, 0xad, 0xf2, 0x5c, 0x9b, 0x0c, - 0x27, 0xf3, 0xf0, 0x14, 0x02, 0x6d, 0x41, 0xc5, 0x0b, 0x62, 0x9f, 0xd1, 0x56, 0x45, 0x08, 0xf8, - 0x56, 0x4e, 0xb0, 0x3d, 0x3e, 0x09, 0xab, 0xb9, 0xe8, 0x31, 0xac, 0x38, 0xe4, 0xd4, 0xe5, 0x32, - 0xad, 0x08, 0x98, 0xbb, 0x79, 0xed, 0x2b, 0x66, 0xe1, 0x64, 0x36, 0x57, 0x7a, 0x4c, 0x49, 0xd4, - 0xaa, 0x4a, 0xa5, 0xf3, 0x36, 0x7a, 0x05, 0x6a, 0xd6, 0x64, 0x12, 0xd8, 0xa6, 0xe3, 0x46, 0xad, - 0x9a, 0x60, 0x54, 0x05, 0x61, 0xcb, 0x8d, 0xda, 0xef, 0x41, 0x35, 0x51, 0xd0, 0x5c, 0xd9, 0xe4, - 0x5f, 0x0b, 0x50, 0x4b, 0x15, 0x82, 0x3e, 0x85, 0xd5, 0xc8, 0x7a, 0x6e, 0x4e, 0x35, 0x2b, 0x8f, - 0xaf, 0x77, 0xf3, 0x6a, 0xd6, 0x7a, 0x3e, 0x55, 0x6e, 0x23, 0xca, 0xf4, 0xd0, 0xe7, 0xb0, 0x36, - 0x71, 0xfd, 0xf8, 0x2c, 0x83, 0x2d, 0xe3, 0xc1, 0xff, 0xe7, 0xc4, 0xde, 0xe5, 0xb3, 0xa7, 0xe8, - 0xcd, 0xc9, 0x4c, 0x5f, 0xff, 0x43, 0x01, 0x1a, 0xd9, 0xe5, 0xb9, 0x12, 0xec, 0x30, 0x16, 0x3f, - 0x50, 0xc2, 0xbc, 0xc9, 0x8f, 0x0c, 0x8f, 0x78, 0x41, 0x74, 0x2e, 0x56, 0x2e, 0x61, 0xd5, 0xe3, - 0xba, 0x76, 0x5c, 0x7a, 0x22, 0x7c, 0xb7, 0x84, 0x45, 0x9b, 0xd3, 0xdc, 0x20, 0xa4, 0x22, 0x9b, - 0x2a, 0x61, 0xd1, 0x46, 0x18, 0xaa, 0x2a, 0x50, 0x70, 0x8f, 0x2b, 0xcd, 0x1f, 0x70, 0x12, 0xe1, - 0x70, 0x8a, 0xa3, 0xff, 0xa6, 0x08, 0x6b, 0x17, 0xb8, 0x5c, 0x4e, 0xe9, 0x06, 0xc9, 0x71, 0x2b, - 0x7b, 0x5c, 0x26, 0xdb, 0x75, 0x92, 0x9c, 0x46, 0xb4, 0xc5, 0x66, 0x0d, 0x55, 0xbe, 0x51, 0x74, - 0x43, 0x6e, 0x68, 0xef, 0xc8, 0x65, 0x52, 0xf0, 0x32, 0x96, 0x1d, 0xf4, 0x0c, 0x9a, 0x11, 0xa1, - 0x24, 0x3a, 0x25, 0x8e, 0x19, 0x06, 0x11, 0x4b, 0xe4, 0xdf, 0x9c, 0x4f, 0xfe, 0x83, 0x20, 0x62, - 0x78, 0x35, 0x41, 0xe2, 0x3d, 0x8a, 0x9e, 0xc2, 0xaa, 0x73, 0xee, 0x5b, 0x9e, 0x6b, 0x2b, 0xe4, - 0xca, 0xc2, 0xc8, 0x0d, 0x05, 0x24, 0x80, 0xf9, 0x35, 0x27, 0xc3, 0xe4, 0x3f, 0x36, 0xb1, 0x8e, - 0xc8, 0x44, 0xe9, 0x44, 0x76, 0x66, 0xfd, 0xba, 0xac, 0xfc, 0x5a, 0xff, 0x49, 0x11, 0x9a, 0xb3, - 0xee, 0x82, 0xfe, 0x17, 0xc0, 0x0e, 0x63, 0x33, 0x24, 0x91, 0x1b, 0x38, 0xca, 0x29, 0x6a, 0x76, - 0x18, 0x1f, 0x08, 0x02, 0xdf, 0x5a, 0x9c, 0xfd, 0x65, 0x1c, 0x30, 0x4b, 0x79, 0x47, 0xd5, 0x0e, - 0xe3, 0x6f, 0xf2, 0x7e, 0x32, 0x57, 0xdc, 0xc3, 0xa9, 0xf2, 0x12, 0x3e, 0x7c, 0x20, 0x08, 0xe8, - 0x1d, 0xb8, 0x2d, 0x1d, 0xc9, 0x9c, 0xb8, 0x9e, 0xcb, 0x4c, 0xd7, 0x37, 0x8f, 0xce, 0xf9, 0xa5, - 0x56, 0xfa, 0x0e, 0x92, 0xcc, 0x5d, 0xce, 0xdb, 0xf1, 0x3f, 0xe6, 0x1c, 0xa4, 0xc3, 0x6a, 0x10, - 0x78, 0x26, 0xb5, 0x83, 0x88, 0x98, 0x96, 0xf3, 0x85, 0x38, 0xc0, 0x4a, 0xb8, 0x1e, 0x04, 0xde, - 0x80, 0xd3, 0xba, 0xce, 0x17, 0xe8, 0x55, 0xa8, 0xdb, 0x61, 0x4c, 0x09, 0x33, 0xf9, 0xa7, 0x55, - 0x11, 0xbf, 0x0d, 0x92, 0xd4, 0x0b, 0x63, 0x9a, 0x19, 0xe0, 0x11, 0x8f, 0x9f, 0x37, 0x99, 0x01, - 0x7b, 0xc4, 0xa3, 0xfa, 0x67, 0x50, 0x16, 0xa7, 0x13, 0xff, 0x3b, 0x11, 0xc0, 0x43, 0x8b, 0x8d, - 0x95, 0xfe, 0xaa, 0x9c, 0x70, 0x60, 0xb1, 0x31, 0x67, 0x8e, 0x03, 0xca, 0x24, 0x53, 0xba, 0x56, - 0x95, 0x13, 0x04, 0xb3, 0x0d, 0xd5, 0x88, 0x58, 0x4e, 0xe0, 0x4f, 0xce, 0xc5, 0x8f, 0x57, 0x71, - 0xda, 0xd7, 0xbf, 0x84, 0x8a, 0x3c, 0xb5, 0xae, 0x81, 0x7f, 0x17, 0x90, 0x3d, 0x8a, 0x82, 0x38, - 0xe4, 0x96, 0xf1, 0x5c, 0x4a, 0xdd, 0xc0, 0xa7, 0xc9, 0xcd, 0x59, 0x72, 0x0e, 0xa6, 0x0c, 0xfd, - 0x2f, 0x05, 0x19, 0xa9, 0xe4, 0x35, 0x88, 0xc7, 0x7a, 0x15, 0x76, 0x16, 0xbe, 0x2b, 0x2a, 0x80, - 0x24, 0x5f, 0x23, 0xea, 0x85, 0x60, 0xde, 0x7c, 0x8d, 0xc8, 0x7c, 0x8d, 0xf0, 0xe4, 0x40, 0x05, - 0x44, 0x09, 0x27, 0xe3, 0x61, 0xdd, 0x49, 0x13, 0x59, 0xa2, 0xff, 0xb3, 0x90, 0x6e, 0xf9, 0x24, - 0xe1, 0x44, 0x9f, 0x43, 0x95, 0xef, 0x1e, 0xd3, 0xb3, 0x42, 0xf5, 0x16, 0xd2, 0x5b, 0x2c, 0x97, - 0xed, 0xf0, 0xcd, 0xb2, 0x67, 0x85, 0x32, 0x52, 0xae, 0x84, 0xb2, 0xc7, 0x8f, 0x0e, 0xcb, 0x99, - 0x1e, 0x1d, 0xbc, 0x8d, 0x5e, 0x87, 0xa6, 0x15, 0xb3, 0xc0, 0xb4, 0x9c, 0x53, 0x12, 0x31, 0x97, - 0x12, 0x65, 0xe1, 0x55, 0x4e, 0xed, 0x26, 0xc4, 0xf6, 0x43, 0x68, 0x64, 0x31, 0xbf, 0x2e, 0xb8, - 0x94, 0xb3, 0xc1, 0xe5, 0xbb, 0x00, 0xd3, 0xbc, 0x8a, 0x7b, 0x02, 0x39, 0x73, 0x99, 0x69, 0x07, - 0x8e, 0x3c, 0xda, 0xca, 0xb8, 0xca, 0x09, 0xbd, 0xc0, 0x21, 0x17, 0xb2, 0xd4, 0x72, 0x92, 0xa5, - 0xf2, 0xcd, 0xc7, 0xb7, 0xca, 0x89, 0x3b, 0x99, 0x10, 0x47, 0x49, 0x58, 0x0b, 0x02, 0xef, 0x89, - 0x20, 0xe8, 0x7f, 0x2e, 0x4a, 0x8f, 0x90, 0x77, 0x84, 0x5c, 0xb9, 0x4b, 0x6a, 0xea, 0xd2, 0xf5, - 0x4c, 0xfd, 0x00, 0x80, 0x32, 0x2b, 0x62, 0xc4, 0x31, 0x2d, 0xa6, 0xae, 0xdd, 0xed, 0x4b, 0x69, - 0xee, 0x30, 0x79, 0x6e, 0xc4, 0x35, 0x35, 0xba, 0xcb, 0xd0, 0xfb, 0xd0, 0xb0, 0x03, 0x2f, 0x9c, - 0x10, 0x35, 0xb9, 0xfc, 0xb5, 0x93, 0xeb, 0xe9, 0xf8, 0x2e, 0xcb, 0xe4, 0xb8, 0x95, 0xeb, 0xe6, - 0xb8, 0x7f, 0x2c, 0xc8, 0xab, 0x4e, 0xf6, 0xa6, 0x85, 0x46, 0x57, 0xbc, 0xcd, 0x3d, 0x5e, 0xf0, - 0xda, 0xf6, 0x55, 0x0f, 0x73, 0xd7, 0x7d, 0x09, 0xfb, 0x53, 0x09, 0x6a, 0xe9, 0x8d, 0xe9, 0x92, - 0xed, 0xef, 0x43, 0x2d, 0x7d, 0xeb, 0x55, 0xb9, 0xc6, 0x57, 0x9a, 0x27, 0x1d, 0x8c, 0x8e, 0x01, - 0x59, 0xa3, 0x51, 0x9a, 0xa9, 0x98, 0x31, 0xb5, 0x46, 0xc9, 0x1d, 0xf3, 0xfe, 0x1c, 0x7a, 0x48, - 0xc2, 0xcf, 0x21, 0x9f, 0x8f, 0x35, 0x6b, 0x34, 0x9a, 0xa1, 0xa0, 0xef, 0xc1, 0xed, 0xd9, 0x35, - 0xcc, 0xa3, 0x73, 0x33, 0x74, 0x1d, 0x95, 0x23, 0x6f, 0xcf, 0x7b, 0x69, 0xec, 0xcc, 0xc0, 0x7f, - 0x7c, 0x7e, 0xe0, 0x3a, 0x52, 0xe7, 0x28, 0xba, 0xc4, 0x68, 0xff, 0x00, 0x5e, 0x7a, 0xc1, 0xf0, - 0x2b, 0x6c, 0xd0, 0xcf, 0xda, 0xe0, 0x3a, 0x4a, 0xc8, 0x58, 0xef, 0xb7, 0x05, 0x79, 0xb7, 0x9d, - 0xd5, 0x49, 0x77, 0x9a, 0xb6, 0xd5, 0x37, 0x37, 0x72, 0xae, 0xd3, 0x3b, 0x38, 0x94, 0xf0, 0x22, - 0xcf, 0xfb, 0x64, 0x26, 0xcf, 0xcb, 0x9f, 0x8b, 0xec, 0x89, 0x49, 0x12, 0x48, 0x21, 0xe8, 0xbf, - 0x2f, 0x41, 0x35, 0x41, 0xe7, 0x11, 0x97, 0x9e, 0x53, 0x46, 0x3c, 0xd3, 0x4b, 0x8e, 0xb0, 0x02, - 0x06, 0x49, 0xda, 0xe3, 0x87, 0xd8, 0x2b, 0x50, 0xe3, 0x99, 0xba, 0x64, 0x17, 0x05, 0xbb, 0xca, - 0x09, 0x82, 0xf9, 0x2a, 0xd4, 0x59, 0xc0, 0xac, 0x89, 0xc9, 0x5c, 0xfb, 0x44, 0x06, 0xb9, 0x02, - 0x06, 0x41, 0x1a, 0x72, 0x0a, 0x7a, 0x13, 0x6e, 0xb0, 0x71, 0x14, 0x30, 0x36, 0xe1, 0x69, 0x9a, - 0x48, 0x4c, 0x64, 0x12, 0xb1, 0x8c, 0xb5, 0x94, 0x21, 0x13, 0x16, 0xca, 0x4f, 0xef, 0xe9, 0x60, - 0xee, 0xba, 0xe2, 0x10, 0x59, 0xc6, 0xab, 0x29, 0x95, 0xbb, 0x36, 0xbf, 0xa2, 0x87, 0x24, 0xb2, - 0x89, 0x2f, 0xcf, 0x8a, 0x02, 0x4e, 0xba, 0xc8, 0x84, 0x35, 0x8f, 0x58, 0x34, 0x8e, 0x88, 0x63, - 0x1e, 0xbb, 0x64, 0xe2, 0xc8, 0x2b, 0x4b, 0x33, 0x77, 0x52, 0x9b, 0xa8, 0xa5, 0xf3, 0x48, 0xcc, - 0xc6, 0xcd, 0x04, 0x4e, 0xf6, 0x79, 0x7e, 0x20, 0x5b, 0x68, 0x0d, 0xea, 0x83, 0x67, 0x83, 0xa1, - 0xb1, 0x67, 0xee, 0xed, 0x6f, 0x19, 0xea, 0x8d, 0x7a, 0x60, 0x60, 0xd9, 0x2d, 0x70, 0xfe, 0x70, - 0x7f, 0xd8, 0xdd, 0x35, 0x87, 0x3b, 0xbd, 0x27, 0x03, 0xad, 0x88, 0x6e, 0xc3, 0x8d, 0xe1, 0x36, - 0xde, 0x1f, 0x0e, 0x77, 0x8d, 0x2d, 0xf3, 0xc0, 0xc0, 0x3b, 0xfb, 0x5b, 0x03, 0xad, 0x84, 0x10, - 0x34, 0xa7, 0xe4, 0xe1, 0xce, 0x9e, 0xa1, 0x2d, 0xa3, 0x3a, 0xac, 0x1c, 0x18, 0xb8, 0x67, 0xf4, - 0x87, 0x5a, 0x59, 0xff, 0x5b, 0x11, 0xea, 0x19, 0x2b, 0x72, 0x47, 0x8e, 0xa8, 0xbc, 0xc4, 0x2c, - 0x63, 0xde, 0xe4, 0x87, 0x89, 0x6d, 0xd9, 0x63, 0x69, 0x9d, 0x65, 0x2c, 0x3b, 0xdc, 0x6e, 0x9e, - 0x75, 0x96, 0xd9, 0xe7, 0xcb, 0xb8, 0xea, 0x59, 0x67, 0x12, 0xe4, 0x35, 0x68, 0x9c, 0x90, 0xc8, - 0x27, 0x13, 0xc5, 0x97, 0x16, 0xa9, 0x4b, 0x9a, 0x1c, 0xb2, 0x0e, 0x9a, 0x1a, 0x32, 0x85, 0x91, - 0xe6, 0x68, 0x4a, 0xfa, 0x5e, 0x02, 0x76, 0x74, 0x59, 0xeb, 0x15, 0xa1, 0xf5, 0x07, 0xf3, 0x3b, - 0xe9, 0x8b, 0x14, 0x3f, 0x48, 0x15, 0xbf, 0x02, 0x25, 0x9c, 0x3c, 0xd7, 0xf6, 0xba, 0xbd, 0x6d, - 0xae, 0xec, 0x55, 0xa8, 0xed, 0x75, 0x3f, 0x35, 0x0f, 0x07, 0xe2, 0xf2, 0x8f, 0x34, 0x68, 0x3c, - 0x31, 0x70, 0xdf, 0xd8, 0x55, 0x94, 0x12, 0xba, 0x05, 0x9a, 0xa2, 0x4c, 0xc7, 0x2d, 0xeb, 0xbf, - 0x2b, 0xc2, 0x9a, 0x3c, 0xd7, 0xd3, 0xf7, 0xa8, 0x17, 0x3f, 0x0c, 0x2d, 0x7e, 0xf4, 0xb6, 0x60, - 0xc5, 0x23, 0x34, 0xb5, 0x43, 0x0d, 0x27, 0x5d, 0xe4, 0x42, 0xdd, 0xf2, 0xfd, 0x80, 0x89, 0x47, - 0x0d, 0xaa, 0x8e, 0xc8, 0xc7, 0x73, 0x3d, 0x9f, 0xa4, 0x92, 0x77, 0xba, 0x53, 0x24, 0x79, 0x42, - 0x66, 0xb1, 0xdb, 0x1f, 0x80, 0x76, 0x71, 0xc0, 0x3c, 0x71, 0xe9, 0x8d, 0x77, 0xa6, 0x61, 0x89, - 0x70, 0x07, 0x3d, 0xec, 0x3f, 0xe9, 0xef, 0x3f, 0xed, 0x6b, 0x4b, 0xbc, 0x83, 0x0f, 0xfb, 0xfd, - 0x9d, 0xfe, 0x63, 0xad, 0x80, 0x00, 0x2a, 0xc6, 0xa7, 0x3b, 0x43, 0x63, 0x4b, 0x2b, 0x6e, 0xfe, - 0x7d, 0x15, 0x2a, 0x52, 0x48, 0xf4, 0x6b, 0x15, 0x92, 0xb3, 0x45, 0x43, 0xf4, 0xc1, 0xdc, 0xa9, - 0xed, 0x4c, 0x21, 0xb2, 0xfd, 0xe1, 0xc2, 0xf3, 0xd5, 0xbb, 0xe0, 0x12, 0xfa, 0x79, 0x01, 0x1a, - 0x33, 0x0f, 0x61, 0x79, 0xdf, 0x70, 0xae, 0xa8, 0x51, 0xb6, 0xbf, 0xb1, 0xd0, 0xdc, 0x54, 0x96, - 0x9f, 0x15, 0xa0, 0x9e, 0xa9, 0xce, 0xa1, 0x07, 0x8b, 0x54, 0xf4, 0xa4, 0x24, 0x0f, 0x17, 0x2f, - 0x06, 0xea, 0x4b, 0x6f, 0x17, 0xd0, 0x4f, 0x0b, 0x50, 0xcf, 0x94, 0xb6, 0x72, 0x8b, 0x72, 0xb9, - 0x10, 0x97, 0x5b, 0x94, 0xab, 0x2a, 0x69, 0x4b, 0xe8, 0x87, 0x05, 0xa8, 0xa5, 0x65, 0x2a, 0x74, - 0x6f, 0xfe, 0xc2, 0x96, 0x14, 0xe2, 0xfe, 0xa2, 0x15, 0x31, 0x7d, 0x09, 0x7d, 0x1f, 0xaa, 0x49, - 0x4d, 0x07, 0xe5, 0x0d, 0x23, 0x17, 0x0a, 0x46, 0xed, 0x7b, 0x73, 0xcf, 0xcb, 0x2e, 0x9f, 0x14, - 0x5a, 0x72, 0x2f, 0x7f, 0xa1, 0x24, 0xd4, 0xbe, 0x37, 0xf7, 0xbc, 0x74, 0x79, 0xee, 0x09, 0x99, - 0x7a, 0x4c, 0x6e, 0x4f, 0xb8, 0x5c, 0x08, 0xca, 0xed, 0x09, 0x57, 0x95, 0x7f, 0xa4, 0x20, 0x99, - 0x8a, 0x4e, 0x6e, 0x41, 0x2e, 0x57, 0x8d, 0x72, 0x0b, 0x72, 0x45, 0x01, 0x49, 0xb9, 0xe4, 0x34, - 0x41, 0xbf, 0x37, 0x77, 0x11, 0x64, 0x4e, 0x97, 0xbc, 0x54, 0x86, 0xd1, 0x97, 0xd0, 0x8f, 0xd4, - 0x93, 0x81, 0xac, 0xa0, 0xa0, 0x79, 0xa0, 0x66, 0x8a, 0x2e, 0xed, 0xf7, 0x16, 0x0b, 0x35, 0xe2, - 0x8c, 0xf8, 0x71, 0x01, 0x60, 0x5a, 0x6b, 0xc9, 0x2d, 0xc4, 0xa5, 0x22, 0x4f, 0xfb, 0xc1, 0x02, - 0x33, 0xb3, 0xdb, 0x23, 0x29, 0xaf, 0xe4, 0xde, 0x1e, 0x17, 0x6a, 0x41, 0xb9, 0xb7, 0xc7, 0xc5, - 0x3a, 0x8e, 0xbe, 0xf4, 0xf1, 0xca, 0xb7, 0xca, 0x32, 0xf6, 0x57, 0xc4, 0xe7, 0xdd, 0xff, 0x04, - 0x00, 0x00, 0xff, 0xff, 0x96, 0x01, 0xe4, 0xc4, 0xf0, 0x23, 0x00, 0x00, + 0xd1, 0x17, 0x9f, 0x22, 0x8b, 0x14, 0x35, 0xdb, 0xbb, 0xfb, 0x99, 0xa6, 0xf1, 0x7d, 0x5e, 0x0f, + 0x60, 0x40, 0xb0, 0xbd, 0x94, 0x2d, 0xe3, 0xf3, 0x3e, 0x02, 0x3f, 0x68, 0x72, 0x76, 0x25, 0xaf, + 0x44, 0x29, 0x4d, 0x0a, 0xeb, 0x4d, 0x62, 0x4f, 0x46, 0x33, 0x2d, 0x72, 0x2c, 0xce, 0xc3, 0xdd, + 0x43, 0xad, 0x84, 0x20, 0x48, 0x90, 0x04, 0x41, 0x12, 0x20, 0x40, 0x2e, 0x41, 0xee, 0xb9, 0x05, + 0xb9, 0xe6, 0x94, 0xc7, 0x25, 0x40, 0xfe, 0x87, 0x1c, 0x73, 0x09, 0x90, 0x6b, 0xfe, 0x83, 0xa0, + 0x1f, 0x33, 0x1c, 0x4a, 0x5a, 0x7b, 0x48, 0xe5, 0xc4, 0xa9, 0xea, 0xee, 0x5f, 0x17, 0xab, 0xaa, + 0xab, 0xaa, 0xbb, 0xa0, 0xee, 0x50, 0xf7, 0x94, 0xd0, 0x76, 0x48, 0x83, 0x28, 0x40, 0xaf, 0x8f, + 0x2d, 0x36, 0x76, 0xed, 0x80, 0x86, 0x6d, 0x3f, 0xf0, 0x2c, 0xa7, 0x1d, 0x4e, 0xa6, 0x23, 0xd7, + 0x67, 0x6d, 0x39, 0x8b, 0xc9, 0x69, 0xad, 0xff, 0x1b, 0x05, 0xc1, 0x68, 0x42, 0x36, 0x05, 0x75, + 0x34, 0x3d, 0xde, 0x74, 0xa6, 0xd4, 0x8a, 0xdc, 0xc0, 0x57, 0xe3, 0xaf, 0x5e, 0x1c, 0x8f, 0x5c, + 0x8f, 0xb0, 0xc8, 0xf2, 0x42, 0x35, 0xe1, 0xa3, 0x91, 0x1b, 0x8d, 0xa7, 0x47, 0x6d, 0x3b, 0xf0, + 0x36, 0x93, 0x2d, 0x37, 0xc5, 0x96, 0x9b, 0x6a, 0xcb, 0x4d, 0x36, 0xb6, 0x28, 0x71, 0x36, 0xc7, + 0xf6, 0x84, 0x85, 0xc4, 0xe6, 0xbf, 0x26, 0xff, 0x90, 0x08, 0xfa, 0xcb, 0xf0, 0xd2, 0xd0, 0x62, + 0x27, 0xdd, 0xc0, 0x3f, 0x76, 0x47, 0x03, 0x7b, 0x4c, 0x3c, 0x0b, 0x93, 0x2f, 0xa7, 0x84, 0x45, + 0xfa, 0x77, 0xa0, 0x79, 0x79, 0x88, 0x85, 0x81, 0xcf, 0x08, 0xfa, 0x08, 0x8a, 0x1c, 0xa4, 0x99, + 0xbb, 0x93, 0xdb, 0xa8, 0x6d, 0xbd, 0xd5, 0x7e, 0xd1, 0xff, 0x95, 0x9b, 0xb7, 0xd5, 0xe6, 0xed, + 0x41, 0x48, 0x6c, 0x2c, 0x56, 0xea, 0xb7, 0xe1, 0x66, 0xd7, 0x0a, 0xad, 0x23, 0x77, 0xe2, 0x46, + 0x2e, 0x61, 0xf1, 0xa6, 0x53, 0xb8, 0x35, 0xcf, 0x56, 0x1b, 0x7e, 0x06, 0x75, 0x3b, 0xc5, 0x57, + 0x1b, 0x3f, 0x68, 0x67, 0x52, 0x74, 0xbb, 0x27, 0xa8, 0x39, 0xe0, 0x39, 0x38, 0xfd, 0x16, 0xa0, + 0x47, 0xae, 0x3f, 0x22, 0x34, 0xa4, 0xae, 0x1f, 0xc5, 0xc2, 0xfc, 0xa2, 0x00, 0x37, 0xe7, 0xd8, + 0x4a, 0x98, 0x2f, 0x00, 0xac, 0x28, 0xa2, 0xee, 0xd1, 0x34, 0x12, 0xa2, 0x14, 0x36, 0x6a, 0x5b, + 0x9f, 0x64, 0x14, 0xe5, 0x0a, 0xbc, 0x76, 0x27, 0x01, 0x33, 0xfc, 0x88, 0x9e, 0xe3, 0x14, 0x3a, + 0xfa, 0x1c, 0xca, 0x63, 0x62, 0x4d, 0xa2, 0x71, 0x33, 0x7f, 0x27, 0xb7, 0xd1, 0xd8, 0x7a, 0x74, + 0x8d, 0x7d, 0xb6, 0x05, 0xd0, 0x20, 0xb2, 0x22, 0x82, 0x15, 0x2a, 0xba, 0x0b, 0x48, 0x7e, 0x99, + 0x0e, 0x61, 0x36, 0x75, 0x43, 0xee, 0x7f, 0xcd, 0xc2, 0x9d, 0xdc, 0x46, 0x15, 0xdf, 0x90, 0x23, + 0xbd, 0xd9, 0x40, 0xeb, 0x7d, 0x58, 0xbf, 0x20, 0x2d, 0xd2, 0xa0, 0x70, 0x42, 0xce, 0x85, 0x45, + 0xaa, 0x98, 0x7f, 0xa2, 0x5b, 0x50, 0x3a, 0xb5, 0x26, 0x53, 0x22, 0x44, 0xae, 0x62, 0x49, 0x3c, + 0xcc, 0xdf, 0xcf, 0xe9, 0x0f, 0xa0, 0x96, 0x12, 0x02, 0x35, 0x00, 0x0e, 0xfb, 0x3d, 0x63, 0x68, + 0x74, 0x87, 0x46, 0x4f, 0x5b, 0x41, 0x6b, 0x50, 0x3d, 0xec, 0x6f, 0x1b, 0x9d, 0xdd, 0xe1, 0xf6, + 0x33, 0x2d, 0x87, 0x6a, 0xb0, 0x1a, 0x13, 0x79, 0xfd, 0x0c, 0x10, 0x26, 0x76, 0x70, 0x4a, 0x28, + 0xf7, 0x4a, 0x65, 0x22, 0xf4, 0x12, 0xac, 0x46, 0x16, 0x3b, 0x31, 0x5d, 0x47, 0x09, 0x50, 0xe6, + 0xe4, 0x8e, 0x83, 0x76, 0xa0, 0x3c, 0xb6, 0x7c, 0x67, 0x22, 0x85, 0xa8, 0x6d, 0xbd, 0x93, 0x51, + 0x6f, 0x1c, 0x7c, 0x5b, 0x2c, 0xc4, 0x0a, 0x80, 0xbb, 0xea, 0xdc, 0xce, 0x52, 0x9b, 0xfa, 0x33, + 0xd0, 0x06, 0x91, 0x45, 0xa3, 0xb4, 0x38, 0x06, 0x14, 0xf9, 0xfe, 0xca, 0x3d, 0x17, 0xd9, 0x53, + 0x1e, 0x33, 0x2c, 0x96, 0xeb, 0xff, 0xce, 0xc3, 0x8d, 0x14, 0xb6, 0x72, 0xbb, 0xa7, 0x50, 0xa6, + 0x84, 0x4d, 0x27, 0x91, 0x80, 0x6f, 0x6c, 0x7d, 0x98, 0x11, 0xfe, 0x12, 0x52, 0x1b, 0x0b, 0x18, + 0xac, 0xe0, 0xd0, 0x06, 0x68, 0x72, 0x85, 0x49, 0x28, 0x0d, 0xa8, 0xe9, 0xb1, 0x91, 0x32, 0x5d, + 0x43, 0xf2, 0x0d, 0xce, 0xde, 0x63, 0xa3, 0x94, 0x56, 0x0b, 0xd7, 0xd4, 0x2a, 0xb2, 0x40, 0xf3, + 0x49, 0xf4, 0x3c, 0xa0, 0x27, 0x26, 0x57, 0x2d, 0x75, 0x1d, 0xd2, 0x2c, 0x0a, 0xd0, 0xf7, 0x32, + 0x82, 0xf6, 0xe5, 0xf2, 0x7d, 0xb5, 0x1a, 0xaf, 0xfb, 0xf3, 0x0c, 0xfd, 0x4d, 0x28, 0xcb, 0x7f, + 0xca, 0x3d, 0x69, 0x70, 0xd8, 0xed, 0x1a, 0x83, 0x81, 0xb6, 0x82, 0xaa, 0x50, 0xc2, 0xc6, 0x10, + 0x73, 0x0f, 0xab, 0x42, 0xe9, 0x51, 0x67, 0xd8, 0xd9, 0xd5, 0xf2, 0xfa, 0x1b, 0xb0, 0xfe, 0xd4, + 0x72, 0xa3, 0x2c, 0xce, 0xa5, 0x07, 0xa0, 0xcd, 0xe6, 0x2a, 0xeb, 0xec, 0xcc, 0x59, 0x27, 0xbb, + 0x6a, 0x8c, 0x33, 0x37, 0xba, 0x60, 0x0f, 0x0d, 0x0a, 0x84, 0x52, 0x65, 0x02, 0xfe, 0xa9, 0x3f, + 0x87, 0xf5, 0x41, 0x14, 0x84, 0x99, 0x3c, 0xff, 0x5d, 0x58, 0xe5, 0x79, 0x22, 0x98, 0x46, 0xca, + 0xf5, 0x5f, 0x6e, 0xcb, 0x3c, 0xd2, 0x8e, 0xf3, 0x48, 0xbb, 0xa7, 0xf2, 0x0c, 0x8e, 0x67, 0xa2, + 0xff, 0x81, 0x32, 0x73, 0x47, 0xbe, 0x35, 0x51, 0x47, 0x5f, 0x51, 0x3a, 0xe2, 0x4e, 0x1e, 0x6f, + 0xac, 0x1c, 0xbf, 0x0b, 0xa8, 0x47, 0x58, 0x44, 0x83, 0xf3, 0x4c, 0xf2, 0xdc, 0x82, 0xd2, 0x71, + 0x40, 0x6d, 0x79, 0x10, 0x2b, 0x58, 0x12, 0xfc, 0x50, 0xcd, 0x81, 0x28, 0xec, 0xbb, 0x80, 0x76, + 0x7c, 0x9e, 0x20, 0xb2, 0x19, 0xe2, 0x57, 0x79, 0xb8, 0x39, 0x37, 0x5f, 0x19, 0x63, 0xf9, 0x73, + 0xc8, 0x03, 0xd3, 0x94, 0xc9, 0x73, 0x88, 0xf6, 0xa1, 0x2c, 0x67, 0x28, 0x4d, 0xde, 0x5b, 0x00, + 0x48, 0xe6, 0x1c, 0x05, 0xa7, 0x60, 0xae, 0x74, 0xfa, 0xc2, 0x7f, 0xdb, 0xe9, 0xb5, 0xf8, 0x7f, + 0xb0, 0xaf, 0xd5, 0xdf, 0xb7, 0xe1, 0x46, 0x6a, 0xb2, 0x52, 0xde, 0x23, 0x28, 0x31, 0xce, 0x50, + 0xda, 0x7b, 0x7b, 0x41, 0xed, 0x31, 0x2c, 0x97, 0xeb, 0x37, 0x25, 0xb8, 0x71, 0x4a, 0xfc, 0x44, + 0x14, 0xbd, 0x07, 0x37, 0x06, 0xc2, 0xb5, 0x32, 0xf9, 0xce, 0xcc, 0x2d, 0xf3, 0x73, 0x6e, 0x79, + 0x0b, 0x50, 0x1a, 0x45, 0x39, 0xcf, 0x39, 0xac, 0x1b, 0x67, 0xc4, 0xce, 0x84, 0xdc, 0x84, 0x55, + 0x3b, 0xf0, 0x3c, 0xcb, 0x77, 0x9a, 0xf9, 0x3b, 0x85, 0x8d, 0x2a, 0x8e, 0xc9, 0xf4, 0xf9, 0x29, + 0x64, 0x3d, 0x3f, 0xfa, 0x2f, 0x73, 0xa0, 0xcd, 0xf6, 0x56, 0x8a, 0xe4, 0xd2, 0x47, 0x0e, 0x07, + 0xe2, 0x7b, 0xd7, 0xb1, 0xa2, 0x14, 0x3f, 0x3e, 0xe2, 0x92, 0x4f, 0x28, 0x4d, 0x85, 0x90, 0xc2, + 0x35, 0x43, 0x88, 0xfe, 0xcf, 0x1c, 0xa0, 0xcb, 0x55, 0x0f, 0x7a, 0x0d, 0xea, 0x8c, 0xf8, 0x8e, + 0x29, 0xd5, 0x28, 0x2d, 0x5c, 0xc1, 0x35, 0xce, 0x93, 0xfa, 0x64, 0x08, 0x41, 0x91, 0x9c, 0x11, + 0x5b, 0x9d, 0x56, 0xf1, 0x8d, 0xc6, 0x50, 0x3f, 0x66, 0xa6, 0xcb, 0x82, 0x89, 0x95, 0x94, 0x07, + 0x8d, 0x2d, 0x63, 0xe9, 0xea, 0xab, 0xfd, 0x68, 0xb0, 0x13, 0x83, 0xe1, 0xda, 0x31, 0x4b, 0x08, + 0xbd, 0x0d, 0xb5, 0xd4, 0x18, 0xaa, 0x40, 0xb1, 0xbf, 0xdf, 0x37, 0xb4, 0x15, 0x04, 0x50, 0xee, + 0x6e, 0xe3, 0xfd, 0xfd, 0xa1, 0x8c, 0xda, 0x3b, 0x7b, 0x9d, 0xc7, 0x86, 0x96, 0xd7, 0xff, 0x5c, + 0x04, 0x98, 0xa5, 0x4f, 0xd4, 0x80, 0x7c, 0x62, 0xe9, 0xbc, 0xeb, 0xf0, 0x3f, 0xe3, 0x5b, 0x5e, + 0x5c, 0x88, 0x88, 0x6f, 0xb4, 0x05, 0xb7, 0x3d, 0x36, 0x0a, 0x2d, 0xfb, 0xc4, 0x54, 0x59, 0xcf, + 0x16, 0x8b, 0xc5, 0xbf, 0xaa, 0xe3, 0x9b, 0x6a, 0x50, 0x49, 0x2d, 0x71, 0x77, 0xa1, 0x40, 0xfc, + 0xd3, 0x66, 0x51, 0x94, 0x7a, 0x0f, 0x17, 0x4e, 0xeb, 0x6d, 0xc3, 0x3f, 0x95, 0xa5, 0x1d, 0x87, + 0x41, 0x7d, 0xa8, 0x52, 0xc2, 0x82, 0x29, 0xb5, 0x09, 0x6b, 0x96, 0x16, 0x3a, 0x64, 0x38, 0x5e, + 0x87, 0x67, 0x10, 0xa8, 0x07, 0x65, 0x2f, 0x98, 0xfa, 0x11, 0x6b, 0x96, 0x85, 0x80, 0x6f, 0x65, + 0x04, 0xdb, 0xe3, 0x8b, 0xb0, 0x5a, 0x8b, 0x1e, 0xc3, 0xaa, 0x43, 0x4e, 0x5d, 0x2e, 0xd3, 0xaa, + 0x80, 0xb9, 0x9b, 0xd5, 0xbe, 0x62, 0x15, 0x8e, 0x57, 0x73, 0xa5, 0x4f, 0x19, 0xa1, 0xcd, 0x8a, + 0x54, 0x3a, 0xff, 0x46, 0xaf, 0x40, 0xd5, 0x9a, 0x4c, 0x02, 0xdb, 0x74, 0x5c, 0xda, 0xac, 0x8a, + 0x81, 0x8a, 0x60, 0xf4, 0x5c, 0x8a, 0x5e, 0x85, 0x9a, 0x3c, 0x19, 0x66, 0x68, 0x45, 0xe3, 0x26, + 0x88, 0x61, 0x90, 0xac, 0x03, 0x2b, 0x1a, 0xab, 0x09, 0x84, 0x52, 0x39, 0xa1, 0x96, 0x4c, 0x20, + 0x94, 0xf2, 0x09, 0xad, 0xf7, 0xa0, 0x12, 0xab, 0x78, 0xa1, 0x7a, 0xf4, 0x6f, 0x39, 0xa8, 0x26, + 0x2a, 0x45, 0x9f, 0xc2, 0x1a, 0xb5, 0x9e, 0x9b, 0x33, 0xdb, 0xc8, 0x00, 0xf8, 0x6e, 0x56, 0xdb, + 0x58, 0xcf, 0x67, 0xe6, 0xa9, 0xd3, 0x14, 0x85, 0x3e, 0x87, 0xf5, 0x89, 0xeb, 0x4f, 0xcf, 0x52, + 0xd8, 0x32, 0xa3, 0xfc, 0x7f, 0x46, 0xec, 0x5d, 0xbe, 0x7a, 0x86, 0xde, 0x98, 0xcc, 0xd1, 0xfa, + 0x1f, 0x72, 0x50, 0x4f, 0x6f, 0xcf, 0x95, 0x60, 0x87, 0x53, 0xf1, 0x07, 0x0a, 0x98, 0x7f, 0xf2, + 0xa0, 0xe3, 0x11, 0x2f, 0xa0, 0xe7, 0x62, 0xe7, 0x02, 0x56, 0x14, 0xb7, 0x96, 0xe3, 0xb2, 0x13, + 0xe1, 0xfd, 0x05, 0x2c, 0xbe, 0x39, 0xcf, 0x0d, 0x42, 0x26, 0xea, 0xb1, 0x02, 0x16, 0xdf, 0x08, + 0x43, 0x45, 0xa5, 0x1a, 0xee, 0xb3, 0x85, 0xc5, 0x53, 0x56, 0x2c, 0x1c, 0x4e, 0x70, 0xf4, 0xdf, + 0xe4, 0x61, 0xfd, 0xc2, 0x28, 0x97, 0x53, 0x3a, 0x52, 0x1c, 0xb0, 0x25, 0xc5, 0x65, 0xb2, 0x5d, + 0x27, 0xae, 0x8a, 0xc4, 0xb7, 0x38, 0xee, 0xa1, 0xaa, 0x58, 0xf2, 0x6e, 0xc8, 0x0d, 0xed, 0x1d, + 0xb9, 0x91, 0x14, 0xbc, 0x84, 0x25, 0x81, 0x9e, 0x41, 0x83, 0x12, 0x46, 0xe8, 0x29, 0x71, 0xcc, + 0x30, 0xa0, 0x51, 0x2c, 0xff, 0xd6, 0x62, 0xf2, 0x1f, 0x04, 0x34, 0xc2, 0x6b, 0x31, 0x12, 0xa7, + 0x18, 0x7a, 0x0a, 0x6b, 0xce, 0xb9, 0x6f, 0x79, 0xae, 0xad, 0x90, 0xcb, 0x4b, 0x23, 0xd7, 0x15, + 0x90, 0x00, 0xe6, 0x17, 0xa5, 0xd4, 0x20, 0xff, 0x63, 0x13, 0xeb, 0x88, 0x4c, 0x94, 0x4e, 0x24, + 0x31, 0xef, 0xd7, 0x25, 0xe5, 0xd7, 0xfa, 0x4f, 0xf2, 0xd0, 0x98, 0x77, 0x17, 0xf4, 0xbf, 0x00, + 0x76, 0x38, 0x35, 0x43, 0x42, 0xdd, 0xc0, 0x51, 0x4e, 0x51, 0xb5, 0xc3, 0xe9, 0x81, 0x60, 0xf0, + 0xc3, 0xc9, 0x87, 0xbf, 0x9c, 0x06, 0x91, 0xa5, 0xbc, 0xa3, 0x62, 0x87, 0xd3, 0x6f, 0x72, 0x3a, + 0x5e, 0x2b, 0x6e, 0xf2, 0x4c, 0x79, 0x09, 0x9f, 0x3e, 0x10, 0x0c, 0xf4, 0x0e, 0xdc, 0x96, 0x8e, + 0x64, 0x4e, 0x5c, 0xcf, 0x8d, 0x4c, 0xd7, 0x37, 0x8f, 0xce, 0xf9, 0xb5, 0x58, 0xfa, 0x0e, 0x92, + 0x83, 0xbb, 0x7c, 0x6c, 0xc7, 0xff, 0x98, 0x8f, 0x20, 0x1d, 0xd6, 0x82, 0xc0, 0x33, 0x99, 0x1d, + 0x50, 0x62, 0x5a, 0xce, 0x17, 0x22, 0x04, 0x16, 0x70, 0x2d, 0x08, 0xbc, 0x01, 0xe7, 0x75, 0x9c, + 0x2f, 0xf8, 0x89, 0xb7, 0xc3, 0x29, 0x23, 0x91, 0xc9, 0x7f, 0x9a, 0x65, 0x79, 0xe2, 0x25, 0xab, + 0x1b, 0x4e, 0x59, 0x6a, 0x82, 0x47, 0x3c, 0x1e, 0xb1, 0x52, 0x13, 0xf6, 0x88, 0xc7, 0xf4, 0xcf, + 0xa0, 0x24, 0xe2, 0x1b, 0xff, 0x77, 0xa2, 0x04, 0x10, 0xa1, 0x43, 0xea, 0xaf, 0xc2, 0x19, 0x22, + 0xb2, 0xbc, 0x02, 0xd5, 0x71, 0xc0, 0x54, 0xe0, 0x91, 0xae, 0x55, 0xe1, 0x0c, 0x31, 0xd8, 0x82, + 0x0a, 0x25, 0x96, 0x13, 0xf8, 0x93, 0x73, 0xf1, 0xc7, 0x2b, 0x38, 0xa1, 0xf5, 0x2f, 0xa1, 0x2c, + 0xe3, 0xde, 0x35, 0xf0, 0xef, 0x02, 0xb2, 0x47, 0x34, 0x98, 0x86, 0xdc, 0x32, 0x9e, 0xcb, 0x98, + 0x1b, 0xf8, 0x2c, 0xbe, 0x7b, 0xcb, 0x91, 0x83, 0xd9, 0x80, 0xfe, 0xd7, 0x9c, 0xcc, 0x75, 0xf2, + 0x22, 0xc5, 0xab, 0x05, 0x95, 0xb8, 0x96, 0xbe, 0x6d, 0x2a, 0x80, 0xb8, 0xe2, 0x23, 0xea, 0x8d, + 0x61, 0xd1, 0x8a, 0x8f, 0xc8, 0x8a, 0x8f, 0xf0, 0xf2, 0x42, 0xa5, 0x54, 0x09, 0x27, 0x33, 0x6a, + 0xcd, 0x49, 0x4a, 0x61, 0xa2, 0xff, 0x2b, 0x97, 0x1c, 0xf9, 0xb8, 0x64, 0x45, 0x9f, 0x43, 0x85, + 0x9f, 0x1e, 0xd3, 0xb3, 0x42, 0xf5, 0x9a, 0xd2, 0x5d, 0xae, 0x1a, 0x6e, 0xf3, 0xc3, 0xb2, 0x67, + 0x85, 0x32, 0xd7, 0xae, 0x86, 0x92, 0xe2, 0xa1, 0xc3, 0x72, 0x66, 0xa1, 0x83, 0x7f, 0xa3, 0xd7, + 0xa1, 0x61, 0x4d, 0xa3, 0xc0, 0xb4, 0x9c, 0x53, 0x42, 0x23, 0x97, 0x11, 0x65, 0xe1, 0x35, 0xce, + 0xed, 0xc4, 0xcc, 0xd6, 0x43, 0xa8, 0xa7, 0x31, 0xbf, 0x2e, 0xb9, 0x94, 0xd2, 0xc9, 0xe5, 0xbb, + 0x00, 0xb3, 0xca, 0x8c, 0x7b, 0x02, 0x39, 0x73, 0x23, 0xd3, 0x0e, 0x1c, 0x19, 0xda, 0x4a, 0xb8, + 0xc2, 0x19, 0xdd, 0xc0, 0x21, 0x17, 0xea, 0xdc, 0x52, 0x5c, 0xe7, 0xf2, 0xc3, 0xc7, 0x8f, 0xca, + 0x89, 0x3b, 0x99, 0x10, 0x47, 0x49, 0x58, 0x0d, 0x02, 0xef, 0x89, 0x60, 0xe8, 0x7f, 0xc9, 0x4b, + 0x8f, 0x90, 0xb7, 0x8c, 0x4c, 0xd5, 0x4f, 0x62, 0xea, 0xc2, 0xf5, 0x4c, 0xfd, 0x00, 0x80, 0x45, + 0x16, 0x8d, 0x88, 0x63, 0x5a, 0x91, 0xba, 0xb8, 0xb7, 0x2e, 0x15, 0xca, 0xc3, 0xf8, 0xc1, 0x12, + 0x57, 0xd5, 0xec, 0x4e, 0x84, 0xde, 0x87, 0xba, 0x1d, 0x78, 0xe1, 0x84, 0xa8, 0xc5, 0xa5, 0xaf, + 0x5d, 0x5c, 0x4b, 0xe6, 0x77, 0xa2, 0x54, 0x95, 0x5c, 0xbe, 0x6e, 0x95, 0xfc, 0xc7, 0x9c, 0xbc, + 0x2c, 0xa5, 0xef, 0x6a, 0x68, 0x74, 0xc5, 0xeb, 0xde, 0xe3, 0x25, 0x2f, 0x7e, 0x5f, 0xf5, 0xb4, + 0x77, 0xdd, 0xb7, 0xb4, 0x3f, 0x15, 0xa0, 0x9a, 0xdc, 0xb9, 0x2e, 0xd9, 0xfe, 0x3e, 0x54, 0x93, + 0xd7, 0x62, 0x55, 0x6b, 0x7c, 0xa5, 0x79, 0x92, 0xc9, 0xe8, 0x18, 0x90, 0x35, 0x1a, 0x25, 0x95, + 0x8a, 0x39, 0x65, 0xd6, 0x28, 0xbe, 0xa5, 0xde, 0x5f, 0x40, 0x0f, 0x71, 0xfa, 0x39, 0xe4, 0xeb, + 0xb1, 0x66, 0x8d, 0x46, 0x73, 0x1c, 0xf4, 0x3d, 0xb8, 0x3d, 0xbf, 0x87, 0x79, 0x74, 0x6e, 0x86, + 0xae, 0xa3, 0xaa, 0xec, 0xed, 0x45, 0xaf, 0x9d, 0xed, 0x39, 0xf8, 0x8f, 0xcf, 0x0f, 0x5c, 0x47, + 0xea, 0x1c, 0xd1, 0x4b, 0x03, 0xad, 0x1f, 0xc0, 0x4b, 0x2f, 0x98, 0x7e, 0x85, 0x0d, 0xfa, 0x69, + 0x1b, 0x5c, 0x47, 0x09, 0x29, 0xeb, 0xfd, 0x36, 0x27, 0x6f, 0xc7, 0xf3, 0x3a, 0xe9, 0xcc, 0xca, + 0xb6, 0xda, 0xd6, 0x66, 0xc6, 0x7d, 0xba, 0x07, 0x87, 0x12, 0x5e, 0xd4, 0x79, 0x9f, 0xcc, 0xd5, + 0x79, 0xd9, 0x6b, 0x91, 0x3d, 0xb1, 0x48, 0x02, 0x29, 0x04, 0xfd, 0xf7, 0x05, 0xa8, 0xc4, 0xe8, + 0xa2, 0x08, 0x3f, 0x67, 0x11, 0xf1, 0x4c, 0x2f, 0x0e, 0x61, 0x39, 0x0c, 0x92, 0xb5, 0xc7, 0x83, + 0xd8, 0x2b, 0x50, 0xe5, 0xb5, 0xbe, 0x1c, 0xce, 0x8b, 0xe1, 0x0a, 0x67, 0x88, 0xc1, 0x57, 0xa1, + 0x16, 0x05, 0x91, 0x35, 0x31, 0x23, 0xd7, 0x3e, 0x91, 0x49, 0x2e, 0x87, 0x41, 0xb0, 0x86, 0x9c, + 0x83, 0xde, 0x84, 0x1b, 0xd1, 0x98, 0x06, 0x51, 0x34, 0xe1, 0x65, 0x9a, 0x28, 0x4c, 0x64, 0x11, + 0x51, 0xc4, 0x5a, 0x32, 0x20, 0x0b, 0x16, 0xc6, 0xa3, 0xf7, 0x6c, 0x32, 0x77, 0x5d, 0x11, 0x44, + 0x8a, 0x78, 0x2d, 0xe1, 0x72, 0xd7, 0xe6, 0x97, 0xfc, 0x90, 0x50, 0x9b, 0xf8, 0x32, 0x56, 0xe4, + 0x70, 0x4c, 0x22, 0x13, 0xd6, 0x3d, 0x62, 0xb1, 0x29, 0x25, 0x8e, 0x79, 0xec, 0x92, 0x89, 0x23, + 0x2f, 0x3d, 0x8d, 0xcc, 0x45, 0x6d, 0xac, 0x96, 0xf6, 0x23, 0xb1, 0x1a, 0x37, 0x62, 0x38, 0x49, + 0xf3, 0xfa, 0x40, 0x7e, 0xa1, 0x75, 0xa8, 0x0d, 0x9e, 0x0d, 0x86, 0xc6, 0x9e, 0xb9, 0xb7, 0xdf, + 0x33, 0xd4, 0x2b, 0xf7, 0xc0, 0xc0, 0x92, 0xcc, 0xf1, 0xf1, 0xe1, 0xfe, 0xb0, 0xb3, 0x6b, 0x0e, + 0x77, 0xba, 0x4f, 0x06, 0x5a, 0x1e, 0xdd, 0x86, 0x1b, 0xc3, 0x6d, 0xbc, 0x3f, 0x1c, 0xee, 0x1a, + 0x3d, 0xf3, 0xc0, 0xc0, 0x3b, 0xfb, 0xbd, 0x81, 0x56, 0x40, 0x08, 0x1a, 0x33, 0xf6, 0x70, 0x67, + 0xcf, 0xd0, 0x8a, 0xa8, 0x06, 0xab, 0x07, 0x06, 0xee, 0x1a, 0xfd, 0xa1, 0x56, 0xd2, 0xff, 0x9e, + 0x87, 0x5a, 0xca, 0x8a, 0xdc, 0x91, 0x29, 0x93, 0x97, 0x98, 0x22, 0xe6, 0x9f, 0x3c, 0x98, 0xd8, + 0x96, 0x3d, 0x96, 0xd6, 0x29, 0x62, 0x49, 0x70, 0xbb, 0x79, 0xd6, 0x59, 0xea, 0x9c, 0x17, 0x71, + 0xc5, 0xb3, 0xce, 0x24, 0xc8, 0x6b, 0x50, 0x3f, 0x21, 0xd4, 0x27, 0x13, 0x35, 0x2e, 0x2d, 0x52, + 0x93, 0x3c, 0x39, 0x65, 0x03, 0x34, 0x35, 0x65, 0x06, 0x23, 0xcd, 0xd1, 0x90, 0xfc, 0xbd, 0x18, + 0xec, 0xe8, 0xb2, 0xd6, 0xcb, 0x42, 0xeb, 0x0f, 0x16, 0x77, 0xd2, 0x17, 0x29, 0x7e, 0x90, 0x28, + 0x7e, 0x15, 0x0a, 0x38, 0x7e, 0xf0, 0xed, 0x76, 0xba, 0xdb, 0x5c, 0xd9, 0x6b, 0x50, 0xdd, 0xeb, + 0x7c, 0x6a, 0x1e, 0x0e, 0xc4, 0xf3, 0x01, 0xd2, 0xa0, 0xfe, 0xc4, 0xc0, 0x7d, 0x63, 0x57, 0x71, + 0x0a, 0xe8, 0x16, 0x68, 0x8a, 0x33, 0x9b, 0x57, 0xd4, 0x7f, 0x97, 0x87, 0x75, 0x19, 0xd7, 0x93, + 0x17, 0xad, 0x17, 0x3f, 0x2d, 0x2d, 0x1f, 0x7a, 0x9b, 0xb0, 0xea, 0x11, 0x96, 0xd8, 0xa1, 0x8a, + 0x63, 0x12, 0xb9, 0x50, 0xb3, 0x7c, 0x3f, 0x88, 0xc4, 0xb3, 0x08, 0x53, 0x21, 0xf2, 0xf1, 0x42, + 0x0f, 0x30, 0x89, 0xe4, 0xed, 0xce, 0x0c, 0x49, 0x46, 0xc8, 0x34, 0x76, 0xeb, 0x03, 0xd0, 0x2e, + 0x4e, 0x58, 0x24, 0x2f, 0xbd, 0xf1, 0xce, 0x2c, 0x2d, 0x11, 0xee, 0xa0, 0x87, 0xfd, 0x27, 0xfd, + 0xfd, 0xa7, 0x7d, 0x6d, 0x85, 0x13, 0xf8, 0xb0, 0xdf, 0xdf, 0xe9, 0x3f, 0xd6, 0x72, 0x08, 0xa0, + 0x6c, 0x7c, 0xba, 0x33, 0x34, 0x7a, 0x5a, 0x7e, 0xeb, 0x1f, 0x6b, 0x50, 0x96, 0x42, 0xa2, 0x5f, + 0xab, 0x94, 0x9c, 0x6e, 0x3b, 0xa2, 0x0f, 0x16, 0x2e, 0x6d, 0xe7, 0x5a, 0x99, 0xad, 0x0f, 0x97, + 0x5e, 0xaf, 0x5e, 0x16, 0x57, 0xd0, 0xcf, 0x73, 0x50, 0x9f, 0x7b, 0x4a, 0xcb, 0xfa, 0x0a, 0x74, + 0x45, 0x97, 0xb3, 0xf5, 0x8d, 0xa5, 0xd6, 0x26, 0xb2, 0xfc, 0x2c, 0x07, 0xb5, 0x54, 0x7f, 0x0f, + 0x3d, 0x58, 0xa6, 0x27, 0x28, 0x25, 0x79, 0xb8, 0x7c, 0x3b, 0x51, 0x5f, 0x79, 0x3b, 0x87, 0x7e, + 0x9a, 0x83, 0x5a, 0xaa, 0x39, 0x96, 0x59, 0x94, 0xcb, 0xad, 0xbc, 0xcc, 0xa2, 0x5c, 0xd5, 0x8b, + 0x5b, 0x41, 0x3f, 0xcc, 0x41, 0x35, 0x69, 0x74, 0xa1, 0x7b, 0x8b, 0xb7, 0xc6, 0xa4, 0x10, 0xf7, + 0x97, 0xed, 0xa9, 0xe9, 0x2b, 0xe8, 0xfb, 0x50, 0x89, 0xbb, 0x42, 0x28, 0x6b, 0x1a, 0xb9, 0xd0, + 0x72, 0x6a, 0xdd, 0x5b, 0x78, 0x5d, 0x7a, 0xfb, 0xb8, 0x55, 0x93, 0x79, 0xfb, 0x0b, 0x4d, 0xa5, + 0xd6, 0xbd, 0x85, 0xd7, 0x25, 0xdb, 0x73, 0x4f, 0x48, 0x75, 0x74, 0x32, 0x7b, 0xc2, 0xe5, 0x56, + 0x52, 0x66, 0x4f, 0xb8, 0xaa, 0x81, 0x24, 0x05, 0x49, 0xf5, 0x84, 0x32, 0x0b, 0x72, 0xb9, 0xef, + 0x94, 0x59, 0x90, 0x2b, 0x5a, 0x50, 0xca, 0x25, 0x67, 0x05, 0xfa, 0xbd, 0x85, 0xdb, 0x28, 0x0b, + 0xba, 0xe4, 0xa5, 0x46, 0x8e, 0xbe, 0x82, 0x7e, 0xa4, 0x9e, 0x0c, 0x64, 0x0f, 0x06, 0x2d, 0x02, + 0x35, 0xd7, 0xb6, 0x69, 0xbd, 0xb7, 0x5c, 0xaa, 0x11, 0x31, 0xe2, 0xc7, 0x39, 0x80, 0x59, 0xb7, + 0x26, 0xb3, 0x10, 0x97, 0xda, 0x44, 0xad, 0x07, 0x4b, 0xac, 0x4c, 0x1f, 0x8f, 0xb8, 0x41, 0x93, + 0xf9, 0x78, 0x5c, 0xe8, 0x26, 0x65, 0x3e, 0x1e, 0x17, 0x3b, 0x41, 0xfa, 0xca, 0xc7, 0xab, 0xdf, + 0x2a, 0xc9, 0xdc, 0x5f, 0x16, 0x3f, 0xef, 0xfe, 0x27, 0x00, 0x00, 0xff, 0xff, 0x44, 0x47, 0xcf, + 0xbe, 0x32, 0x24, 0x00, 0x00, } diff --git a/plugins/drivers/proto/driver.proto b/plugins/drivers/proto/driver.proto index 9c870d27e..a5e44e36a 100644 --- a/plugins/drivers/proto/driver.proto +++ b/plugins/drivers/proto/driver.proto @@ -325,6 +325,12 @@ message TaskConfig { // AllocDir is the directory on the host where the allocation directory // exists. string alloc_dir = 9; + + // StdoutPath is the path to the file to open and write task stdout to + string stdout_path = 10; + + // StderrPath is the path to the file to open and write task stderr to + string stderr_path = 11; } message Resources { diff --git a/plugins/drivers/testing.go b/plugins/drivers/testing.go index 48dd00dae..a3ba7b3c0 100644 --- a/plugins/drivers/testing.go +++ b/plugins/drivers/testing.go @@ -87,21 +87,20 @@ func (h *DriverHarness) MkAllocDir(t *TaskConfig, enableLogs bool) func() { //logmon if enableLogs { - var stdoutFifo, stderrFifo string if runtime.GOOS == "windows" { id := uuid.Generate()[:8] - stdoutFifo = fmt.Sprintf("//./pipe/%s-%s.stdout", t.Name, id) - stderrFifo = fmt.Sprintf("//./pipe/%s-%s.stderr", t.Name, id) + t.StdoutPath = fmt.Sprintf("//./pipe/%s-%s.stdout", t.Name, id) + t.StderrPath = fmt.Sprintf("//./pipe/%s-%s.stderr", t.Name, id) } else { - stdoutFifo = filepath.Join(taskDir.LogDir, fmt.Sprintf(".%s.stdout.fifo", t.Name)) - stderrFifo = filepath.Join(taskDir.LogDir, fmt.Sprintf(".%s.stderr.fifo", t.Name)) + t.StdoutPath = filepath.Join(taskDir.LogDir, fmt.Sprintf(".%s.stdout.fifo", t.Name)) + t.StderrPath = filepath.Join(taskDir.LogDir, fmt.Sprintf(".%s.stderr.fifo", t.Name)) } err = h.lm.Start(&logmon.LogConfig{ LogDir: taskDir.LogDir, StdoutLogFile: fmt.Sprintf("%s.stdout", t.Name), StderrLogFile: fmt.Sprintf("%s.stderr", t.Name), - StdoutFifo: stdoutFifo, - StderrFifo: stderrFifo, + StdoutFifo: t.StdoutPath, + StderrFifo: t.StderrPath, MaxFiles: 10, MaxFileSizeMB: 10, }) @@ -114,6 +113,9 @@ func (h *DriverHarness) MkAllocDir(t *TaskConfig, enableLogs bool) func() { } return func() { + if h.lm != nil { + h.lm.Stop() + } allocDir.Destroy() } } diff --git a/plugins/drivers/utils.go b/plugins/drivers/utils.go index 41605aede..b74e13d4a 100644 --- a/plugins/drivers/utils.go +++ b/plugins/drivers/utils.go @@ -58,6 +58,8 @@ func taskConfigFromProto(pb *proto.TaskConfig) *TaskConfig { Mounts: []MountConfig{}, //TODO User: pb.User, AllocDir: pb.AllocDir, + StdoutPath: pb.StdoutPath, + StderrPath: pb.StderrPath, } } @@ -75,6 +77,8 @@ func taskConfigToProto(cfg *TaskConfig) *proto.TaskConfig { User: cfg.User, AllocDir: cfg.AllocDir, MsgpackDriverConfig: cfg.rawDriverConfig, + StdoutPath: cfg.StdoutPath, + StderrPath: cfg.StderrPath, } return pb } From 28e8e5852c5b5059e85233901336fdef4e3c37ee Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 26 Sep 2018 15:33:25 -0400 Subject: [PATCH 082/152] drivers/raw_exec: added unix specific tests --- drivers/raw_exec/driver.go | 23 ++--- drivers/raw_exec/driver_unix_test.go | 125 +++++++++++++++++++++++++++ 2 files changed, 137 insertions(+), 11 deletions(-) create mode 100644 drivers/raw_exec/driver_unix_test.go diff --git a/drivers/raw_exec/driver.go b/drivers/raw_exec/driver.go index e0bcc6e6d..a0f2b832b 100644 --- a/drivers/raw_exec/driver.go +++ b/drivers/raw_exec/driver.go @@ -246,13 +246,14 @@ func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, } execCmd := &executor.ExecCommand{ - Cmd: driverConfig.Command, - Args: driverConfig.Args, - Env: cfg.EnvList(), - User: cfg.User, - //TaskKillSignal: os.Interrupt, + Cmd: driverConfig.Command, + Args: driverConfig.Args, + Env: cfg.EnvList(), + User: cfg.User, BasicProcessCgroup: !r.config.NoCgroups, TaskDir: cfg.TaskDir().Dir, + StdoutPath: cfg.StdoutPath, + StderrPath: cfg.StderrPath, } ps, err := exec.Launch(execCmd) @@ -371,12 +372,12 @@ func (r *RawExecDriver) InspectTask(taskID string) (*drivers.TaskStatus, error) } status := &drivers.TaskStatus{ - ID: handle.task.ID, - Name: handle.task.Name, - State: handle.procState, - StartedAt: handle.startedAt, - CompletedAt: handle.completedAt, - ExitResult: handle.exitResult, + ID: handle.task.ID, + Name: handle.task.Name, + State: handle.procState, + StartedAt: handle.startedAt, + CompletedAt: handle.completedAt, + ExitResult: handle.exitResult, DriverAttributes: map[string]string{ "pid": strconv.Itoa(handle.pid), }, diff --git a/drivers/raw_exec/driver_unix_test.go b/drivers/raw_exec/driver_unix_test.go new file mode 100644 index 000000000..e20518782 --- /dev/null +++ b/drivers/raw_exec/driver_unix_test.go @@ -0,0 +1,125 @@ +// +build !windows + +package raw_exec + +import ( + "context" + "fmt" + "io/ioutil" + "path/filepath" + "runtime" + "strings" + "testing" + "time" + + "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/helper/testtask" + "github.com/hashicorp/nomad/helper/uuid" + "github.com/hashicorp/nomad/plugins/drivers" + "github.com/hashicorp/nomad/testutil" + "github.com/stretchr/testify/require" +) + +func TestRawExecDriver_User(t *testing.T) { + t.Parallel() + if runtime.GOOS != "linux" { + t.Skip("Linux only test") + } + require := require.New(t) + + d := NewRawExecDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + Name: "sleep", + User: "alice", + } + + cleanup := harness.MkAllocDir(task, false) + defer cleanup() + + task.EncodeDriverConfig(&TaskConfig{ + Command: testtask.Path(), + Args: []string{"sleep", "45s"}, + }) + + testtask.SetTaskConfigEnv(task) + _, err := harness.StartTask(task) + require.Error(err) + msg := "unknown user alice" + require.Contains(err.Error(), msg) +} + +func TestRawExecDriver_Signal(t *testing.T) { + t.Parallel() + if runtime.GOOS != "linux" { + t.Skip("Linux only test") + } + require := require.New(t) + + d := NewRawExecDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + Name: "signal", + } + + cleanup := harness.MkAllocDir(task, true) + defer cleanup() + + task.EncodeDriverConfig(&TaskConfig{ + Command: "/bin/bash", + Args: []string{"test.sh"}, + }) + + testFile := filepath.Join(task.TaskDir().Dir, "test.sh") + testData := []byte(` +at_term() { + echo 'Terminated.' + exit 3 +} +trap at_term USR1 +while true; do + sleep 1 +done + `) + require.NoError(ioutil.WriteFile(testFile, testData, 0777)) + + testtask.SetTaskConfigEnv(task) + _, err := harness.StartTask(task) + require.NoError(err) + + go func() { + time.Sleep(100 * time.Millisecond) + require.NoError(harness.SignalTask(task.ID, "SIGUSR1")) + }() + + // Task should terminate quickly + waitCh, err := harness.WaitTask(context.Background(), task.ID) + require.NoError(err) + select { + case res := <-waitCh: + require.False(res.Successful()) + require.Equal(3, res.ExitCode) + case <-time.After(time.Duration(testutil.TestMultiplier()*6) * time.Second): + require.Fail("WaitTask timeout") + } + + // Check the log file to see it exited because of the signal + outputFile := filepath.Join(task.TaskDir().LogDir, "signal.stdout.0") + exp := "Terminated." + testutil.WaitForResult(func() (bool, error) { + act, err := ioutil.ReadFile(outputFile) + if err != nil { + return false, fmt.Errorf("Couldn't read expected output: %v", err) + } + + if strings.TrimSpace(string(act)) != exp { + t.Logf("Read from %v", outputFile) + return false, fmt.Errorf("Command outputted %v; want %v", act, exp) + } + return true, nil + }, func(err error) { require.NoError(err) }) +} From a95cbc38baf732b56e8f9ada8109291a495ef479 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 26 Sep 2018 15:44:54 -0400 Subject: [PATCH 083/152] drivers/raw_exec: sync access to task state --- drivers/raw_exec/driver.go | 36 +++++++++++++++++++----------------- 1 file changed, 19 insertions(+), 17 deletions(-) diff --git a/drivers/raw_exec/driver.go b/drivers/raw_exec/driver.go index a0f2b832b..a41739865 100644 --- a/drivers/raw_exec/driver.go +++ b/drivers/raw_exec/driver.go @@ -371,6 +371,9 @@ func (r *RawExecDriver) InspectTask(taskID string) (*drivers.TaskStatus, error) return nil, drivers.ErrTaskNotFound } + handle.stateLock.RLock() + defer handle.stateLock.RUnlock() + status := &drivers.TaskStatus{ ID: handle.task.ID, Name: handle.task.Name, @@ -474,16 +477,6 @@ func (ts *taskStore) Delete(id string) { delete(ts.store, id) } -func (ts *taskStore) Range(f func(id string, handle *rawExecTaskHandle) bool) { - ts.lock.RLock() - defer ts.lock.RUnlock() - for k, v := range ts.store { - if f(k, v) { - break - } - } -} - type RawExecTaskState struct { ReattachConfig *utils.ReattachConfig TaskConfig *drivers.TaskConfig @@ -495,12 +488,16 @@ type rawExecTaskHandle struct { exec executor.Executor pid int pluginClient *plugin.Client - task *drivers.TaskConfig - procState drivers.TaskState - startedAt time.Time - completedAt time.Time - exitResult *drivers.ExitResult logger hclog.Logger + + // stateLock syncs access to all fields below + stateLock sync.RWMutex + + task *drivers.TaskConfig + procState drivers.TaskState + startedAt time.Time + completedAt time.Time + exitResult *drivers.ExitResult } func (h *rawExecTaskHandle) IsRunning() bool { @@ -508,11 +505,18 @@ func (h *rawExecTaskHandle) IsRunning() bool { } func (h *rawExecTaskHandle) run() { + + // since run is called immediatly after the handle is created this + // ensures the exitResult is initialized so we avoid a nil pointer + // thus it does not need to be included in the lock if h.exitResult == nil { h.exitResult = &drivers.ExitResult{} } ps, err := h.exec.Wait() + h.stateLock.Lock() + defer h.stateLock.Unlock() + if err != nil { h.exitResult.Err = err h.procState = drivers.TaskStateUnknown @@ -523,6 +527,4 @@ func (h *rawExecTaskHandle) run() { h.exitResult.ExitCode = ps.ExitCode h.exitResult.Signal = ps.Signal h.completedAt = ps.Time - - //h.pluginClient.Kill() } From cc76555814c9024fe09f2b30bc85e908b0e10eed Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Wed, 26 Sep 2018 13:38:46 -0700 Subject: [PATCH 084/152] Internal plugin catalog --- plugins/shared/catalog/catalog.go | 53 +++++++++++++++++++++++++++++++ 1 file changed, 53 insertions(+) create mode 100644 plugins/shared/catalog/catalog.go diff --git a/plugins/shared/catalog/catalog.go b/plugins/shared/catalog/catalog.go new file mode 100644 index 000000000..36bb7c6c0 --- /dev/null +++ b/plugins/shared/catalog/catalog.go @@ -0,0 +1,53 @@ +// Package catalog is used to register internal plugins such that they can be +// loaded. +package catalog + +import ( + "sync" + + "github.com/hashicorp/nomad/plugins/shared/loader" +) + +var ( + // catalog is the set of registered internal plugins + catalog = map[loader.PluginID]*Registration{} + mu sync.Mutex +) + +// Registration is the registration of an internal plugin +type Registration struct { + Config *loader.InternalPluginConfig + ConfigLoader ConfigFromOptions +} + +// ConfigFromOptions is used to retrieve a plugin config when passed a node's +// option map. This allows upgrade pathing from the old configuration format to +// the new config format. +type ConfigFromOptions func(options map[string]string) (config map[string]interface{}, err error) + +// Register is used to register an internal plugin. +func Register(id loader.PluginID, config *loader.InternalPluginConfig) { + mu.Lock() + defer mu.Unlock() + catalog[id] = &Registration{ + Config: config, + } +} + +// RegisterDeferredConfig is used to register an internal plugin that sets its +// config using the client's option map. +func RegisterDeferredConfig(id loader.PluginID, config *loader.InternalPluginConfig, configLoader ConfigFromOptions) { + mu.Lock() + defer mu.Unlock() + catalog[id] = &Registration{ + Config: config, + ConfigLoader: configLoader, + } +} + +// Catalog returns the catalog of internal plugins +func Catalog() map[loader.PluginID]*Registration { + mu.Lock() + defer mu.Unlock() + return catalog +} From 183561cf8240cda70df9d123524d9379b22bc1da Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Wed, 26 Sep 2018 13:39:09 -0700 Subject: [PATCH 085/152] Plugin loader initialization --- command/agent/agent.go | 15 +++++ command/agent/plugins.go | 82 ++++++++++++++++++++++++++++ plugins/shared/loader/init.go | 13 ++++- plugins/shared/loader/loader_test.go | 69 +++++++++++++++++++++++ 4 files changed, 176 insertions(+), 3 deletions(-) create mode 100644 command/agent/plugins.go diff --git a/command/agent/agent.go b/command/agent/agent.go index 9c91f6e40..dc7e21bcc 100644 --- a/command/agent/agent.go +++ b/command/agent/agent.go @@ -26,6 +26,7 @@ import ( "github.com/hashicorp/nomad/nomad" "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/nomad/structs/config" + "github.com/hashicorp/nomad/plugins/shared/loader" "github.com/hashicorp/raft" ) @@ -62,10 +63,21 @@ type Agent struct { // consulCatalog is the subset of Consul's Catalog API Nomad uses. consulCatalog consul.CatalogAPI + // client is the launched Nomad Client. Can be nil if the agent isn't + // configured to run a client. client *client.Client + // server is the launched Nomad Server. Can be nil if the agent isn't + // configured to run a server. server *nomad.Server + // pluginLoader is used to load plugins + pluginLoader loader.PluginCatalog + + // pluginSingletonLoader is a plugin loader that will returns singleton + // instances of the plugins. + pluginSingletonLoader loader.PluginCatalog + shutdown bool shutdownCh chan struct{} shutdownLock sync.Mutex @@ -99,6 +111,9 @@ func NewAgent(config *Config, logOutput io.Writer, inmem *metrics.InmemSink) (*A } // TODO setup plugin loader + if err := a.setupPlugins(); err != nil { + return nil, err + } if err := a.setupServer(); err != nil { return nil, err diff --git a/command/agent/plugins.go b/command/agent/plugins.go new file mode 100644 index 000000000..a0226246a --- /dev/null +++ b/command/agent/plugins.go @@ -0,0 +1,82 @@ +package agent + +import ( + "fmt" + + "github.com/hashicorp/nomad/plugins/shared/catalog" + "github.com/hashicorp/nomad/plugins/shared/loader" + "github.com/hashicorp/nomad/plugins/shared/singleton" +) + +// setupPlugins is used to setup the plugin loaders. +func (a *Agent) setupPlugins() error { + // Get our internal plugins + internal, err := a.internalPluginConfigs() + if err != nil { + return err + } + + // Build the plugin loader + config := &loader.PluginLoaderConfig{ + Logger: a.logger, + PluginDir: a.config.PluginDir, + Configs: a.config.Plugins, + InternalPlugins: internal, + } + l, err := loader.NewPluginLoader(config) + if err != nil { + return fmt.Errorf("failed to create plugin loader: %v", err) + } + a.pluginLoader = l + + // Wrap the loader to get our singleton loader + a.pluginSingletonLoader = singleton.NewSingletonLoader(a.logger, l) + + for k, plugins := range a.pluginLoader.Catalog() { + for _, p := range plugins { + a.logger.Info("detected plugin", "name", p.Name, "type", k, "plugin_version", p.PluginVersion) + } + } + + return nil +} + +func (a *Agent) internalPluginConfigs() (map[loader.PluginID]*loader.InternalPluginConfig, error) { + // Get the registered plugins + catalog := catalog.Catalog() + + // Create our map of plugins + internal := make(map[loader.PluginID]*loader.InternalPluginConfig, len(catalog)) + + // Grab the client options map if we can + var options map[string]string + if a.config != nil && a.config.Client != nil { + options = a.config.Client.Options + } + + for id, reg := range catalog { + if reg.Config == nil { + a.logger.Warn("skipping loading internal plugin because it is missing its configuration", "plugin", id) + continue + } + + pluginConfig := reg.Config.Config + if reg.ConfigLoader != nil { + pc, err := reg.ConfigLoader(options) + if err != nil { + return nil, fmt.Errorf("failed to retrieve config for internal plugin %v: %v", id, err) + } + + pluginConfig = pc + + // TODO We should log the config to warn users about upgrade pathing + } + + internal[id] = &loader.InternalPluginConfig{ + Factory: reg.Config.Factory, + Config: pluginConfig, + } + } + + return internal, nil +} diff --git a/plugins/shared/loader/init.go b/plugins/shared/loader/init.go index 3972b72e8..5ca4dca48 100644 --- a/plugins/shared/loader/init.go +++ b/plugins/shared/loader/init.go @@ -61,8 +61,11 @@ func validateConfig(config *PluginLoaderConfig) error { // init initializes the plugin loader by compiling both internal and external // plugins and selecting the highest versioned version of any given plugin. func (l *PluginLoader) init(config *PluginLoaderConfig) error { + // Create a mapping of name to config + configMap := configMap(config.Configs) + // Initialize the internal plugins - internal, err := l.initInternal(config.InternalPlugins) + internal, err := l.initInternal(config.InternalPlugins, configMap) if err != nil { return fmt.Errorf("failed to fingerprint internal plugins: %v", err) } @@ -74,7 +77,6 @@ func (l *PluginLoader) init(config *PluginLoaderConfig) error { } // Fingerprint the passed plugins - configMap := configMap(config.Configs) external, err := l.fingerprintPlugins(plugins, configMap) if err != nil { return fmt.Errorf("failed to fingerprint plugins: %v", err) @@ -92,7 +94,7 @@ func (l *PluginLoader) init(config *PluginLoaderConfig) error { } // initInternal initializes internal plugins. -func (l *PluginLoader) initInternal(plugins map[PluginID]*InternalPluginConfig) (map[PluginID]*pluginInfo, error) { +func (l *PluginLoader) initInternal(plugins map[PluginID]*InternalPluginConfig, configs map[string]*config.PluginConfig) (map[PluginID]*pluginInfo, error) { var mErr multierror.Error fingerprinted := make(map[PluginID]*pluginInfo, len(plugins)) for k, config := range plugins { @@ -109,6 +111,11 @@ func (l *PluginLoader) initInternal(plugins map[PluginID]*InternalPluginConfig) config: config.Config, } + // Try to retrieve a user specified config + if userConfig, ok := configs[k.Name]; ok && userConfig.Config != nil { + info.config = userConfig.Config + } + // Fingerprint base info i, err := base.PluginInfo() if err != nil { diff --git a/plugins/shared/loader/loader_test.go b/plugins/shared/loader/loader_test.go index 982be7968..3bc4d53a0 100644 --- a/plugins/shared/loader/loader_test.go +++ b/plugins/shared/loader/loader_test.go @@ -427,6 +427,75 @@ func TestPluginLoader_Internal_Config(t *testing.T) { require.EqualValues(expected, detected) } +// Tests that an external config can override the config of an internal plugin +func TestPluginLoader_Internal_ExternalConfig(t *testing.T) { + t.Parallel() + require := require.New(t) + + // Create the harness + h := newHarness(t, nil) + defer h.cleanup() + + plugin := "mock-device" + pluginVersion := "v0.0.1" + + id := PluginID{ + Name: plugin, + PluginType: base.PluginTypeDevice, + } + expectedConfig := map[string]interface{}{ + "foo": "2", + "bar": "3", + } + + logger := testlog.HCLogger(t) + logger.SetLevel(log.Trace) + lconfig := &PluginLoaderConfig{ + Logger: logger, + PluginDir: h.pluginDir(), + InternalPlugins: map[PluginID]*InternalPluginConfig{ + id: { + Factory: mockFactory(plugin, base.PluginTypeDevice, pluginVersion, true), + Config: map[string]interface{}{ + "foo": "1", + "bar": "2", + }, + }, + }, + Configs: []*config.PluginConfig{ + { + Name: plugin, + Config: expectedConfig, + }, + }, + } + + l, err := NewPluginLoader(lconfig) + require.NoError(err) + + // Get the catalog and assert we have the two plugins + c := l.Catalog() + require.Len(c, 1) + require.Contains(c, base.PluginTypeDevice) + detected := c[base.PluginTypeDevice] + require.Len(detected, 1) + + expected := []*base.PluginInfoResponse{ + { + Name: plugin, + Type: base.PluginTypeDevice, + PluginVersion: pluginVersion, + PluginApiVersion: "v0.1.0", + }, + } + require.EqualValues(expected, detected) + + // Check the config + loaded, ok := l.plugins[id] + require.True(ok) + require.EqualValues(expectedConfig, loaded.config) +} + // Pass a config but make sure it is fatal func TestPluginLoader_Internal_Config_Bad(t *testing.T) { t.Parallel() From 6f0ed6184b9152bef8ba0ad6fd3b8a5c850b6ac8 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Wed, 26 Sep 2018 18:14:36 -0700 Subject: [PATCH 086/152] Fix client reloading and pass the plugin loaders to server and client --- CHANGELOG.md | 1 + client/config/config.go | 8 ++ client/testing.go | 10 +- command/agent/agent.go | 210 ++++++++++++++++++------------ command/agent/command.go | 27 ++-- nomad/config.go | 8 ++ nomad/testing.go | 6 + plugins/shared/catalog/testing.go | 65 +++++++++ plugins/shared/loader/init.go | 11 +- 9 files changed, 251 insertions(+), 95 deletions(-) create mode 100644 plugins/shared/catalog/testing.go diff --git a/CHANGELOG.md b/CHANGELOG.md index 2c43b556c..60f849bff 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -15,6 +15,7 @@ IMPROVEMENTS: BUG FIXES: * core: Fixed bug in reconciler where allocs already stopped were being unnecessarily updated [[GH-4764](https://github.com/hashicorp/nomad/issues/4764)] + * client: Fix an issue reloading the client config [[GH-4730](https://github.com/hashicorp/nomad/issues/4730)] ## 0.8.6 (September 26, 2018) diff --git a/client/config/config.go b/client/config/config.go index 07ab2bf34..dfdd8cbee 100644 --- a/client/config/config.go +++ b/client/config/config.go @@ -13,6 +13,7 @@ import ( "github.com/hashicorp/nomad/helper" "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/nomad/structs/config" + "github.com/hashicorp/nomad/plugins/shared/loader" "github.com/hashicorp/nomad/version" ) @@ -206,6 +207,13 @@ type Config struct { // This period is meant to be long enough for a leader election to take // place, and a small jitter is applied to avoid a thundering herd. RPCHoldTimeout time.Duration + + // PluginLoader is used to load plugins. + PluginLoader loader.PluginCatalog + + // PluginSingletonLoader is a plugin loader that will returns singleton + // instances of the plugins. + PluginSingletonLoader loader.PluginCatalog } func (c *Config) Copy() *Config { diff --git a/client/testing.go b/client/testing.go index cde5e8b80..a67caa0dd 100644 --- a/client/testing.go +++ b/client/testing.go @@ -8,12 +8,16 @@ import ( "github.com/hashicorp/nomad/helper" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/shared/catalog" + "github.com/hashicorp/nomad/plugins/shared/singleton" "github.com/mitchellh/go-testing-interface" ) // TestClient creates an in-memory client for testing purposes. func TestClient(t testing.T, cb func(c *config.Config)) *Client { conf := config.DefaultConfig() + logger := testlog.HCLogger(t) + conf.Logger = logger conf.VaultConfig.Enabled = helper.BoolToPtr(false) conf.DevMode = true conf.Node = &structs.Node{ @@ -32,12 +36,14 @@ func TestClient(t testing.T, cb func(c *config.Config)) *Client { } conf.Options[fingerprint.TightenNetworkTimeoutsConfig] = "true" + // Set the plugin loaders + conf.PluginLoader = catalog.TestPluginLoader(t) + conf.PluginSingletonLoader = singleton.NewSingletonLoader(logger, conf.PluginLoader) + if cb != nil { cb(conf) } - logger := testlog.HCLogger(t) - conf.Logger = logger catalog := consul.NewMockCatalog(logger) mockService := consulApi.NewMockConsulServiceClient(t, logger) client, err := NewClient(conf, catalog, mockService) diff --git a/command/agent/agent.go b/command/agent/agent.go index dc7e21bcc..0dc02044c 100644 --- a/command/agent/agent.go +++ b/command/agent/agent.go @@ -110,7 +110,6 @@ func NewAgent(config *Config, logOutput io.Writer, inmem *metrics.InmemSink) (*A return nil, fmt.Errorf("Failed to initialize Consul client: %v", err) } - // TODO setup plugin loader if err := a.setupPlugins(); err != nil { return nil, err } @@ -129,14 +128,13 @@ func NewAgent(config *Config, logOutput io.Writer, inmem *metrics.InmemSink) (*A } // convertServerConfig takes an agent config and log output and returns a Nomad -// Config. -func convertServerConfig(agentConfig *Config, logger log.Logger, logOutput io.Writer) (*nomad.Config, error) { +// Config. There may be missing fields that must be set by the agent. To do this +// call finalizeServerConfig +func convertServerConfig(agentConfig *Config) (*nomad.Config, error) { conf := agentConfig.NomadConfig if conf == nil { conf = nomad.DefaultConfig() } - conf.Logger = logger - conf.LogOutput = logOutput conf.DevMode = agentConfig.DevMode conf.Build = agentConfig.Version.VersionNumber() if agentConfig.Region != "" { @@ -335,61 +333,73 @@ func convertServerConfig(agentConfig *Config, logger log.Logger, logOutput io.Wr // serverConfig is used to generate a new server configuration struct // for initializing a nomad server. func (a *Agent) serverConfig() (*nomad.Config, error) { - return convertServerConfig(a.config, a.logger, a.logOutput) + c, err := convertServerConfig(a.config) + if err != nil { + return nil, err + } + + a.finalizeServerConfig(c) + return c, nil } -// clientConfig is used to generate a new client configuration struct -// for initializing a Nomad client. +// finalizeServerConfig sets configuration fields on the server config that are +// not staticly convertable and are from the agent. +func (a *Agent) finalizeServerConfig(c *nomad.Config) { + // Setup the logging + c.Logger = a.logger + c.LogOutput = a.logOutput + + // Setup the plugin loaders + c.PluginLoader = a.pluginLoader + c.PluginSingletonLoader = a.pluginSingletonLoader +} + +// clientConfig is used to generate a new client configuration struct for +// initializing a Nomad client. func (a *Agent) clientConfig() (*clientconfig.Config, error) { - // Setup the configuration - conf := a.config.ClientConfig - if conf == nil { - conf = clientconfig.DefaultConfig() + c, err := convertClientConfig(a.config) + if err != nil { + return nil, err } + if err := a.finalizeClientConfig(c); err != nil { + return nil, err + } + + return c, nil +} + +// finalizeClientConfig sets configuration fields on the client config that are +// not staticly convertable and are from the agent. +func (a *Agent) finalizeClientConfig(c *clientconfig.Config) error { + // Setup the logging + c.Logger = a.logger + c.LogOutput = a.logOutput + // If we are running a server, append both its bind and advertise address so // we are able to at least talk to the local server even if that isn't // configured explicitly. This handles both running server and client on one // host and -dev mode. - conf.Servers = a.config.Client.Servers if a.server != nil { if a.config.AdvertiseAddrs == nil || a.config.AdvertiseAddrs.RPC == "" { - return nil, fmt.Errorf("AdvertiseAddrs is nil or empty") + return fmt.Errorf("AdvertiseAddrs is nil or empty") } else if a.config.normalizedAddrs == nil || a.config.normalizedAddrs.RPC == "" { - return nil, fmt.Errorf("normalizedAddrs is nil or empty") + return fmt.Errorf("normalizedAddrs is nil or empty") } - conf.Servers = append(conf.Servers, + c.Servers = append(c.Servers, a.config.normalizedAddrs.RPC, a.config.AdvertiseAddrs.RPC) } - conf.Logger = a.logger - conf.LogOutput = a.logOutput - conf.LogLevel = a.config.LogLevel - conf.DevMode = a.config.DevMode - if a.config.Region != "" { - conf.Region = a.config.Region - } - if a.config.DataDir != "" { - conf.StateDir = filepath.Join(a.config.DataDir, "client") - conf.AllocDir = filepath.Join(a.config.DataDir, "alloc") - } - if a.config.Client.StateDir != "" { - conf.StateDir = a.config.Client.StateDir - } - if a.config.Client.AllocDir != "" { - conf.AllocDir = a.config.Client.AllocDir - } - if a.config.Client.NetworkInterface != "" { - conf.NetworkInterface = a.config.Client.NetworkInterface - } - conf.ChrootEnv = a.config.Client.ChrootEnv - conf.Options = a.config.Client.Options - // Logging deprecation messages about consul related configuration in client + // Setup the plugin loaders + c.PluginLoader = a.pluginLoader + c.PluginSingletonLoader = a.pluginSingletonLoader + + // Log deprecation messages about Consul related configuration in client // options var invalidConsulKeys []string - for key := range conf.Options { + for key := range c.Options { if strings.HasPrefix(key, "consul") { invalidConsulKeys = append(invalidConsulKeys, fmt.Sprintf("options.%s", key)) } @@ -401,34 +411,68 @@ func (a *Agent) clientConfig() (*clientconfig.Config, error) { to configure Nomad to work with Consul.`) } - if a.config.Client.NetworkSpeed != 0 { - conf.NetworkSpeed = a.config.Client.NetworkSpeed + return nil +} + +// convertClientConfig takes an agent config and log output and returns a client +// Config. There may be missing fields that must be set by the agent. To do this +// call finalizeServerConfig +func convertClientConfig(agentConfig *Config) (*clientconfig.Config, error) { + // Setup the configuration + conf := agentConfig.ClientConfig + if conf == nil { + conf = clientconfig.DefaultConfig() } - if a.config.Client.CpuCompute != 0 { - conf.CpuCompute = a.config.Client.CpuCompute + + conf.Servers = agentConfig.Client.Servers + conf.LogLevel = agentConfig.LogLevel + conf.DevMode = agentConfig.DevMode + if agentConfig.Region != "" { + conf.Region = agentConfig.Region } - if a.config.Client.MemoryMB != 0 { - conf.MemoryMB = a.config.Client.MemoryMB + if agentConfig.DataDir != "" { + conf.StateDir = filepath.Join(agentConfig.DataDir, "client") + conf.AllocDir = filepath.Join(agentConfig.DataDir, "alloc") } - if a.config.Client.MaxKillTimeout != "" { - dur, err := time.ParseDuration(a.config.Client.MaxKillTimeout) + if agentConfig.Client.StateDir != "" { + conf.StateDir = agentConfig.Client.StateDir + } + if agentConfig.Client.AllocDir != "" { + conf.AllocDir = agentConfig.Client.AllocDir + } + if agentConfig.Client.NetworkInterface != "" { + conf.NetworkInterface = agentConfig.Client.NetworkInterface + } + conf.ChrootEnv = agentConfig.Client.ChrootEnv + conf.Options = agentConfig.Client.Options + if agentConfig.Client.NetworkSpeed != 0 { + conf.NetworkSpeed = agentConfig.Client.NetworkSpeed + } + if agentConfig.Client.CpuCompute != 0 { + conf.CpuCompute = agentConfig.Client.CpuCompute + } + if agentConfig.Client.MemoryMB != 0 { + conf.MemoryMB = agentConfig.Client.MemoryMB + } + if agentConfig.Client.MaxKillTimeout != "" { + dur, err := time.ParseDuration(agentConfig.Client.MaxKillTimeout) if err != nil { return nil, fmt.Errorf("Error parsing max kill timeout: %s", err) } conf.MaxKillTimeout = dur } - conf.ClientMaxPort = uint(a.config.Client.ClientMaxPort) - conf.ClientMinPort = uint(a.config.Client.ClientMinPort) + conf.ClientMaxPort = uint(agentConfig.Client.ClientMaxPort) + conf.ClientMinPort = uint(agentConfig.Client.ClientMinPort) // Setup the node conf.Node = new(structs.Node) - conf.Node.Datacenter = a.config.Datacenter - conf.Node.Name = a.config.NodeName - conf.Node.Meta = a.config.Client.Meta - conf.Node.NodeClass = a.config.Client.NodeClass + conf.Node.Datacenter = agentConfig.Datacenter + conf.Node.Name = agentConfig.NodeName + conf.Node.Meta = agentConfig.Client.Meta + conf.Node.NodeClass = agentConfig.Client.NodeClass // Set up the HTTP advertise address - conf.Node.HTTPAddr = a.config.AdvertiseAddrs.HTTP + conf.Node.HTTPAddr = agentConfig.AdvertiseAddrs.HTTP // Reserve resources on the node. // COMPAT(0.10): Remove in 0.10 @@ -437,58 +481,58 @@ func (a *Agent) clientConfig() (*clientconfig.Config, error) { r = new(structs.Resources) conf.Node.Reserved = r } - r.CPU = a.config.Client.Reserved.CPU - r.MemoryMB = a.config.Client.Reserved.MemoryMB - r.DiskMB = a.config.Client.Reserved.DiskMB - r.IOPS = a.config.Client.Reserved.IOPS + r.CPU = agentConfig.Client.Reserved.CPU + r.MemoryMB = agentConfig.Client.Reserved.MemoryMB + r.DiskMB = agentConfig.Client.Reserved.DiskMB + r.IOPS = agentConfig.Client.Reserved.IOPS res := conf.Node.ReservedResources if res == nil { res = new(structs.NodeReservedResources) conf.Node.ReservedResources = res } - res.Cpu.CpuShares = int64(a.config.Client.Reserved.CPU) - res.Memory.MemoryMB = int64(a.config.Client.Reserved.MemoryMB) - res.Disk.DiskMB = int64(a.config.Client.Reserved.DiskMB) - res.Networks.ReservedHostPorts = a.config.Client.Reserved.ReservedPorts + res.Cpu.CpuShares = int64(agentConfig.Client.Reserved.CPU) + res.Memory.MemoryMB = int64(agentConfig.Client.Reserved.MemoryMB) + res.Disk.DiskMB = int64(agentConfig.Client.Reserved.DiskMB) + res.Networks.ReservedHostPorts = agentConfig.Client.Reserved.ReservedPorts - conf.Version = a.config.Version + conf.Version = agentConfig.Version - if *a.config.Consul.AutoAdvertise && a.config.Consul.ClientServiceName == "" { + if *agentConfig.Consul.AutoAdvertise && agentConfig.Consul.ClientServiceName == "" { return nil, fmt.Errorf("client_service_name must be set when auto_advertise is enabled") } - conf.ConsulConfig = a.config.Consul - conf.VaultConfig = a.config.Vault + conf.ConsulConfig = agentConfig.Consul + conf.VaultConfig = agentConfig.Vault // Set up Telemetry configuration - conf.StatsCollectionInterval = a.config.Telemetry.collectionInterval - conf.PublishNodeMetrics = a.config.Telemetry.PublishNodeMetrics - conf.PublishAllocationMetrics = a.config.Telemetry.PublishAllocationMetrics - conf.DisableTaggedMetrics = a.config.Telemetry.DisableTaggedMetrics - conf.BackwardsCompatibleMetrics = a.config.Telemetry.BackwardsCompatibleMetrics + conf.StatsCollectionInterval = agentConfig.Telemetry.collectionInterval + conf.PublishNodeMetrics = agentConfig.Telemetry.PublishNodeMetrics + conf.PublishAllocationMetrics = agentConfig.Telemetry.PublishAllocationMetrics + conf.DisableTaggedMetrics = agentConfig.Telemetry.DisableTaggedMetrics + conf.BackwardsCompatibleMetrics = agentConfig.Telemetry.BackwardsCompatibleMetrics // Set the TLS related configs - conf.TLSConfig = a.config.TLSConfig + conf.TLSConfig = agentConfig.TLSConfig conf.Node.TLSEnabled = conf.TLSConfig.EnableHTTP // Set the GC related configs - conf.GCInterval = a.config.Client.GCInterval - conf.GCParallelDestroys = a.config.Client.GCParallelDestroys - conf.GCDiskUsageThreshold = a.config.Client.GCDiskUsageThreshold - conf.GCInodeUsageThreshold = a.config.Client.GCInodeUsageThreshold - conf.GCMaxAllocs = a.config.Client.GCMaxAllocs - if a.config.Client.NoHostUUID != nil { - conf.NoHostUUID = *a.config.Client.NoHostUUID + conf.GCInterval = agentConfig.Client.GCInterval + conf.GCParallelDestroys = agentConfig.Client.GCParallelDestroys + conf.GCDiskUsageThreshold = agentConfig.Client.GCDiskUsageThreshold + conf.GCInodeUsageThreshold = agentConfig.Client.GCInodeUsageThreshold + conf.GCMaxAllocs = agentConfig.Client.GCMaxAllocs + if agentConfig.Client.NoHostUUID != nil { + conf.NoHostUUID = *agentConfig.Client.NoHostUUID } else { // Default no_host_uuid to true conf.NoHostUUID = true } // Setup the ACLs - conf.ACLEnabled = a.config.ACL.Enabled - conf.ACLTokenTTL = a.config.ACL.TokenTTL - conf.ACLPolicyTTL = a.config.ACL.PolicyTTL + conf.ACLEnabled = agentConfig.ACL.Enabled + conf.ACLTokenTTL = agentConfig.ACL.TokenTTL + conf.ACLPolicyTTL = agentConfig.ACL.PolicyTTL return conf, nil } diff --git a/command/agent/command.go b/command/agent/command.go index e15555c33..5c049be50 100644 --- a/command/agent/command.go +++ b/command/agent/command.go @@ -751,25 +751,36 @@ func (c *Command) handleReload() { if s := c.agent.Server(); s != nil { c.agent.logger.Debug("starting reload of server config") - sconf, err := convertServerConfig(newConf, c.agent.logger, c.logOutput) + sconf, err := convertServerConfig(newConf) if err != nil { c.agent.logger.Error("failed to convert server config", "error", err) return - } else { - if err := s.Reload(sconf); err != nil { - c.agent.logger.Error("reloading server config failed", "error", err) - return - } + } + + // Finalize the config to get the agent objects injected in + c.agent.finalizeServerConfig(sconf) + + // Reload the config + if err := s.Reload(sconf); err != nil { + c.agent.logger.Error("reloading server config failed", "error", err) + return } } if s := c.agent.Client(); s != nil { - clientConfig, err := c.agent.clientConfig() c.agent.logger.Debug("starting reload of client config") + clientConfig, err := convertClientConfig(newConf) if err != nil { - c.agent.logger.Error("reloading client config failed", "error", err) + c.agent.logger.Error("failed to convert client config", "error", err) return } + + // Finalize the config to get the agent objects injected in + if err := c.agent.finalizeClientConfig(clientConfig); err != nil { + c.agent.logger.Error("failed to finalize client config", "error", err) + return + } + if err := c.agent.Client().Reload(clientConfig); err != nil { c.agent.logger.Error("reloading client config failed", "error", err) return diff --git a/nomad/config.go b/nomad/config.go index 2e0b49ed7..c712e5833 100644 --- a/nomad/config.go +++ b/nomad/config.go @@ -14,6 +14,7 @@ import ( "github.com/hashicorp/nomad/helper/uuid" "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/nomad/structs/config" + "github.com/hashicorp/nomad/plugins/shared/loader" "github.com/hashicorp/nomad/scheduler" "github.com/hashicorp/raft" "github.com/hashicorp/serf/serf" @@ -295,6 +296,13 @@ type Config struct { // autopilot tasks, such as promoting eligible non-voters and removing // dead servers. AutopilotInterval time.Duration + + // PluginLoader is used to load plugins. + PluginLoader loader.PluginCatalog + + // PluginSingletonLoader is a plugin loader that will returns singleton + // instances of the plugins. + PluginSingletonLoader loader.PluginCatalog } // CheckVersion is used to check if the ProtocolVersion is valid diff --git a/nomad/testing.go b/nomad/testing.go index 9696aa386..7d0f9fcd0 100644 --- a/nomad/testing.go +++ b/nomad/testing.go @@ -12,6 +12,8 @@ import ( "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/nomad/mock" "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/shared/catalog" + "github.com/hashicorp/nomad/plugins/shared/singleton" "github.com/mitchellh/go-testing-interface" ) @@ -69,6 +71,10 @@ func TestServer(t testing.T, cb func(*Config)) *Server { config.ServerHealthInterval = 50 * time.Millisecond config.AutopilotInterval = 100 * time.Millisecond + // Set the plugin loaders + config.PluginLoader = catalog.TestPluginLoader(t) + config.PluginSingletonLoader = singleton.NewSingletonLoader(config.Logger, config.PluginLoader) + // Invoke the callback if any if cb != nil { cb(config) diff --git a/plugins/shared/catalog/testing.go b/plugins/shared/catalog/testing.go new file mode 100644 index 000000000..59bcc6cf1 --- /dev/null +++ b/plugins/shared/catalog/testing.go @@ -0,0 +1,65 @@ +package catalog + +import ( + "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/nomad/structs/config" + "github.com/hashicorp/nomad/plugins/shared/loader" + "github.com/mitchellh/go-testing-interface" +) + +// TestPluginLoader returns a plugin loader populated only with internal plugins +func TestPluginLoader(t testing.T) loader.PluginCatalog { + return TestPluginLoaderWithOptions(t, "", nil, nil) +} + +// TestPluginLoaderWithOptions allows configuring the plugin loader fully. +func TestPluginLoaderWithOptions(t testing.T, + pluginDir string, + options map[string]string, + configs []*config.PluginConfig) loader.PluginCatalog { + + // Get a logger + logger := testlog.HCLogger(t) + + // Get the registered plugins + catalog := Catalog() + + // Create our map of plugins + internal := make(map[loader.PluginID]*loader.InternalPluginConfig, len(catalog)) + + for id, reg := range catalog { + if reg.Config == nil { + logger.Warn("skipping loading internal plugin because it is missing its configuration", "plugin", id) + continue + } + + pluginConfig := reg.Config.Config + if reg.ConfigLoader != nil { + pc, err := reg.ConfigLoader(options) + if err != nil { + t.Fatalf("failed to retrieve config for internal plugin %v: %v", id, err) + } + + pluginConfig = pc + } + + internal[id] = &loader.InternalPluginConfig{ + Factory: reg.Config.Factory, + Config: pluginConfig, + } + } + + // Build the plugin loader + config := &loader.PluginLoaderConfig{ + Logger: logger, + PluginDir: "", + Configs: configs, + InternalPlugins: internal, + } + l, err := loader.NewPluginLoader(config) + if err != nil { + t.Fatalf("failed to create plugin loader: %v", err) + } + + return l +} diff --git a/plugins/shared/loader/init.go b/plugins/shared/loader/init.go index 5ca4dca48..1f9a48f54 100644 --- a/plugins/shared/loader/init.go +++ b/plugins/shared/loader/init.go @@ -32,8 +32,6 @@ func validateConfig(config *PluginLoaderConfig) error { return fmt.Errorf("nil config passed") } else if config.Logger == nil { multierror.Append(&mErr, fmt.Errorf("nil logger passed")) - } else if config.PluginDir == "" { - multierror.Append(&mErr, fmt.Errorf("invalid plugin dir %q passed", config.PluginDir)) } // Validate that all plugins have a binary name @@ -149,9 +147,18 @@ func (l *PluginLoader) initInternal(plugins map[PluginID]*InternalPluginConfig, // scan scans the plugin directory and retrieves potentially eligible binaries func (l *PluginLoader) scan() ([]os.FileInfo, error) { + if l.pluginDir == "" { + return nil, nil + } + // Capture the list of binaries in the plugins folder f, err := os.Open(l.pluginDir) if err != nil { + // There are no plugins to scan + if os.IsNotExist(err) { + return nil, nil + } + return nil, fmt.Errorf("failed to open plugin directory %q: %v", l.pluginDir, err) } files, err := f.Readdirnames(-1) From 84ce8c348782783b66a8c7db97812f0108a5da1d Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Thu, 27 Sep 2018 09:34:58 -0700 Subject: [PATCH 087/152] extra logging --- command/agent/plugins.go | 2 +- plugins/shared/loader/init.go | 1 + 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/command/agent/plugins.go b/command/agent/plugins.go index a0226246a..6637d69f9 100644 --- a/command/agent/plugins.go +++ b/command/agent/plugins.go @@ -56,7 +56,7 @@ func (a *Agent) internalPluginConfigs() (map[loader.PluginID]*loader.InternalPlu for id, reg := range catalog { if reg.Config == nil { - a.logger.Warn("skipping loading internal plugin because it is missing its configuration", "plugin", id) + a.logger.Error("skipping loading internal plugin because it is missing its configuration", "plugin", id) continue } diff --git a/plugins/shared/loader/init.go b/plugins/shared/loader/init.go index 1f9a48f54..da15623e3 100644 --- a/plugins/shared/loader/init.go +++ b/plugins/shared/loader/init.go @@ -156,6 +156,7 @@ func (l *PluginLoader) scan() ([]os.FileInfo, error) { if err != nil { // There are no plugins to scan if os.IsNotExist(err) { + l.logger.Debug("skipping external plugins since plugin_dir doesn't exist") return nil, nil } From 3ccc091a72a46a08e65fd770d708a53f4c8d60cf Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 19 Sep 2018 17:32:50 -0700 Subject: [PATCH 088/152] ar: lock around accessing tasks Specify that Alloc() does not return updated task states. --- client/allocrunnerv2/alloc_runner.go | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 73e5773cd..b4b47326f 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -206,7 +206,8 @@ func (ar *allocRunner) runImpl() <-chan struct{} { return waitCh } -// Alloc returns the current allocation being run by this runner. +// Alloc returns the current allocation being run by this runner as sent by the +// server. This view of the allocation does not have updated task states. func (ar *allocRunner) Alloc() *structs.Allocation { ar.allocLock.RLock() defer ar.allocLock.RUnlock() @@ -277,6 +278,7 @@ func (ar *allocRunner) TaskStateUpdated(taskName string, state *structs.TaskStat } // Gather the state of the other tasks + ar.tasksLock.RLock() states := make(map[string]*structs.TaskState, len(ar.tasks)) for name, tr := range ar.tasks { if name == taskName { @@ -285,6 +287,7 @@ func (ar *allocRunner) TaskStateUpdated(taskName string, state *structs.TaskStat states[name] = tr.TaskState() } } + ar.tasksLock.RUnlock() // Get the client allocation calloc := ar.clientAlloc(states) From 55ab491801690e9824ee6085546e2f168ba31e03 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 19 Sep 2018 17:33:41 -0700 Subject: [PATCH 089/152] tr: remove wip comments --- .../taskrunner/task_runner_hooks.go | 47 ------------------- 1 file changed, 47 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 8baf4a6c5..4e3d550f7 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -69,8 +69,6 @@ func (tr *TaskRunner) initHooks() { // prestart is used to run the runners prestart hooks. func (tr *TaskRunner) prestart() error { - //XXX is this necessary? maybe we should have a generic cancelletion - // method instead of peeking into the alloc // Determine if the allocation is terminaland we should avoid running // prestart hooks. alloc := tr.Alloc() @@ -391,48 +389,3 @@ func (tr *TaskRunner) kill() { } } } - -/* -TR Hooks: - -> @schmichael -Task Validate: -Require: Client config, task definiton -Return: error -Implement: Prestart - -> DONE -Task Dir Build: -Requires: Folder structure, driver isolation, client config -Return env, error -Implement: Prestart - -> @alex -Vault: Task, RPC to talk to server to derive token, Node SecretID -Return vault token (Call a setter), error, env -Implement: Prestart - -> @alex -Consul Template: -Require: Task, alloc directory, way to signal/restart task, updates when vault token changes -Return env, error -Implement: Prestart and Update (for new Vault token) and Destroy - -> @schmichael -Consul Service Reg: -Require: Task, interpolation/ENV -Return: error -Implement: Poststart, Update, Kill, Exited - -> @alex -Dispatch Payload: -Require: Alloc -Return error -Implement: Prestart - -> @schmichael -Artifacts: -Require: Folder structure, task, interpolation/ENV -Return: error -Implement: Prestart -*/ From 8d1419c62bd6a9f1f4be469f42e2638f8f9cdaca Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 19 Sep 2018 17:34:18 -0700 Subject: [PATCH 090/152] client: fix accessing alloc runners * GetClientAlloc() gains nothing from using allAllocs() * getAllocatedResources was calling getAllocRunners() twice --- client/client.go | 10 ++++++---- 1 file changed, 6 insertions(+), 4 deletions(-) diff --git a/client/client.go b/client/client.go index 83508769f..1056520fd 100644 --- a/client/client.go +++ b/client/client.go @@ -639,12 +639,14 @@ func (c *Client) GetAllocFS(allocID string) (allocdir.AllocDirFS, error) { // GetClientAlloc returns the allocation from the client func (c *Client) GetClientAlloc(allocID string) (*structs.Allocation, error) { - all := c.allAllocs() - alloc, ok := all[allocID] + c.allocLock.RLock() + ar, ok := c.allocs[allocID] + c.allocLock.RUnlock() if !ok { return nil, structs.NewErrUnknownAllocation(allocID) } - return alloc, nil + + return ar.Alloc(), nil } // GetServers returns the list of nomad servers this client is aware of. @@ -2527,7 +2529,7 @@ func (c *Client) getAllocatedResources(selfNode *structs.Node) *structs.Comparab func (c *Client) allAllocs() map[string]*structs.Allocation { ars := c.getAllocRunners() allocs := make(map[string]*structs.Allocation, len(ars)) - for _, ar := range c.getAllocRunners() { + for _, ar := range ars { a := ar.Alloc() allocs[a.ID] = a } From 23253480533354a92750757bd2696f6017b0f9ab Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 19 Sep 2018 17:35:37 -0700 Subject: [PATCH 091/152] mock_driver: close waitCh after exiting mock_driver wasn't behaving like other driver handles. --- client/driver/mock_driver.go | 2 ++ 1 file changed, 2 insertions(+) diff --git a/client/driver/mock_driver.go b/client/driver/mock_driver.go index 5f2436ea5..f17c9d8ab 100644 --- a/client/driver/mock_driver.go +++ b/client/driver/mock_driver.go @@ -415,6 +415,8 @@ func (h *mockDriverHandle) Stats() (*cstructs.TaskResourceUsage, error) { // run waits for the configured amount of time and then indicates the task has // terminated func (h *mockDriverHandle) run() { + defer close(h.waitCh) + // Setup logging output if h.stdoutString != "" { go h.handleLogging() From 951ed174361a1a0341f005021a7c7f4dd1b024b5 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 19 Sep 2018 17:36:34 -0700 Subject: [PATCH 092/152] client: do not inspect task state to follow logs "Ask forgiveness, not permission." Instead of peaking at TaskStates (which are no longer updated on the AR.Alloc() view of the world) to only read logs for running tasks, just try to read the logs and improve the error handling if they don't exist. This should make log streaming less dependent on AR/TR behavior. Also fixed a race where the log streamer could exit before reading an error. This caused no logs or errors to be displayed sometimes when an error occurred. --- client/fs_endpoint.go | 40 ++++++++--- client/fs_endpoint_test.go | 135 ++++++++++--------------------------- 2 files changed, 68 insertions(+), 107 deletions(-) diff --git a/client/fs_endpoint.go b/client/fs_endpoint.go index 0578ab519..41123dbf4 100644 --- a/client/fs_endpoint.go +++ b/client/fs_endpoint.go @@ -6,6 +6,7 @@ import ( "fmt" "io" "math" + "net/http" "os" "path/filepath" "sort" @@ -395,13 +396,6 @@ func (f *FileSystem) logs(conn io.ReadWriteCloser) { return } - state, ok := alloc.TaskStates[req.Task] - if !ok || state.StartedAt.IsZero() { - f.handleStreamResultError(fmt.Errorf("task %q not started yet. No logs available", req.Task), - helper.Int64ToPtr(404), encoder) - return - } - ctx, cancel := context.WithCancel(context.Background()) defer cancel() @@ -447,6 +441,15 @@ OUTER: break OUTER case frame, ok := <-frames: if !ok { + // framer may have been closed when an error + // occurred. Check once more for an error. + select { + case streamErr = <-errCh: + // There was a pending error! + default: + // No error, continue on + } + break OUTER } @@ -473,7 +476,12 @@ OUTER: } if streamErr != nil { - f.handleStreamResultError(streamErr, helper.Int64ToPtr(500), encoder) + // If error has a Code, use it + var code int64 = 500 + if codedErr, ok := streamErr.(interface{ Code() int }); ok { + code = int64(codedErr.Code()) + } + f.handleStreamResultError(streamErr, &code, encoder) return } } @@ -819,6 +827,20 @@ func logIndexes(entries []*cstructs.AllocFileInfo, task, logType string) (indexT return indexTupleArray(indexes), nil } +type notFoundErr struct { + taskName string + logType string +} + +func (e notFoundErr) Error() string { + return fmt.Sprintf("log entry for task %q and log type %q not found", e.taskName, e.logType) +} + +// Code returns a 404 to avoid returning a 500 +func (e notFoundErr) Code() int { + return http.StatusNotFound +} + // findClosest takes a list of entries, the desired log index and desired log // offset (which can be negative, treated as offset from end), task name and log // type and returns the log entry, the log index, the offset to read from and a @@ -832,7 +854,7 @@ func findClosest(entries []*cstructs.AllocFileInfo, desiredIdx, desiredOffset in return nil, 0, 0, err } if len(indexes) == 0 { - return nil, 0, 0, fmt.Errorf("log entry for task %q and log type %q not found", task, logType) + return nil, 0, 0, notFoundErr{taskName: task, logType: logType} } // Binary search the indexes to get the desiredIdx diff --git a/client/fs_endpoint_test.go b/client/fs_endpoint_test.go index ec555fb6c..7258b3778 100644 --- a/client/fs_endpoint_test.go +++ b/client/fs_endpoint_test.go @@ -1175,23 +1175,12 @@ func TestFS_Logs(t *testing.T) { defer c.Shutdown() // Force an allocation onto the node - expected := "Hello from the other side" - a := mock.Alloc() - a.Job.Type = structs.JobTypeBatch - a.NodeID = c.NodeID() - a.Job.TaskGroups[0].Count = 1 - a.Job.TaskGroups[0].Tasks[0] = &structs.Task{ - Name: "web", - Driver: "mock_driver", - Config: map[string]interface{}{ - "run_for": "2s", - "stdout_string": expected, - }, - LogConfig: structs.DefaultLogConfig(), - Resources: &structs.Resources{ - CPU: 500, - MemoryMB: 256, - }, + expected := "Hello from the other side\n" + job := mock.BatchJob() + job.TaskGroups[0].Count = 1 + job.TaskGroups[0].Tasks[0].Config = map[string]interface{}{ + "run_for": "2s", + "stdout_string": expected, } // Wait for the client to connect @@ -1209,33 +1198,21 @@ func TestFS_Logs(t *testing.T) { t.Fatal(err) }) - // Upsert the allocation - state := s.State() - require.Nil(state.UpsertJob(999, a.Job)) - require.Nil(state.UpsertAllocs(1003, []*structs.Allocation{a})) + // Wait for client to be running job + testutil.WaitForRunning(t, s.RPC, job) - // Wait for the client to run the allocation - testutil.WaitForResult(func() (bool, error) { - alloc, err := state.AllocByID(nil, a.ID) - if err != nil { - return false, err - } - if alloc == nil { - return false, fmt.Errorf("unknown alloc") - } - if alloc.ClientStatus != structs.AllocClientStatusComplete { - return false, fmt.Errorf("alloc client status: %v", alloc.ClientStatus) - } - - return true, nil - }, func(err error) { - t.Fatalf("Alloc on node %q not finished: %v", c.NodeID(), err) - }) + // Get the allocation ID + args := structs.AllocListRequest{} + args.Region = "global" + resp := structs.AllocListResponse{} + require.NoError(s.RPC("Alloc.List", &args, &resp)) + require.Len(resp.Allocations, 1) + allocID := resp.Allocations[0].ID // Make the request req := &cstructs.FsLogsRequest{ - AllocID: a.ID, - Task: a.Job.TaskGroups[0].Tasks[0].Name, + AllocID: allocID, + Task: job.TaskGroups[0].Tasks[0].Name, LogType: "stdout", Origin: "start", PlainText: true, @@ -1315,71 +1292,33 @@ func TestFS_Logs_Follow(t *testing.T) { defer c.Shutdown() // Force an allocation onto the node - expectedBase := "Hello from the other side" + expectedBase := "Hello from the other side\n" repeat := 10 - a := mock.Alloc() - a.Job.Type = structs.JobTypeBatch - a.NodeID = c.NodeID() - a.Job.TaskGroups[0].Count = 1 - a.Job.TaskGroups[0].Tasks[0] = &structs.Task{ - Name: "web", - Driver: "mock_driver", - Config: map[string]interface{}{ - "run_for": "20s", - "stdout_string": expectedBase, - "stdout_repeat": repeat, - "stdout_repeat_duration": 200 * time.Millisecond, - }, - LogConfig: structs.DefaultLogConfig(), - Resources: &structs.Resources{ - CPU: 500, - MemoryMB: 256, - }, + job := mock.BatchJob() + job.TaskGroups[0].Count = 1 + job.TaskGroups[0].Tasks[0].Config = map[string]interface{}{ + "run_for": "20s", + "stdout_string": expectedBase, + "stdout_repeat": repeat, + "stdout_repeat_duration": 200 * time.Millisecond, } - // Wait for the client to connect - testutil.WaitForResult(func() (bool, error) { - node, err := s.State().NodeByID(nil, c.NodeID()) - if err != nil { - return false, err - } - if node == nil { - return false, fmt.Errorf("unknown node") - } + // Wait for client to be running job + testutil.WaitForRunning(t, s.RPC, job) - return node.Status == structs.NodeStatusReady, fmt.Errorf("bad node status") - }, func(err error) { - t.Fatal(err) - }) - - // Upsert the allocation - state := s.State() - require.Nil(state.UpsertJob(999, a.Job)) - require.Nil(state.UpsertAllocs(1003, []*structs.Allocation{a})) - - // Wait for the client to run the allocation - testutil.WaitForResult(func() (bool, error) { - alloc, err := state.AllocByID(nil, a.ID) - if err != nil { - return false, err - } - if alloc == nil { - return false, fmt.Errorf("unknown alloc") - } - if alloc.ClientStatus != structs.AllocClientStatusRunning { - return false, fmt.Errorf("alloc client status: %v", alloc.ClientStatus) - } - - return true, nil - }, func(err error) { - t.Fatalf("Alloc on node %q not running: %v", c.NodeID(), err) - }) + // Get the allocation ID + args := structs.AllocListRequest{} + args.Region = "global" + resp := structs.AllocListResponse{} + require.NoError(s.RPC("Alloc.List", &args, &resp)) + require.Len(resp.Allocations, 1) + allocID := resp.Allocations[0].ID // Make the request req := &cstructs.FsLogsRequest{ - AllocID: a.ID, - Task: a.Job.TaskGroups[0].Tasks[0].Name, + AllocID: allocID, + Task: job.TaskGroups[0].Tasks[0].Name, LogType: "stdout", Origin: "start", PlainText: true, @@ -1389,7 +1328,7 @@ func TestFS_Logs_Follow(t *testing.T) { // Get the handler handler, err := c.StreamingRpcHandler("FileSystem.Logs") - require.Nil(err) + require.NoError(err) // Create a pipe p1, p2 := net.Pipe() From 211b96bb5cd422cceab71bcb55c69fcefc8d3287 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 20 Sep 2018 15:44:27 -0700 Subject: [PATCH 093/152] tr: fix shutdown/destroy/WaitResult handling Multiple receivers raced for the WaitResult when killing tasks which could lead to a deadlock if the "wrong" receiver won. Wrap handlers in an ugly little proxy to avoid this. At first I wanted to push this into drivers, but the result is tied to the TR's handle lifecycle -- not the lifecycle of an alloc or task. --- .../allocrunnerv2/taskrunner/handleproxy.go | 57 ++++++++++++++ .../taskrunner/handleproxy_test.go | 76 +++++++++++++++++++ client/allocrunnerv2/taskrunner/lifecycle.go | 28 ++++--- .../allocrunnerv2/taskrunner/task_runner.go | 69 ++++++++++------- .../taskrunner/task_runner_getters.go | 9 ++- .../taskrunner/task_runner_hooks.go | 2 +- 6 files changed, 194 insertions(+), 47 deletions(-) create mode 100644 client/allocrunnerv2/taskrunner/handleproxy.go create mode 100644 client/allocrunnerv2/taskrunner/handleproxy_test.go diff --git a/client/allocrunnerv2/taskrunner/handleproxy.go b/client/allocrunnerv2/taskrunner/handleproxy.go new file mode 100644 index 000000000..88e4ab7d2 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/handleproxy.go @@ -0,0 +1,57 @@ +package taskrunner + +import ( + "context" + "sync" + + "github.com/hashicorp/nomad/client/driver/structs" +) + +// handleResult multiplexes a single WaitResult to multiple waiters. Useful +// because DriverHandle.WaitCh is closed after it returns a single WaitResult. +type handleResult struct { + doneCh <-chan struct{} + + result *structs.WaitResult + mu sync.RWMutex +} + +func newHandleResult(waitCh <-chan *structs.WaitResult) *handleResult { + doneCh := make(chan struct{}) + + h := &handleResult{ + doneCh: doneCh, + } + + go func() { + // Wait for result + res := <-waitCh + + // Set result + h.mu.Lock() + h.result = res + h.mu.Unlock() + + // Notify waiters + close(doneCh) + + }() + + return h +} + +// Wait blocks until a task's result is available or the passed-in context is +// canceled. Safe for concurrent callers. +func (h *handleResult) Wait(ctx context.Context) *structs.WaitResult { + // Block until done or canceled + select { + case <-h.doneCh: + case <-ctx.Done(): + } + + h.mu.RLock() + res := h.result + h.mu.RUnlock() + + return res +} diff --git a/client/allocrunnerv2/taskrunner/handleproxy_test.go b/client/allocrunnerv2/taskrunner/handleproxy_test.go new file mode 100644 index 000000000..5231d5297 --- /dev/null +++ b/client/allocrunnerv2/taskrunner/handleproxy_test.go @@ -0,0 +1,76 @@ +package taskrunner + +import ( + "context" + "testing" + "time" + + "github.com/hashicorp/nomad/client/driver/structs" + "github.com/stretchr/testify/require" +) + +// TestHandleResult_Wait_Result asserts multiple waiters on a handleResult all +// receive the wait result. +func TestHandleResult_Wait_Result(t *testing.T) { + t.Parallel() + + waitCh := make(chan *structs.WaitResult) + h := newHandleResult(waitCh) + + outCh1 := make(chan *structs.WaitResult) + outCh2 := make(chan *structs.WaitResult) + + // Create two recievers + go func() { + outCh1 <- h.Wait(context.Background()) + }() + go func() { + outCh2 <- h.Wait(context.Background()) + }() + + // Send a single result + go func() { + waitCh <- &structs.WaitResult{ExitCode: 1} + }() + + // Assert both receivers got the result + assert := func(outCh chan *structs.WaitResult) { + select { + case result := <-outCh: + require.NotNil(t, result) + require.Equal(t, 1, result.ExitCode) + case <-time.After(time.Second): + t.Fatalf("timeout waiting for result") + } + } + + assert(outCh1) + assert(outCh2) +} + +// TestHandleResult_Wait_Cancel asserts that canceling the context unblocks the +// waiter. +func TestHandleResult_Wait_Cancel(t *testing.T) { + t.Parallel() + + waitCh := make(chan *structs.WaitResult) + h := newHandleResult(waitCh) + + ctx, cancel := context.WithCancel(context.Background()) + outCh := make(chan *structs.WaitResult) + + go func() { + outCh <- h.Wait(ctx) + }() + + // Cancelling the context should unblock the Wait + cancel() + + // Assert the result is nil + select { + case result := <-outCh: + require.Nil(t, result) + case <-time.After(time.Second): + t.Fatalf("timeout waiting for result") + } +} diff --git a/client/allocrunnerv2/taskrunner/lifecycle.go b/client/allocrunnerv2/taskrunner/lifecycle.go index 848929636..4c9eabe10 100644 --- a/client/allocrunnerv2/taskrunner/lifecycle.go +++ b/client/allocrunnerv2/taskrunner/lifecycle.go @@ -7,9 +7,11 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) +// Restart a task. Returns immediately if no task is running. Blocks until +// existing task exits or passed-in context is canceled. func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error { // Grab the handle - handle := tr.getDriverHandle() + handle, result := tr.getDriverHandle() // Check it is running if handle == nil { @@ -29,19 +31,14 @@ func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, fai tr.logger.Error("failed to kill task. Resources may have been leaked", "error", err) } - // Drain the wait channel or wait for the request context to be cancelled - select { - case <-handle.WaitCh(): - case <-ctx.Done(): - return ctx.Err() - } - + // Drain the wait channel or wait for the request context to be canceled + result.Wait(ctx) return nil } func (tr *TaskRunner) Signal(event *structs.TaskEvent, s os.Signal) error { // Grab the handle - handle := tr.getDriverHandle() + handle, _ := tr.getDriverHandle() // Check it is running if handle == nil { @@ -58,8 +55,12 @@ func (tr *TaskRunner) Signal(event *structs.TaskEvent, s os.Signal) error { // Kill a task. Blocks until task exits or context is canceled. State is set to // dead. func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error { + // Cancel the task runner to break out of restart delay or the main run + // loop. + tr.ctxCancel() + // Grab the handle - handle := tr.getDriverHandle() + handle, result := tr.getDriverHandle() // Check if the handle is running if handle == nil { @@ -82,11 +83,8 @@ func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error tr.logger.Error("failed to kill task. Resources may have been leaked", "error", destroyErr) } - // Drain the wait channel or wait for the request context to be cancelled - select { - case <-handle.WaitCh(): - case <-ctx.Done(): - } + // Block until task has exited. + result.Wait(ctx) // Store that the task has been destroyed and any associated error. tr.UpdateState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskKilled).SetKillError(destroyErr)) diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 4b5bed7be..84578621b 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -16,6 +16,7 @@ import ( "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" + dstructs "github.com/hashicorp/nomad/client/driver/structs" cstate "github.com/hashicorp/nomad/client/state" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/client/vaultclient" @@ -72,8 +73,8 @@ type TaskRunner struct { // unnecessary writes persistedHash []byte - // ctx is the task runner's context and is done whe the task runner - // should exit. Shutdown hooks are run. + // ctx is the task runner's context representing the tasks's lifecycle. + // Canceling the context will cause the task to be destroyed. ctx context.Context // ctxCancel is used to exit the task runner's Run loop without @@ -95,8 +96,9 @@ type TaskRunner struct { // driver is the driver for the task. driver driver.Driver - handle driver.DriverHandle // the handle to the running driver - handleLock sync.Mutex + handle driver.DriverHandle // the handle to the running driver + handleResult *handleResult // proxy for handle results + handleLock sync.Mutex // task is the task being run task *structs.Task @@ -259,7 +261,7 @@ func (tr *TaskRunner) initLabels() { func (tr *TaskRunner) Run() { defer close(tr.waitCh) - var handle driver.DriverHandle + var waitRes *dstructs.WaitResult // Updates are handled asynchronously with the other hooks but each // triggered update - whether due to alloc updates or a new vault token @@ -291,42 +293,48 @@ MAIN: tr.logger.Error("poststart failed", "error", err) } - // Grab the handle - handle = tr.getDriverHandle() + // Grab the result proxy and wait for task to exit + { + _, result := tr.getDriverHandle() - select { - case waitRes := <-handle.WaitCh(): - // Clear the handle - tr.clearDriverHandle() - - // Store the wait result on the restart tracker - tr.restartTracker.SetWaitResult(waitRes) - case <-tr.ctx.Done(): - tr.logger.Debug("task killed") + // Do *not* use tr.ctx here as it would cause Wait() to + // unblock before the task exits when Kill() is called. + waitRes = result.Wait(context.Background()) } + // Clear the handle + tr.clearDriverHandle() + + // Store the wait result on the restart tracker + tr.restartTracker.SetWaitResult(waitRes) + if err := tr.exited(); err != nil { tr.logger.Error("exited hooks failed", "error", err) } RESTART: - // Actually restart by sleeping and also watching for destroy events - restart, restartWait := tr.shouldRestart() + restart, restartDelay := tr.shouldRestart() if !restart { break MAIN } - deadline := time.Now().Add(restartWait) - timer := time.NewTimer(restartWait) - for time.Now().Before(deadline) { - select { - case <-timer.C: - case <-tr.ctx.Done(): - tr.logger.Debug("task runner cancelled") - break MAIN - } + // Actually restart by sleeping and also watching for destroy events + select { + case <-time.After(restartDelay): + case <-tr.ctx.Done(): + tr.logger.Trace("task killed between restarts", "delay", restartDelay) + break MAIN } - timer.Stop() + } + + // If task terminated, update server. All other exit conditions (eg + // killed or out of restarts) will perform their own server updates. + if waitRes != nil { + event := structs.NewTaskEvent(structs.TaskTerminated). + SetExitCode(waitRes.ExitCode). + SetSignal(waitRes.Signal). + SetExitMessage(waitRes.Err) + tr.UpdateState(structs.TaskStateDead, event) } // Run the stop hooks @@ -361,13 +369,16 @@ func (tr *TaskRunner) handleUpdates() { } } +// shouldRestart determines whether the task should be restarted and updates +// the task state unless the task is killed or terminated. func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { // Determine if we should restart state, when := tr.restartTracker.GetState() reason := tr.restartTracker.GetReason() switch state { case structs.TaskKilled: - // The task was killed. Nothing to do + // Never restart an explicitly killed task. Kill method handles + // updating the server. return false, 0 case structs.TaskNotRestarting, structs.TaskTerminated: tr.logger.Info("not restarting task", "reason", reason) diff --git a/client/allocrunnerv2/taskrunner/task_runner_getters.go b/client/allocrunnerv2/taskrunner/task_runner_getters.go index 8a40c65bc..abce06593 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_getters.go +++ b/client/allocrunnerv2/taskrunner/task_runner_getters.go @@ -49,20 +49,25 @@ func (tr *TaskRunner) setVaultToken(token string) { tr.envBuilder.SetVaultToken(token, tr.task.Vault.Env) } -func (tr *TaskRunner) getDriverHandle() driver.DriverHandle { +// getDriverHandle returns a driver handle and its result proxy. Use the +// result proxy instead of the handle's WaitCh. +func (tr *TaskRunner) getDriverHandle() (driver.DriverHandle, *handleResult) { tr.handleLock.Lock() defer tr.handleLock.Unlock() - return tr.handle + return tr.handle, tr.handleResult } +// setDriverHanlde sets the driver handle and creates a new result proxy. func (tr *TaskRunner) setDriverHandle(handle driver.DriverHandle) { tr.handleLock.Lock() defer tr.handleLock.Unlock() tr.handle = handle + tr.handleResult = newHandleResult(handle.WaitCh()) } func (tr *TaskRunner) clearDriverHandle() { tr.handleLock.Lock() defer tr.handleLock.Unlock() tr.handle = nil + tr.handleResult = nil } diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunnerv2/taskrunner/task_runner_hooks.go index 4e3d550f7..14caba832 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunnerv2/taskrunner/task_runner_hooks.go @@ -168,7 +168,7 @@ func (tr *TaskRunner) poststart() error { }() } - handle := tr.getDriverHandle() + handle, _ := tr.getDriverHandle() net := handle.Network() var merr multierror.Error From 1c9ccdeab5d37bd43462ba2fd8c567f2b5afe591 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 20 Sep 2018 15:47:38 -0700 Subject: [PATCH 094/152] tests: fix races caused by sharing a buffer httptest.ResponseRecorder exposes a bytes.Buffer which we were reading and writing concurrently to test streaming log APIs. This is a race, so I wrapped the struct in a lock with some helpers. --- command/agent/fs_endpoint_test.go | 29 ++++++------ testutil/responsewriter.go | 76 +++++++++++++++++++++++++++++++ 2 files changed, 89 insertions(+), 16 deletions(-) create mode 100644 testutil/responsewriter.go diff --git a/command/agent/fs_endpoint_test.go b/command/agent/fs_endpoint_test.go index 9be39497c..538203c9e 100644 --- a/command/agent/fs_endpoint_test.go +++ b/command/agent/fs_endpoint_test.go @@ -145,25 +145,22 @@ func TestHTTP_FS_ReadAt_MissingParams(t *testing.T) { require := require.New(t) httpTest(t, nil, func(s *TestAgent) { req, err := http.NewRequest("GET", "/v1/client/fs/readat/", nil) - require.Nil(err) - respW := httptest.NewRecorder() + require.NoError(err) - _, err = s.Server.FileReadAtRequest(respW, req) - require.NotNil(err) + _, err = s.Server.FileReadAtRequest(httptest.NewRecorder(), req) + require.Error(err) req, err = http.NewRequest("GET", "/v1/client/fs/readat/foo", nil) - require.Nil(err) - respW = httptest.NewRecorder() + require.NoError(err) - _, err = s.Server.FileReadAtRequest(respW, req) - require.NotNil(err) + _, err = s.Server.FileReadAtRequest(httptest.NewRecorder(), req) + require.Error(err) req, err = http.NewRequest("GET", "/v1/client/fs/readat/foo?path=/path/to/file", nil) - require.Nil(err) - respW = httptest.NewRecorder() + require.NoError(err) - _, err = s.Server.FileReadAtRequest(respW, req) - require.NotNil(err) + _, err = s.Server.FileReadAtRequest(httptest.NewRecorder(), req) + require.Error(err) }) } @@ -398,7 +395,7 @@ func TestHTTP_FS_Logs(t *testing.T) { p, _ := io.Pipe() req, err := http.NewRequest("GET", path, p) require.Nil(err) - respW := httptest.NewRecorder() + respW := testutil.NewResponseRecorder() go func() { _, err = s.Server.Logs(respW, req) require.Nil(err) @@ -406,7 +403,7 @@ func TestHTTP_FS_Logs(t *testing.T) { out := "" testutil.WaitForResult(func() (bool, error) { - output, err := ioutil.ReadAll(respW.Body) + output, err := ioutil.ReadAll(respW) if err != nil { return false, err } @@ -436,7 +433,7 @@ func TestHTTP_FS_Logs_Follow(t *testing.T) { p, _ := io.Pipe() req, err := http.NewRequest("GET", path, p) require.Nil(err) - respW := httptest.NewRecorder() + respW := testutil.NewResponseRecorder() errCh := make(chan error) go func() { _, err := s.Server.Logs(respW, req) @@ -445,7 +442,7 @@ func TestHTTP_FS_Logs_Follow(t *testing.T) { out := "" testutil.WaitForResult(func() (bool, error) { - output, err := ioutil.ReadAll(respW.Body) + output, err := ioutil.ReadAll(respW) if err != nil { return false, err } diff --git a/testutil/responsewriter.go b/testutil/responsewriter.go new file mode 100644 index 000000000..20bb38bb6 --- /dev/null +++ b/testutil/responsewriter.go @@ -0,0 +1,76 @@ +package testutil + +import ( + "net/http" + "net/http/httptest" + "sync" +) + +// assert ResponseRecorder implements the http.ResponseWriter interface +var _ http.ResponseWriter = (*ResponseRecorder)(nil) + +// ResponseRecorder implements a ResponseWriter which can be written to and +// read from concurrently. For use in testing streaming APIs where +// httptest.ResponseRecorder is unsafe for concurrent access. Uses +// httptest.ResponseRecorder internally and exposes most of the functionality. +type ResponseRecorder struct { + rr *httptest.ResponseRecorder + mu sync.Mutex +} + +func NewResponseRecorder() *ResponseRecorder { + return &ResponseRecorder{ + rr: httptest.NewRecorder(), + } +} + +// Flush sets Flushed=true. +func (r *ResponseRecorder) Flush() { + r.mu.Lock() + defer r.mu.Unlock() + r.rr.Flush() +} + +// Flushed returns true if Flush has been called. +func (r *ResponseRecorder) Flushed() bool { + r.mu.Lock() + defer r.mu.Unlock() + return r.rr.Flushed +} + +// Header returns the response headers. Readers should call HeaderMap() to +// avoid races due to the server concurrently mutating headers. +func (r *ResponseRecorder) Header() http.Header { + r.mu.Lock() + defer r.mu.Unlock() + return r.rr.Header() +} + +// HeaderMap returns the HTTP headers written before WriteHeader was called. +func (r *ResponseRecorder) HeaderMap() http.Header { + r.mu.Lock() + defer r.mu.Unlock() + return r.rr.HeaderMap +} + +// Write to the underlying response buffer. Safe to call concurrent with Read. +func (r *ResponseRecorder) Write(p []byte) (int, error) { + r.mu.Lock() + defer r.mu.Unlock() + return r.rr.Body.Write(p) +} + +// WriteHeader sets the response code and freezes the headers returned by +// HeaderMap. Safe to call concurrent with Read and HeaderMap. +func (r *ResponseRecorder) WriteHeader(statusCode int) { + r.mu.Lock() + defer r.mu.Unlock() + r.rr.WriteHeader(statusCode) +} + +// Read available response bytes. Safe to call concurrently with Write(). +func (r *ResponseRecorder) Read(p []byte) (int, error) { + r.mu.Lock() + defer r.mu.Unlock() + return r.rr.Body.Read(p) +} From 4b44b9039bdd8e8fe861256bfe9a55252ef1d8ab Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 25 Sep 2018 11:03:10 -0700 Subject: [PATCH 095/152] tr: remove unneeded lock; chan synchronizes access --- client/allocrunnerv2/taskrunner/handleproxy.go | 12 ++---------- 1 file changed, 2 insertions(+), 10 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/handleproxy.go b/client/allocrunnerv2/taskrunner/handleproxy.go index 88e4ab7d2..3b24f7d2f 100644 --- a/client/allocrunnerv2/taskrunner/handleproxy.go +++ b/client/allocrunnerv2/taskrunner/handleproxy.go @@ -2,7 +2,6 @@ package taskrunner import ( "context" - "sync" "github.com/hashicorp/nomad/client/driver/structs" ) @@ -11,9 +10,7 @@ import ( // because DriverHandle.WaitCh is closed after it returns a single WaitResult. type handleResult struct { doneCh <-chan struct{} - result *structs.WaitResult - mu sync.RWMutex } func newHandleResult(waitCh <-chan *structs.WaitResult) *handleResult { @@ -28,9 +25,7 @@ func newHandleResult(waitCh <-chan *structs.WaitResult) *handleResult { res := <-waitCh // Set result - h.mu.Lock() h.result = res - h.mu.Unlock() // Notify waiters close(doneCh) @@ -47,11 +42,8 @@ func (h *handleResult) Wait(ctx context.Context) *structs.WaitResult { select { case <-h.doneCh: case <-ctx.Done(): + return nil } - h.mu.RLock() - res := h.result - h.mu.RUnlock() - - return res + return h.result } From 9a7e6be2b662d04789b7964d6e41475917f5a825 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 25 Sep 2018 13:43:58 -0700 Subject: [PATCH 096/152] client: fix potentially dropped streaming errors --- client/fs_endpoint.go | 9 +++++++++ 1 file changed, 9 insertions(+) diff --git a/client/fs_endpoint.go b/client/fs_endpoint.go index 41123dbf4..307ccb8e7 100644 --- a/client/fs_endpoint.go +++ b/client/fs_endpoint.go @@ -276,6 +276,15 @@ OUTER: break OUTER case frame, ok := <-frames: if !ok { + // frame may have been closed when an error + // occurred. Check once more for an error. + select { + case streamErr = <-errCh: + // There was a pending error! + default: + // No error, continue on + } + break OUTER } From fb4aa74153f68a73be760b2f00fb2c49d092396a Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 25 Sep 2018 13:52:54 -0700 Subject: [PATCH 097/152] client: add comment --- client/fs_endpoint.go | 3 +++ 1 file changed, 3 insertions(+) diff --git a/client/fs_endpoint.go b/client/fs_endpoint.go index 307ccb8e7..95f1a9695 100644 --- a/client/fs_endpoint.go +++ b/client/fs_endpoint.go @@ -836,6 +836,9 @@ func logIndexes(entries []*cstructs.AllocFileInfo, task, logType string) (indexT return indexTupleArray(indexes), nil } +// notFoundErr is returned when a log is requested but cannot be found. +// Implements agent.HTTPCodedError but does not reference it to avoid circular +// imports. type notFoundErr struct { taskName string logType string From 960f3be76ce928adc01ed95b1ff827c7d44691a3 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 26 Sep 2018 17:08:43 -0700 Subject: [PATCH 098/152] client: expose task state to client The interesting decision in this commit was to expose AR's state and not a fully materialized Allocation struct. AR.clientAlloc builds an Alloc that contains the task state, so I considered simply memoizing and exposing that method. However, that would lead to AR having two awkwardly similar methods: - Alloc() - which returns the server-sent alloc - ClientAlloc() - which returns the fully materialized client alloc Since ClientAlloc() could be memoized it would be just as cheap to call as Alloc(), so why not replace Alloc() entirely? Replacing Alloc() entirely would require Update() to immediately materialize the task states on server-sent Allocs as there may have been local task state changes since the server received an Alloc update. This quickly becomes difficult to reason about: should Update hooks use the TaskStates? Are state changes caused by TR Update hooks immediately reflected in the Alloc? Should AR persist its copy of the Alloc? If so, are its TaskStates canonical or the TaskStates on TR? So! Forget that. Let's separate the static Allocation from the dynamic AR & TR state! - AR.Alloc() is for static Allocation access (often for the Job) - AR.AllocState() is for the dynamic AR & TR runtime state (deployment status, task states, etc). If code needs to know the status of a task: AllocState() If code needs to know the names of tasks: Alloc() It should be very easy for a developer to reason about which method they should call and what they can do with the return values. --- client/allocrunnerv2/alloc_runner.go | 35 +++++++++++++++---- client/allocrunnerv2/state/state.go | 18 +++++++++- .../allocrunnerv2/taskrunner/task_runner.go | 21 ++++++----- client/client.go | 9 +++-- client/fs_endpoint.go | 20 ++++++----- command/agent/helpers.go | 2 +- 6 files changed, 74 insertions(+), 31 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index b4b47326f..8b1ace5f4 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -56,9 +56,7 @@ type allocRunner struct { alloc *structs.Allocation allocLock sync.RWMutex - //XXX implement for local state - // state captures the state of the alloc runner - state *state.State + state *state.State // alloc runner state stateLock sync.RWMutex stateDB cstate.StateDB @@ -305,20 +303,22 @@ func (ar *allocRunner) clientAlloc(taskStates map[string]*structs.TaskState) *st ar.stateLock.RLock() defer ar.stateLock.RUnlock() + // store task states for AllocState to expose + ar.state.TaskStates = taskStates + a := &structs.Allocation{ ID: ar.id, TaskStates: taskStates, } - s := ar.state - if d := s.DeploymentStatus; d != nil { + if d := ar.state.DeploymentStatus; d != nil { a.DeploymentStatus = d.Copy() } // Compute the ClientStatus - if s.ClientStatus != "" { + if ar.state.ClientStatus != "" { // The client status is being forced - a.ClientStatus, a.ClientDescription = s.ClientStatus, s.ClientDescription + a.ClientStatus, a.ClientDescription = ar.state.ClientStatus, ar.state.ClientDescription } else { a.ClientStatus, a.ClientDescription = getClientStatus(taskStates) } @@ -388,6 +388,27 @@ func getClientStatus(taskStates map[string]*structs.TaskState) (status, descript return "", "" } +// AllocState returns a copy of allocation state including a snapshot of task +// states. +func (ar *allocRunner) AllocState() *state.State { + // Must acquire write-lock in case TaskStates needs to be set. + ar.stateLock.Lock() + defer ar.stateLock.Unlock() + + // If TaskStateUpdated has not been called yet, ar.state.TaskStates + // won't be set as it is not the canonical source of TaskStates. + if len(ar.state.TaskStates) == 0 { + ar.tasksLock.RLock() + ar.state.TaskStates = make(map[string]*structs.TaskState, len(ar.tasks)) + for k, tr := range ar.tasks { + ar.state.TaskStates[k] = tr.TaskState() + } + ar.tasksLock.RUnlock() + } + + return ar.state.Copy() +} + // Update the running allocation with a new version received from the server. // // This method sends the updated alloc to Run for serially processing updates. diff --git a/client/allocrunnerv2/state/state.go b/client/allocrunnerv2/state/state.go index 85ec7c3da..9754e369c 100644 --- a/client/allocrunnerv2/state/state.go +++ b/client/allocrunnerv2/state/state.go @@ -6,7 +6,6 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) -// XXX Why its own package? // State captures the state of the allocation runner. type State struct { // ClientStatus captures the overall state of the allocation @@ -18,6 +17,9 @@ type State struct { // DeploymentStatus captures the status of the deployment DeploymentStatus *structs.AllocDeploymentStatus + + // TaskStates is a snapshot of task states. + TaskStates map[string]*structs.TaskState } // SetDeploymentStatus is a helper for updating the client-controlled @@ -43,3 +45,17 @@ func (s *State) ClearDeploymentStatus() { s.DeploymentStatus.Healthy = nil s.DeploymentStatus.Timestamp = time.Time{} } + +// Copy returns a deep copy of State. +func (s *State) Copy() *State { + taskStates := make(map[string]*structs.TaskState, len(s.TaskStates)) + for k, v := range s.TaskStates { + taskStates[k] = v.Copy() + } + return &State{ + ClientStatus: s.ClientStatus, + ClientDescription: s.ClientDescription, + DeploymentStatus: s.DeploymentStatus.Copy(), + TaskStates: taskStates, + } +} diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index 84578621b..c206860e2 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -177,17 +177,16 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { ) tr := &TaskRunner{ - alloc: config.Alloc, - allocID: config.Alloc.ID, - clientConfig: config.ClientConfig, - task: config.Task, - taskDir: config.TaskDir, - taskName: config.Task.Name, - envBuilder: envBuilder, - consulClient: config.Consul, - vaultClient: config.VaultClient, - //XXX Make a Copy to avoid races? - state: config.Alloc.TaskStates[config.Task.Name], + alloc: config.Alloc, + allocID: config.Alloc.ID, + clientConfig: config.ClientConfig, + task: config.Task, + taskDir: config.TaskDir, + taskName: config.Task.Name, + envBuilder: envBuilder, + consulClient: config.Consul, + vaultClient: config.VaultClient, + state: config.Alloc.TaskStates[config.Task.Name].Copy(), localState: config.LocalState, stateDB: config.StateDB, stateUpdater: config.StateUpdater, diff --git a/client/client.go b/client/client.go index 1056520fd..2fbeaf7c5 100644 --- a/client/client.go +++ b/client/client.go @@ -18,6 +18,7 @@ import ( consulapi "github.com/hashicorp/consul/api" hclog "github.com/hashicorp/go-hclog" multierror "github.com/hashicorp/go-multierror" + arstate "github.com/hashicorp/nomad/client/allocrunnerv2/state" consulApi "github.com/hashicorp/nomad/client/consul" cstructs "github.com/hashicorp/nomad/client/structs" hstats "github.com/hashicorp/nomad/helper/stats" @@ -103,6 +104,7 @@ type ClientStatsReporter interface { //TODO Create via factory to allow testing Client with mock AllocRunners. type AllocRunner interface { Alloc() *structs.Allocation + AllocState() *arstate.State Destroy() GetAllocDir() *allocdir.AllocDir IsDestroyed() bool @@ -637,8 +639,9 @@ func (c *Client) GetAllocFS(allocID string) (allocdir.AllocDirFS, error) { return ar.GetAllocDir(), nil } -// GetClientAlloc returns the allocation from the client -func (c *Client) GetClientAlloc(allocID string) (*structs.Allocation, error) { +// GetAllocState returns a copy of an allocation's state on this client. It +// returns either an AllocState or an unknown allocation error. +func (c *Client) GetAllocState(allocID string) (*arstate.State, error) { c.allocLock.RLock() ar, ok := c.allocs[allocID] c.allocLock.RUnlock() @@ -646,7 +649,7 @@ func (c *Client) GetClientAlloc(allocID string) (*structs.Allocation, error) { return nil, structs.NewErrUnknownAllocation(allocID) } - return ar.Alloc(), nil + return ar.AllocState(), nil } // GetServers returns the list of nomad servers this client is aware of. diff --git a/client/fs_endpoint.go b/client/fs_endpoint.go index 95f1a9695..ded3ae0aa 100644 --- a/client/fs_endpoint.go +++ b/client/fs_endpoint.go @@ -380,7 +380,7 @@ func (f *FileSystem) logs(conn io.ReadWriteCloser) { return } - alloc, err := f.c.GetClientAlloc(req.AllocID) + allocState, err := f.c.GetAllocState(req.AllocID) if err != nil { code := helper.Int64ToPtr(500) if structs.IsErrUnknownAllocation(err) { @@ -392,19 +392,23 @@ func (f *FileSystem) logs(conn io.ReadWriteCloser) { } // Check that the task is there - tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup) - if tg == nil { - f.handleStreamResultError(fmt.Errorf("Failed to lookup task group for allocation"), - helper.Int64ToPtr(500), encoder) - return - } else if taskStruct := tg.LookupTask(req.Task); taskStruct == nil { + taskState := allocState.TaskStates[req.Task] + if taskState == nil { f.handleStreamResultError( - fmt.Errorf("task group %q does not have task with name %q", alloc.TaskGroup, req.Task), + fmt.Errorf("unknown task name %q", req.Task), helper.Int64ToPtr(400), encoder) return } + if taskState.StartedAt.IsZero() { + f.handleStreamResultError( + fmt.Errorf("task %q not started yet. No logs available", req.Task), + helper.Int64ToPtr(404), + encoder) + return + } + ctx, cancel := context.WithCancel(context.Background()) defer cancel() diff --git a/command/agent/helpers.go b/command/agent/helpers.go index 50542416c..7248ea5f8 100644 --- a/command/agent/helpers.go +++ b/command/agent/helpers.go @@ -12,7 +12,7 @@ func (s *HTTPServer) rpcHandlerForAlloc(allocID string) (localClient, remoteClie if c != nil { // If there is an error it means that the client doesn't have the // allocation so we can't use the local client - _, err := c.GetClientAlloc(allocID) + _, err := c.GetAllocState(allocID) if err == nil { localAlloc = true } From 6bcf772f3c5ff024aa0fa0d620cf8a4a7013f2ed Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 26 Sep 2018 17:25:20 -0700 Subject: [PATCH 099/152] tests: test via ServeMux so http codes are set --- command/agent/fs_endpoint_test.go | 25 +++++++++++++++++-------- 1 file changed, 17 insertions(+), 8 deletions(-) diff --git a/command/agent/fs_endpoint_test.go b/command/agent/fs_endpoint_test.go index 538203c9e..7d7e4e784 100644 --- a/command/agent/fs_endpoint_test.go +++ b/command/agent/fs_endpoint_test.go @@ -211,37 +211,46 @@ func TestHTTP_FS_Stream_MissingParams(t *testing.T) { }) } +// TestHTTP_FS_Logs_MissingParams asserts proper error codes and messages are +// returned for incorrect parameters (eg missing tasks). func TestHTTP_FS_Logs_MissingParams(t *testing.T) { t.Parallel() require := require.New(t) httpTest(t, nil, func(s *TestAgent) { + // AllocID Not Present req, err := http.NewRequest("GET", "/v1/client/fs/logs/", nil) require.Nil(err) respW := httptest.NewRecorder() - _, err = s.Server.Logs(respW, req) - require.EqualError(err, allocIDNotPresentErr.Error()) + s.Server.mux.ServeHTTP(respW, req) + require.Equal(respW.Body.String(), allocIDNotPresentErr.Error()) + require.Equal(500, respW.Code) // 500 for backward compat + // Task Not Present req, err = http.NewRequest("GET", "/v1/client/fs/logs/foo", nil) require.Nil(err) respW = httptest.NewRecorder() - _, err = s.Server.Logs(respW, req) - require.EqualError(err, taskNotPresentErr.Error()) + s.Server.mux.ServeHTTP(respW, req) + require.Equal(respW.Body.String(), taskNotPresentErr.Error()) + require.Equal(500, respW.Code) // 500 for backward compat + // Log Type Not Present req, err = http.NewRequest("GET", "/v1/client/fs/logs/foo?task=foo", nil) require.Nil(err) respW = httptest.NewRecorder() - _, err = s.Server.Logs(respW, req) - require.EqualError(err, logTypeNotPresentErr.Error()) + s.Server.mux.ServeHTTP(respW, req) + require.Equal(respW.Body.String(), logTypeNotPresentErr.Error()) + require.Equal(500, respW.Code) // 500 for backward compat + // Ok req, err = http.NewRequest("GET", "/v1/client/fs/logs/foo?task=foo&type=stdout", nil) require.Nil(err) respW = httptest.NewRecorder() - _, err = s.Server.Logs(respW, req) - require.Nil(err) + s.Server.mux.ServeHTTP(respW, req) + require.Equal(200, respW.Code) }) } From 1d747048ea18b9f62a9e62b7e8d8b4d0bffa2ab1 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 26 Sep 2018 17:29:08 -0700 Subject: [PATCH 100/152] tests: ensure task state is initialized in NewAR Also expose NoopDB for use in tests. --- client/allocrunnerv2/alloc_runner_test.go | 38 +++++++++++++++++++++++ client/state/noopdb.go | 19 ++++++------ client/state/state_database.go | 2 +- 3 files changed, 49 insertions(+), 10 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner_test.go b/client/allocrunnerv2/alloc_runner_test.go index 4ec222614..99460f758 100644 --- a/client/allocrunnerv2/alloc_runner_test.go +++ b/client/allocrunnerv2/alloc_runner_test.go @@ -1,5 +1,43 @@ package allocrunnerv2 +import ( + "testing" + + "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/state" + "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/nomad/mock" + "github.com/stretchr/testify/require" +) + +// TestAllocRunner_AllocState_Initialized asserts that getting TaskStates via +// AllocState() are initialized even before the AllocRunner has run. +func TestAllocRunner_AllocState_Initialized(t *testing.T) { + t.Parallel() + + alloc := mock.Alloc() + logger := testlog.HCLogger(t) + + conf := &Config{ + Alloc: alloc, + Logger: logger, + ClientConfig: config.TestClientConfig(), + StateDB: state.NoopDB{}, + Consul: nil, + Vault: nil, + StateUpdater: nil, + PrevAllocWatcher: nil, + } + + ar, err := NewAllocRunner(conf) + require.NoError(t, err) + + allocState := ar.AllocState() + + require.NotNil(t, allocState) + require.NotNil(t, allocState.TaskStates[alloc.Job.TaskGroups[0].Tasks[0].Name]) +} + /* import ( diff --git a/client/state/noopdb.go b/client/state/noopdb.go index 65d105fd3..a98461e88 100644 --- a/client/state/noopdb.go +++ b/client/state/noopdb.go @@ -5,36 +5,37 @@ import ( "github.com/hashicorp/nomad/nomad/structs" ) -type noopDB struct{} +// NoopDB implements a StateDB that does not persist any data. +type NoopDB struct{} -func (n noopDB) GetAllAllocations() ([]*structs.Allocation, map[string]error, error) { +func (n NoopDB) GetAllAllocations() ([]*structs.Allocation, map[string]error, error) { return nil, nil, nil } -func (n noopDB) PutAllocation(*structs.Allocation) error { +func (n NoopDB) PutAllocation(*structs.Allocation) error { return nil } -func (n noopDB) GetTaskRunnerState(allocID string, taskName string) (*state.LocalState, *structs.TaskState, error) { +func (n NoopDB) GetTaskRunnerState(allocID string, taskName string) (*state.LocalState, *structs.TaskState, error) { return nil, nil, nil } -func (n noopDB) PutTaskRunnerLocalState(allocID string, taskName string, val interface{}) error { +func (n NoopDB) PutTaskRunnerLocalState(allocID string, taskName string, val interface{}) error { return nil } -func (n noopDB) PutTaskState(allocID string, taskName string, state *structs.TaskState) error { +func (n NoopDB) PutTaskState(allocID string, taskName string, state *structs.TaskState) error { return nil } -func (n noopDB) DeleteTaskBucket(allocID, taskName string) error { +func (n NoopDB) DeleteTaskBucket(allocID, taskName string) error { return nil } -func (n noopDB) DeleteAllocationBucket(allocID string) error { +func (n NoopDB) DeleteAllocationBucket(allocID string) error { return nil } -func (n noopDB) Close() error { +func (n NoopDB) Close() error { return nil } diff --git a/client/state/state_database.go b/client/state/state_database.go index 50f811209..ba90da2d0 100644 --- a/client/state/state_database.go +++ b/client/state/state_database.go @@ -46,7 +46,7 @@ func GetStateDBFactory(devMode bool) NewStateDBFunc { // Return a noop state db implementation when in debug mode if devMode { return func(string) (StateDB, error) { - return noopDB{}, nil + return NoopDB{}, nil } } From dd4227f84af3fbd4207664be63ae0c491686d90e Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 26 Sep 2018 17:29:54 -0700 Subject: [PATCH 101/152] tests: make a test client/config easier to generate Sadly can't move the fingerprint timeout tweak into the helper due to circular imports. --- client/config/testing.go | 23 +++++++++++++++++++++++ client/testing.go | 21 +++------------------ 2 files changed, 26 insertions(+), 18 deletions(-) create mode 100644 client/config/testing.go diff --git a/client/config/testing.go b/client/config/testing.go new file mode 100644 index 000000000..c98cad8c9 --- /dev/null +++ b/client/config/testing.go @@ -0,0 +1,23 @@ +package config + +import ( + "github.com/hashicorp/nomad/helper" + "github.com/hashicorp/nomad/nomad/structs" +) + +// TestClientConfig returns a default client configuration for test clients. +func TestClientConfig() *Config { + conf := DefaultConfig() + conf.VaultConfig.Enabled = helper.BoolToPtr(false) + conf.DevMode = true + conf.Node = &structs.Node{ + Reserved: &structs.Resources{ + DiskMB: 0, + }, + } + + // Loosen GC threshold + conf.GCDiskUsageThreshold = 98.0 + conf.GCInodeUsageThreshold = 98.0 + return conf +} diff --git a/client/testing.go b/client/testing.go index a67caa0dd..deb19eda2 100644 --- a/client/testing.go +++ b/client/testing.go @@ -5,9 +5,6 @@ import ( consulApi "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/fingerprint" "github.com/hashicorp/nomad/command/agent/consul" - "github.com/hashicorp/nomad/helper" - "github.com/hashicorp/nomad/helper/testlog" - "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/plugins/shared/catalog" "github.com/hashicorp/nomad/plugins/shared/singleton" "github.com/mitchellh/go-testing-interface" @@ -15,22 +12,10 @@ import ( // TestClient creates an in-memory client for testing purposes. func TestClient(t testing.T, cb func(c *config.Config)) *Client { - conf := config.DefaultConfig() - logger := testlog.HCLogger(t) - conf.Logger = logger - conf.VaultConfig.Enabled = helper.BoolToPtr(false) - conf.DevMode = true - conf.Node = &structs.Node{ - Reserved: &structs.Resources{ - DiskMB: 0, - }, - } + conf := config.TestClientConfig() - // Loosen GC threshold - conf.GCDiskUsageThreshold = 98.0 - conf.GCInodeUsageThreshold = 98.0 - - // Tighten the fingerprinter timeouts + // Tighten the fingerprinter timeouts (must be done in client package + // to avoid circular dependencies) if conf.Options == nil { conf.Options = make(map[string]string) } From f279b1d1b16a129c8bc4203a18fc7ba8301807ca Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 26 Sep 2018 17:33:32 -0700 Subject: [PATCH 102/152] tests: test logs endpoint against pending task Although the really exciting change is making WaitForRunning return the allocations that it started. This should cut down test boilerplate significantly. --- client/driver/mock_driver.go | 3 +- client/fs_endpoint_test.go | 338 +++++++++++---------------- command/agent/alloc_endpoint_test.go | 2 +- testutil/wait.go | 52 ++--- 4 files changed, 170 insertions(+), 225 deletions(-) diff --git a/client/driver/mock_driver.go b/client/driver/mock_driver.go index f17c9d8ab..11b502518 100644 --- a/client/driver/mock_driver.go +++ b/client/driver/mock_driver.go @@ -45,7 +45,8 @@ type MockDriverConfig struct { // StartErrRecoverable marks the error returned is recoverable StartErrRecoverable bool `mapstructure:"start_error_recoverable"` - // StartBlockFor specifies a duration in which to block before returning + // StartBlockFor specifies a duration in which to block Start before + // returning. Useful for testing the behavior of tasks in pending. StartBlockFor time.Duration `mapstructure:"start_block_for"` // KillAfter is the duration after which the mock driver indicates the task diff --git a/client/fs_endpoint_test.go b/client/fs_endpoint_test.go index 7258b3778..b9e141acb 100644 --- a/client/fs_endpoint_test.go +++ b/client/fs_endpoint_test.go @@ -32,7 +32,7 @@ import ( // tempAllocDir returns a new alloc dir that is rooted in a temp dir. The caller // should destroy the temp dir. func tempAllocDir(t testing.TB) *allocdir.AllocDir { - dir, err := ioutil.TempDir("", "") + dir, err := ioutil.TempDir("", "nomadtest") if err != nil { t.Fatalf("TempDir() failed: %v", err) } @@ -536,68 +536,21 @@ func TestFS_Stream(t *testing.T) { }) defer c.Shutdown() - // Force an allocation onto the node expected := "Hello from the other side" - a := mock.Alloc() - a.Job.Type = structs.JobTypeBatch - a.NodeID = c.NodeID() - a.Job.TaskGroups[0].Count = 1 - a.Job.TaskGroups[0].Tasks[0] = &structs.Task{ - Name: "web", - Driver: "mock_driver", - Config: map[string]interface{}{ - "run_for": "2s", - "stdout_string": expected, - }, - LogConfig: structs.DefaultLogConfig(), - Resources: &structs.Resources{ - CPU: 500, - MemoryMB: 256, - }, + job := mock.BatchJob() + job.TaskGroups[0].Count = 1 + job.TaskGroups[0].Tasks[0].Config = map[string]interface{}{ + "run_for": "2s", + "stdout_string": expected, } - // Wait for the client to connect - testutil.WaitForResult(func() (bool, error) { - node, err := s.State().NodeByID(nil, c.NodeID()) - if err != nil { - return false, err - } - if node == nil { - return false, fmt.Errorf("unknown node") - } - - return node.Status == structs.NodeStatusReady, fmt.Errorf("bad node status") - }, func(err error) { - t.Fatal(err) - }) - - // Upsert the allocation - state := s.State() - require.Nil(state.UpsertJob(999, a.Job)) - require.Nil(state.UpsertAllocs(1003, []*structs.Allocation{a})) - - // Wait for the client to run the allocation - testutil.WaitForResult(func() (bool, error) { - alloc, err := state.AllocByID(nil, a.ID) - if err != nil { - return false, err - } - if alloc == nil { - return false, fmt.Errorf("unknown alloc") - } - if alloc.ClientStatus != structs.AllocClientStatusComplete { - return false, fmt.Errorf("alloc client status: %v", alloc.ClientStatus) - } - - return true, nil - }, func(err error) { - t.Fatalf("Alloc on node %q not finished: %v", c.NodeID(), err) - }) + // Wait for alloc to be running + alloc := testutil.WaitForRunning(t, s.RPC, job)[0] // Make the request req := &cstructs.FsStreamRequest{ - AllocID: a.ID, - Path: "alloc/logs/web.stdout.0", + AllocID: alloc.ID, + Path: "alloc/logs/worker.stdout.0", PlainText: true, QueryOptions: structs.QueryOptions{Region: "global"}, } @@ -693,72 +646,25 @@ func TestFS_Stream_Follow(t *testing.T) { }) defer c.Shutdown() - // Force an allocation onto the node expectedBase := "Hello from the other side" repeat := 10 - a := mock.Alloc() - a.Job.Type = structs.JobTypeBatch - a.NodeID = c.NodeID() - a.Job.TaskGroups[0].Count = 1 - a.Job.TaskGroups[0].Tasks[0] = &structs.Task{ - Name: "web", - Driver: "mock_driver", - Config: map[string]interface{}{ - "run_for": "20s", - "stdout_string": expectedBase, - "stdout_repeat": repeat, - "stdout_repeat_duration": 200 * time.Millisecond, - }, - LogConfig: structs.DefaultLogConfig(), - Resources: &structs.Resources{ - CPU: 500, - MemoryMB: 256, - }, + job := mock.BatchJob() + job.TaskGroups[0].Count = 1 + job.TaskGroups[0].Tasks[0].Config = map[string]interface{}{ + "run_for": "20s", + "stdout_string": expectedBase, + "stdout_repeat": repeat, + "stdout_repeat_duration": 200 * time.Millisecond, } - // Wait for the client to connect - testutil.WaitForResult(func() (bool, error) { - node, err := s.State().NodeByID(nil, c.NodeID()) - if err != nil { - return false, err - } - if node == nil { - return false, fmt.Errorf("unknown node") - } - - return node.Status == structs.NodeStatusReady, fmt.Errorf("bad node status") - }, func(err error) { - t.Fatal(err) - }) - - // Upsert the allocation - state := s.State() - require.Nil(state.UpsertJob(999, a.Job)) - require.Nil(state.UpsertAllocs(1003, []*structs.Allocation{a})) - - // Wait for the client to run the allocation - testutil.WaitForResult(func() (bool, error) { - alloc, err := state.AllocByID(nil, a.ID) - if err != nil { - return false, err - } - if alloc == nil { - return false, fmt.Errorf("unknown alloc") - } - if alloc.ClientStatus != structs.AllocClientStatusRunning { - return false, fmt.Errorf("alloc client status: %v", alloc.ClientStatus) - } - - return true, nil - }, func(err error) { - t.Fatalf("Alloc on node %q not running: %v", c.NodeID(), err) - }) + // Wait for alloc to be running + alloc := testutil.WaitForRunning(t, s.RPC, job)[0] // Make the request req := &cstructs.FsStreamRequest{ - AllocID: a.ID, - Path: "alloc/logs/web.stdout.0", + AllocID: alloc.ID, + Path: "alloc/logs/worker.stdout.0", PlainText: true, Follow: true, QueryOptions: structs.QueryOptions{Region: "global"}, @@ -837,69 +743,22 @@ func TestFS_Stream_Limit(t *testing.T) { }) defer c.Shutdown() - // Force an allocation onto the node var limit int64 = 5 full := "Hello from the other side" expected := full[:limit] - a := mock.Alloc() - a.Job.Type = structs.JobTypeBatch - a.NodeID = c.NodeID() - a.Job.TaskGroups[0].Count = 1 - a.Job.TaskGroups[0].Tasks[0] = &structs.Task{ - Name: "web", - Driver: "mock_driver", - Config: map[string]interface{}{ - "run_for": "2s", - "stdout_string": full, - }, - LogConfig: structs.DefaultLogConfig(), - Resources: &structs.Resources{ - CPU: 500, - MemoryMB: 256, - }, + job := mock.BatchJob() + job.TaskGroups[0].Count = 1 + job.TaskGroups[0].Tasks[0].Config = map[string]interface{}{ + "run_for": "2s", + "stdout_string": full, } - // Wait for the client to connect - testutil.WaitForResult(func() (bool, error) { - node, err := s.State().NodeByID(nil, c.NodeID()) - if err != nil { - return false, err - } - if node == nil { - return false, fmt.Errorf("unknown node") - } - - return node.Status == structs.NodeStatusReady, fmt.Errorf("bad node status") - }, func(err error) { - t.Fatal(err) - }) - - // Upsert the allocation - state := s.State() - require.Nil(state.UpsertJob(999, a.Job)) - require.Nil(state.UpsertAllocs(1003, []*structs.Allocation{a})) - - // Wait for the client to run the allocation - testutil.WaitForResult(func() (bool, error) { - alloc, err := state.AllocByID(nil, a.ID) - if err != nil { - return false, err - } - if alloc == nil { - return false, fmt.Errorf("unknown alloc") - } - if alloc.ClientStatus != structs.AllocClientStatusComplete { - return false, fmt.Errorf("alloc client status: %v", alloc.ClientStatus) - } - - return true, nil - }, func(err error) { - t.Fatalf("Alloc on node %q not finished: %v", c.NodeID(), err) - }) + // Wait for alloc to be running + alloc := testutil.WaitForRunning(t, s.RPC, job)[0] // Make the request req := &cstructs.FsStreamRequest{ - AllocID: a.ID, + AllocID: alloc.ID, Path: "alloc/logs/web.stdout.0", PlainText: true, Limit: limit, @@ -1040,6 +899,116 @@ OUTER: } } +// TestFS_Logs_TaskPending asserts that trying to stream logs for tasks which +// have not started returns a 404 error. +func TestFS_Logs_TaskPending(t *testing.T) { + t.Parallel() + require := require.New(t) + + // Start a server and client + s := nomad.TestServer(t, nil) + defer s.Shutdown() + testutil.WaitForLeader(t, s.RPC) + + c := TestClient(t, func(c *config.Config) { + c.Servers = []string{s.GetConfig().RPCAddr.String()} + }) + defer c.Shutdown() + + job := mock.BatchJob() + job.TaskGroups[0].Count = 1 + job.TaskGroups[0].Tasks[0].Config = map[string]interface{}{ + "start_block_for": "4s", + } + + // Register job + args := &structs.JobRegisterRequest{} + args.Job = job + args.WriteRequest.Region = "global" + var jobResp structs.JobRegisterResponse + require.NoError(s.RPC("Job.Register", args, &jobResp)) + + // Get the allocation ID + var allocID string + testutil.WaitForResult(func() (bool, error) { + args := structs.AllocListRequest{} + args.Region = "global" + resp := structs.AllocListResponse{} + if err := s.RPC("Alloc.List", &args, &resp); err != nil { + return false, err + } + + if len(resp.Allocations) != 1 { + return false, fmt.Errorf("expected 1 alloc, found %d", len(resp.Allocations)) + } + + allocID = resp.Allocations[0].ID + return true, nil + }, func(err error) { + t.Fatalf("error getting alloc id: %v", err) + }) + + // Make the request + req := &cstructs.FsLogsRequest{ + AllocID: allocID, + Task: job.TaskGroups[0].Tasks[0].Name, + LogType: "stdout", + Origin: "start", + PlainText: true, + QueryOptions: structs.QueryOptions{Region: "global"}, + } + + // Get the handler + handler, err := c.StreamingRpcHandler("FileSystem.Logs") + require.Nil(err) + + // Create a pipe + p1, p2 := net.Pipe() + defer p1.Close() + defer p2.Close() + + errCh := make(chan error) + streamMsg := make(chan *cstructs.StreamErrWrapper) + + // Start the handler + go handler(p2) + + // Start the decoder + go func() { + decoder := codec.NewDecoder(p1, structs.MsgpackHandle) + for { + var msg cstructs.StreamErrWrapper + if err := decoder.Decode(&msg); err != nil { + if err == io.EOF || strings.Contains(err.Error(), "closed") { + return + } + errCh <- fmt.Errorf("error decoding: %v", err) + } + + streamMsg <- &msg + } + }() + + // Send the request + encoder := codec.NewEncoder(p1, structs.MsgpackHandle) + require.Nil(encoder.Encode(req)) + + for { + select { + case <-time.After(3 * time.Second): + t.Fatal("timeout") + case err := <-errCh: + t.Fatalf("unexpected stream error: %v", err) + case msg := <-streamMsg: + require.NotNil(msg.Error) + require.NotNil(msg.Error.Code) + require.EqualValues(404, *msg.Error.Code) + require.Contains(msg.Error.Message, "not started") + return + } + } +} + func TestFS_Logs_ACL(t *testing.T) { t.Parallel() require := require.New(t) @@ -1174,7 +1143,6 @@ func TestFS_Logs(t *testing.T) { }) defer c.Shutdown() - // Force an allocation onto the node expected := "Hello from the other side\n" job := mock.BatchJob() job.TaskGroups[0].Count = 1 @@ -1183,21 +1151,6 @@ func TestFS_Logs(t *testing.T) { "stdout_string": expected, } - // Wait for the client to connect - testutil.WaitForResult(func() (bool, error) { - node, err := s.State().NodeByID(nil, c.NodeID()) - if err != nil { - return false, err - } - if node == nil { - return false, fmt.Errorf("unknown node") - } - - return node.Status == structs.NodeStatusReady, fmt.Errorf("bad node status") - }, func(err error) { - t.Fatal(err) - }) - // Wait for client to be running job testutil.WaitForRunning(t, s.RPC, job) @@ -1291,7 +1244,6 @@ func TestFS_Logs_Follow(t *testing.T) { }) defer c.Shutdown() - // Force an allocation onto the node expectedBase := "Hello from the other side\n" repeat := 10 @@ -1305,19 +1257,11 @@ func TestFS_Logs_Follow(t *testing.T) { } // Wait for client to be running job - testutil.WaitForRunning(t, s.RPC, job) - - // Get the allocation ID - args := structs.AllocListRequest{} - args.Region = "global" - resp := structs.AllocListResponse{} - require.NoError(s.RPC("Alloc.List", &args, &resp)) - require.Len(resp.Allocations, 1) - allocID := resp.Allocations[0].ID + alloc := testutil.WaitForRunning(t, s.RPC, job)[0] // Make the request req := &cstructs.FsLogsRequest{ - AllocID: allocID, + AllocID: alloc.ID, Task: job.TaskGroups[0].Tasks[0].Name, LogType: "stdout", Origin: "start", diff --git a/command/agent/alloc_endpoint_test.go b/command/agent/alloc_endpoint_test.go index 5d85894c2..d576817f1 100644 --- a/command/agent/alloc_endpoint_test.go +++ b/command/agent/alloc_endpoint_test.go @@ -453,7 +453,7 @@ func TestHTTP_AllocSnapshot_Atomic(t *testing.T) { // Wait for the client to run it testutil.WaitForResult(func() (bool, error) { - if _, err := s.client.GetClientAlloc(alloc.ID); err != nil { + if _, err := s.client.GetAllocState(alloc.ID); err != nil { return false, err } diff --git a/testutil/wait.go b/testutil/wait.go index 583f2f1f7..3e0ea166f 100644 --- a/testutil/wait.go +++ b/testutil/wait.go @@ -87,42 +87,40 @@ func WaitForLeader(t testing.T, rpc rpcFn) { }) } -// WaitForRunning runs a job and blocks until it is running. -func WaitForRunning(t testing.T, rpc rpcFn, job *structs.Job) { - registered := false +// WaitForRunning runs a job and blocks until all allocs are out of pending. +func WaitForRunning(t testing.T, rpc rpcFn, job *structs.Job) []*structs.AllocListStub { WaitForResult(func() (bool, error) { - if !registered { - args := &structs.JobRegisterRequest{} - args.Job = job - args.WriteRequest.Region = "global" - var jobResp structs.JobRegisterResponse - err := rpc("Job.Register", args, &jobResp) - if err != nil { - return false, fmt.Errorf("Job.Register error: %v", err) - } + args := &structs.JobRegisterRequest{} + args.Job = job + args.WriteRequest.Region = "global" + var jobResp structs.JobRegisterResponse + err := rpc("Job.Register", args, &jobResp) + return err == nil, fmt.Errorf("Job.Register error: %v", err) + }, func(err error) { + t.Fatalf("error registering job: %v", err) + }) - // Only register once - registered = true - } + t.Logf("Job %q registered", job.ID) - args := &structs.JobSummaryRequest{} + var resp structs.JobAllocationsResponse + + WaitForResult(func() (bool, error) { + args := &structs.JobSpecificRequest{} args.JobID = job.ID args.QueryOptions.Region = "global" - var resp structs.JobSummaryResponse - err := rpc("Job.Summary", args, &resp) + err := rpc("Job.Allocations", args, &resp) if err != nil { - return false, fmt.Errorf("Job.Summary error: %v", err) + return false, fmt.Errorf("Job.Allocations error: %v", err) } - tgs := len(job.TaskGroups) - summaries := len(resp.JobSummary.Summary) - if tgs != summaries { - return false, fmt.Errorf("task_groups=%d summaries=%d", tgs, summaries) + if len(resp.Allocations) == 0 { + return false, fmt.Errorf("0 allocations") } - for tg, summary := range resp.JobSummary.Summary { - if summary.Running == 0 { - return false, fmt.Errorf("task_group=%s %#v", tg, resp.JobSummary.Summary) + for _, alloc := range resp.Allocations { + if alloc.ClientStatus != structs.AllocClientStatusRunning { + return false, fmt.Errorf("alloc not running: id=%v tg=%v status=%v", + alloc.ID, alloc.TaskGroup, alloc.ClientStatus) } } @@ -130,4 +128,6 @@ func WaitForRunning(t testing.T, rpc rpcFn, job *structs.Job) { }, func(err error) { t.Fatalf("job not running: %v", err) }) + + return resp.Allocations } From 9e49ed3464cd6ead0f53181b998956d5ea13514e Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 27 Sep 2018 17:30:10 -0700 Subject: [PATCH 103/152] ar: AllocState should not mutate ar.state If ar.state.TaskStates has not been set, set it on the copy of ar.state. That keeps ar.state manipulations in one location and allows AllocState to only acquire read-locks. --- client/allocrunnerv2/alloc_runner.go | 15 ++++++++------- 1 file changed, 8 insertions(+), 7 deletions(-) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 8b1ace5f4..19ff1b32b 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -56,7 +56,8 @@ type allocRunner struct { alloc *structs.Allocation allocLock sync.RWMutex - state *state.State // alloc runner state + // state is the alloc runner's state + state *state.State stateLock sync.RWMutex stateDB cstate.StateDB @@ -391,22 +392,22 @@ func getClientStatus(taskStates map[string]*structs.TaskState) (status, descript // AllocState returns a copy of allocation state including a snapshot of task // states. func (ar *allocRunner) AllocState() *state.State { - // Must acquire write-lock in case TaskStates needs to be set. - ar.stateLock.Lock() - defer ar.stateLock.Unlock() + ar.stateLock.RLock() + state := ar.state.Copy() + ar.stateLock.RUnlock() // If TaskStateUpdated has not been called yet, ar.state.TaskStates // won't be set as it is not the canonical source of TaskStates. - if len(ar.state.TaskStates) == 0 { + if len(state.TaskStates) == 0 { ar.tasksLock.RLock() ar.state.TaskStates = make(map[string]*structs.TaskState, len(ar.tasks)) for k, tr := range ar.tasks { - ar.state.TaskStates[k] = tr.TaskState() + state.TaskStates[k] = tr.TaskState() } ar.tasksLock.RUnlock() } - return ar.state.Copy() + return state } // Update the running allocation with a new version received from the server. From 048510b13ed354c8d326389308dc0a9b515245cc Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Thu, 27 Sep 2018 17:32:17 -0700 Subject: [PATCH 104/152] tr: properly comment handle fields --- client/allocrunnerv2/taskrunner/task_runner.go | 11 ++++++++--- 1 file changed, 8 insertions(+), 3 deletions(-) diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index c206860e2..cf541e720 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -96,9 +96,14 @@ type TaskRunner struct { // driver is the driver for the task. driver driver.Driver - handle driver.DriverHandle // the handle to the running driver - handleResult *handleResult // proxy for handle results - handleLock sync.Mutex + // handleLock guards access to handle and handleResult + handleLock sync.Mutex + + // handle to the running driver + handle driver.DriverHandle + + // handleResult proxies wait results from drivers + handleResult *handleResult // task is the task being run task *structs.Task From 5f696608a65c877a91040ab43cb92f4c124b2e5d Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Fri, 28 Sep 2018 16:54:41 -0700 Subject: [PATCH 105/152] tests: fix missing logger caused by bad merge --- client/testing.go | 3 +++ 1 file changed, 3 insertions(+) diff --git a/client/testing.go b/client/testing.go index deb19eda2..bfeef18e4 100644 --- a/client/testing.go +++ b/client/testing.go @@ -5,6 +5,7 @@ import ( consulApi "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/fingerprint" "github.com/hashicorp/nomad/command/agent/consul" + "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/plugins/shared/catalog" "github.com/hashicorp/nomad/plugins/shared/singleton" "github.com/mitchellh/go-testing-interface" @@ -21,6 +22,8 @@ func TestClient(t testing.T, cb func(c *config.Config)) *Client { } conf.Options[fingerprint.TightenNetworkTimeoutsConfig] = "true" + logger := testlog.HCLogger(t) + // Set the plugin loaders conf.PluginLoader = catalog.TestPluginLoader(t) conf.PluginSingletonLoader = singleton.NewSingletonLoader(logger, conf.PluginLoader) From 5742a6b932ae1cccbe24638ad8578cd753ba975a Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Sat, 29 Sep 2018 23:03:49 -0400 Subject: [PATCH 106/152] plugin/drivers: rework eventer and change naming stream -> consumer --- plugins/drivers/utils/eventer.go | 99 ++++++++++++++++++--------- plugins/drivers/utils/eventer_test.go | 52 +++++++++----- 2 files changed, 103 insertions(+), 48 deletions(-) diff --git a/plugins/drivers/utils/eventer.go b/plugins/drivers/utils/eventer.go index da6d0c500..8cea3209c 100644 --- a/plugins/drivers/utils/eventer.go +++ b/plugins/drivers/utils/eventer.go @@ -1,41 +1,49 @@ package utils import ( + "fmt" "sync" "time" + hclog "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/plugins/drivers" "golang.org/x/net/context" ) var ( - //DefaultSendEventTimeout is the timeout used when publishing events to consumers + // DefaultSendEventTimeout is the timeout used when publishing events to consumers DefaultSendEventTimeout = 2 * time.Second ) // Eventer is a utility to control broadcast of TaskEvents to multiple consumers. -// It also implements the TaskStats func in the DriverPlugin interface so that +// It also implements the TaskEvents func in the DriverPlugin interface so that // it can be embedded in a implementing driver struct. type Eventer struct { - sync.RWMutex + consumersLock sync.RWMutex // events is a channel were events to be broadcasted are sent events chan *drivers.TaskEvent - // streamers is a slice of consumers to broadcast events to - // access is gaurded by RWMutex - streamers []*eventStreamer + // consumers is a slice of eventConsumers to broadcast events to. + // access is gaurded by consumersLock RWMutex + consumers []*eventConsumer - // stop chan to allow control of event loop shutdown - stop chan struct{} + // ctx to allow control of event loop shutdown + ctx context.Context + + // done tracks if the event loop has stopped due to the ctx being done + done bool + + logger hclog.Logger } // NewEventer returns an Eventer with a running event loop that can be stopped // by closing the given stop channel -func NewEventer(stop chan struct{}) *Eventer { +func NewEventer(ctx context.Context, logger hclog.Logger) *Eventer { e := &Eventer{ events: make(chan *drivers.TaskEvent), - stop: stop, + ctx: ctx, + logger: logger, } go e.eventLoop() return e @@ -46,56 +54,83 @@ func NewEventer(stop chan struct{}) *Eventer { func (e *Eventer) eventLoop() { for { select { - case <-e.stop: - for _, stream := range e.streamers { - close(stream.ch) - } + case <-e.ctx.Done(): + e.done = true + close(e.events) return case event := <-e.events: - e.RLock() - for _, stream := range e.streamers { - stream.send(event) + e.consumersLock.RLock() + for _, consumer := range e.consumers { + consumer.send(event) } - e.RUnlock() + e.consumersLock.RUnlock() } } } -type eventStreamer struct { +type eventConsumer struct { timeout time.Duration ctx context.Context ch chan *drivers.TaskEvent + logger hclog.Logger } -func (s *eventStreamer) send(event *drivers.TaskEvent) { +func (c *eventConsumer) send(event *drivers.TaskEvent) { select { - case <-time.After(s.timeout): - case <-s.ctx.Done(): - case s.ch <- event: + case <-time.After(c.timeout): + c.logger.Warn("timeout sending event", "task_id", event.TaskID, "message", event.Message) + case <-c.ctx.Done(): + case c.ch <- event: } } -func (e *Eventer) newStream(ctx context.Context) <-chan *drivers.TaskEvent { - e.Lock() - defer e.Unlock() +func (e *Eventer) newConsumer(ctx context.Context) *eventConsumer { + e.consumersLock.Lock() + defer e.consumersLock.Unlock() - stream := &eventStreamer{ + consumer := &eventConsumer{ ch: make(chan *drivers.TaskEvent), ctx: ctx, timeout: DefaultSendEventTimeout, + logger: e.logger, } - e.streamers = append(e.streamers, stream) + e.consumers = append(e.consumers, consumer) - return stream.ch + return consumer } // TaskEvents is an implementation of the DriverPlugin.TaskEvents function func (e *Eventer) TaskEvents(ctx context.Context) (<-chan *drivers.TaskEvent, error) { - stream := e.newStream(ctx) - return stream, nil + consumer := e.newConsumer(ctx) + go e.handleConsumer(consumer) + return consumer.ch, nil +} + +func (e *Eventer) handleConsumer(consumer *eventConsumer) { + // wait for consumer or eventer ctx to finish + select { + case <-consumer.ctx.Done(): + case <-e.ctx.Done(): + } + e.consumersLock.Lock() + defer e.consumersLock.Unlock() + defer close(consumer.ch) + + filtered := e.consumers[:0] + for _, c := range e.consumers { + if c != consumer { + filtered = append(filtered, c) + } + } + e.consumers = filtered } // EmitEvent can be used to broadcast a new event -func (e *Eventer) EmitEvent(event *drivers.TaskEvent) { +func (e *Eventer) EmitEvent(event *drivers.TaskEvent) error { + if e.done { + return fmt.Errorf("error sending event, context canceled") + } + e.events <- event + return nil } diff --git a/plugins/drivers/utils/eventer_test.go b/plugins/drivers/utils/eventer_test.go index ad14dd283..6c7972a9a 100644 --- a/plugins/drivers/utils/eventer_test.go +++ b/plugins/drivers/utils/eventer_test.go @@ -6,6 +6,7 @@ import ( "testing" "time" + "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/plugins/drivers" "github.com/stretchr/testify/require" ) @@ -14,8 +15,8 @@ func TestEventer(t *testing.T) { t.Parallel() require := require.New(t) - stop := make(chan struct{}) - e := NewEventer(stop) + ctx, cancel := context.WithCancel(context.Background()) + e := NewEventer(ctx, testlog.HCLogger(t)) events := []*drivers.TaskEvent{ { @@ -32,9 +33,11 @@ func TestEventer(t *testing.T) { }, } - consumer1, err := e.TaskEvents(context.Background()) + ctx1, cancel1 := context.WithCancel(context.Background()) + consumer1, err := e.TaskEvents(ctx1) require.NoError(err) - consumer2, err := e.TaskEvents(context.Background()) + ctx2 := (context.Background()) + consumer2, err := e.TaskEvents(ctx2) require.NoError(err) var buffer1, buffer2 []*drivers.TaskEvent @@ -42,31 +45,48 @@ func TestEventer(t *testing.T) { wg.Add(2) go func() { defer wg.Done() - for { - event, ok := <-consumer1 - if !ok { - return - } + var i int + for event := range consumer1 { + i++ buffer1 = append(buffer1, event) + if i == 3 { + break + } } }() go func() { defer wg.Done() - for { - event, ok := <-consumer2 - if !ok { - return - } + var i int + for event := range consumer2 { + i++ buffer2 = append(buffer2, event) + if i == 3 { + break + } } }() for _, event := range events { - e.EmitEvent(event) + require.NoError(e.EmitEvent(event)) } - close(stop) wg.Wait() require.Exactly(events, buffer1) require.Exactly(events, buffer2) + cancel1() + time.Sleep(100 * time.Millisecond) + require.Equal(1, len(e.consumers)) + + require.NoError(e.EmitEvent(&drivers.TaskEvent{})) + ev, ok := <-consumer1 + require.Nil(ev) + require.False(ok) + ev, ok = <-consumer2 + require.NotNil(ev) + require.True(ok) + + cancel() + time.Sleep(100 * time.Millisecond) + require.Zero(len(e.consumers)) + require.Error(e.EmitEvent(&drivers.TaskEvent{})) } From 8cf32eb9d232d09bc989c75a68ae2f6068a3b3e6 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Sat, 29 Sep 2018 23:05:14 -0400 Subject: [PATCH 107/152] drivers/rawexec: PR comments and feedback --- drivers/{raw_exec => rawexec}/driver.go | 210 +++++++----------- drivers/{raw_exec => rawexec}/driver_test.go | 2 +- .../{raw_exec => rawexec}/driver_unix_test.go | 2 +- drivers/rawexec/state.go | 86 +++++++ plugins/drivers/task_handle.go | 5 +- .../drivers/utils/plugin_reattach_config.go | 2 +- 6 files changed, 172 insertions(+), 135 deletions(-) rename drivers/{raw_exec => rawexec}/driver.go (73%) rename drivers/{raw_exec => rawexec}/driver_test.go (99%) rename drivers/{raw_exec => rawexec}/driver_unix_test.go (99%) create mode 100644 drivers/rawexec/state.go diff --git a/drivers/raw_exec/driver.go b/drivers/rawexec/driver.go similarity index 73% rename from drivers/raw_exec/driver.go rename to drivers/rawexec/driver.go index a41739865..cf1aa6b94 100644 --- a/drivers/raw_exec/driver.go +++ b/drivers/rawexec/driver.go @@ -1,11 +1,10 @@ -package raw_exec +package rawexec import ( "fmt" "os" "path/filepath" "strconv" - "sync" "time" "github.com/hashicorp/consul-template/signals" @@ -29,6 +28,7 @@ const ( ) var ( + // pluginInfo is the response returned for the PluginInfo RPC pluginInfo = &base.PluginInfoResponse{ Type: base.PluginTypeDriver, PluginApiVersion: "0.0.1", @@ -36,6 +36,7 @@ var ( Name: pluginName, } + // configSpec is the hcl specification returned by the ConfigSchema RPC configSpec = hclspec.NewObject(map[string]*hclspec.Spec{ "enabled": hclspec.NewDefault( hclspec.NewAttr("enabled", "bool", false), @@ -47,11 +48,15 @@ var ( ), }) + // taskConfigSpec is the hcl specification for the driver config section of + // a task within a job. It is returned in the TaskConfigSchema RPC taskConfigSpec = hclspec.NewObject(map[string]*hclspec.Spec{ "command": hclspec.NewAttr("command", "string", true), "args": hclspec.NewAttr("command", "list(string)", false), }) + // capabilities is returned by the Capabilities RPC and indicates what + // optional features this driver supports capabilities = &drivers.Capabilities{ SendSignals: true, Exec: true, @@ -59,23 +64,34 @@ var ( } ) -// The RawExecDriver is a privileged version of the exec driver. It provides no +// RawExecDriver is a privileged version of the exec driver. It provides no // resource isolation and just fork/execs. The Exec driver should be preferred // and this should only be used when explicitly needed. type RawExecDriver struct { - *utils.Eventer + // eventer is used to handle multiplexing of TaskEvents calls such that an + // event can be broadcast to all callers + eventer *utils.Eventer + + // config is the driver configuration set by the SetConfig RPC config *Config - tasks *taskStore - // fingerprintCh is a channel which other funcs can send fingerprints to - // that will immediately be sent - fingerprintCh chan *drivers.Fingerprint + // tasks is the in memory datastore mapping taskIDs to rawExecDriverHandles + tasks *taskStore - stopCh chan struct{} + // ctx is the context for the driver. It is passed to other subsystems to + // coordinate shutdown + ctx context.Context + // signalShutdown is called when the driver is shutting down and cancels the + // ctx passed to any subsystems + signalShutdown context.CancelFunc + + // logger will log to the plugin output which is usually an 'executor.out' + // file located in the root of the TaskDir logger hclog.Logger } +// Config is the driver configuration set by the SetConfig RPC call type Config struct { // NoCgroups tracks whether we should use a cgroup to manage the process // tree @@ -85,20 +101,33 @@ type Config struct { Enabled bool `codec:"enabled"` } +// TaskConfig is the driver configuration of a task within a job type TaskConfig struct { Command string `codec:"command"` Args []string `codec:"args"` } +// RawExecTaskState is the state which is encoded in the handle returned in +// StartTask. This information is needed to rebuild the task state and handler +// during recovery. +type RawExecTaskState struct { + ReattachConfig *utils.ReattachConfig + TaskConfig *drivers.TaskConfig + Pid int + StartedAt time.Time +} + +// NewRawExecDriver returns a new DriverPlugin implementation func NewRawExecDriver(logger hclog.Logger) drivers.DriverPlugin { - stopCh := make(chan struct{}) + ctx, cancel := context.WithCancel(context.Background()) + logger = logger.Named(pluginName) return &RawExecDriver{ - Eventer: utils.NewEventer(stopCh), - config: &Config{}, - tasks: newTaskStore(), - fingerprintCh: make(chan *drivers.Fingerprint), - stopCh: stopCh, - logger: logger.Named(pluginName), + eventer: utils.NewEventer(ctx, logger), + config: &Config{}, + tasks: newTaskStore(), + ctx: ctx, + signalShutdown: cancel, + logger: logger, } } @@ -117,12 +146,11 @@ func (r *RawExecDriver) SetConfig(data []byte) error { } r.config = &config - go r.fingerprintNow() return nil } func (r *RawExecDriver) Shutdown(ctx context.Context) error { - close(r.stopCh) + r.signalShutdown() return nil } @@ -136,32 +164,26 @@ func (r *RawExecDriver) Capabilities() (*drivers.Capabilities, error) { func (r *RawExecDriver) Fingerprint(ctx context.Context) (<-chan *drivers.Fingerprint, error) { ch := make(chan *drivers.Fingerprint) - go r.fingerprint(ctx, ch) + go r.handleFingerprint(ctx, ch) return ch, nil } -func (r *RawExecDriver) fingerprint(ctx context.Context, ch chan *drivers.Fingerprint) { - defer close(r.fingerprintCh) - +func (r *RawExecDriver) handleFingerprint(ctx context.Context, ch chan *drivers.Fingerprint) { ticker := time.NewTimer(0) for { select { case <-ctx.Done(): return + case <-r.ctx.Done(): + return case <-ticker.C: ticker.Reset(fingerprintPeriod) - go r.fingerprintNow() - case f := <-r.fingerprintCh: - ch <- f + ch <- r.buildFingerprint() } } } -func (r *RawExecDriver) fingerprintNow() { - if r.fingerprintCh == nil { - r.logger.Debug("fingerprint channel was nil, skipping fingerprint") - return - } +func (r *RawExecDriver) buildFingerprint() *drivers.Fingerprint { var health drivers.HealthState var desc string attrs := map[string]string{} @@ -173,7 +195,8 @@ func (r *RawExecDriver) fingerprintNow() { health = drivers.HealthStateUndetected desc = "raw_exec disabled" } - r.fingerprintCh <- &drivers.Fingerprint{ + + return &drivers.Fingerprint{ Attributes: map[string]string{}, Health: health, HealthDescription: desc, @@ -181,18 +204,20 @@ func (r *RawExecDriver) fingerprintNow() { } func (r *RawExecDriver) RecoverTask(handle *drivers.TaskHandle) error { - var taskState RawExecTaskState + if handle == nil { + return fmt.Errorf("error: handle cannot be nil") + } - err := handle.GetDriverState(&taskState) - if err != nil { - r.logger.Error("failed to recover task", "error", err, "task_id", handle.Config.ID) - return err + var taskState RawExecTaskState + if err := handle.GetDriverState(&taskState); err != nil { + r.logger.Error("failed to decode task state from handle", "error", err, "task_id", handle.Config.ID) + return fmt.Errorf("failed to decode task state from handle: %v", err) } plugRC, err := utils.ReattachConfigToGoPlugin(taskState.ReattachConfig) if err != nil { - r.logger.Error("failed to recover task", "error", err, "task_id", handle.Config.ID) - return err + r.logger.Error("failed to build ReattachConfig from task state", "error", err, "task_id", handle.Config.ID) + return fmt.Errorf("failed to build ReattachConfig from task state: %v", err) } pluginConfig := &plugin.ClientConfig{ @@ -201,8 +226,8 @@ func (r *RawExecDriver) RecoverTask(handle *drivers.TaskHandle) error { exec, pluginClient, err := utils.CreateExecutorWithConfig(pluginConfig, os.Stderr) if err != nil { - r.logger.Error("failed to recover task", "error", err, "task_id", handle.Config.ID) - return err + r.logger.Error("failed to reattach to executor", "error", err, "task_id", handle.Config.ID) + return fmt.Errorf("failed to reattach to executor: %v", err) } h := &rawExecTaskHandle{ @@ -223,12 +248,12 @@ func (r *RawExecDriver) RecoverTask(handle *drivers.TaskHandle) error { func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) { if _, ok := r.tasks.Get(cfg.ID); ok { - return nil, fmt.Errorf("task with ID '%s' already started", cfg.ID) + return nil, fmt.Errorf("task with ID %q already started", cfg.ID) } var driverConfig TaskConfig if err := cfg.DecodeDriverConfig(&driverConfig); err != nil { - return nil, err + return nil, fmt.Errorf("failed to decode driver config: %v", err) } handle := drivers.NewTaskHandle(pluginName) @@ -240,9 +265,10 @@ func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, LogLevel: "debug", } + // TODO: best way to pass port ranges in from client config exec, pluginClient, err := utils.CreateExecutor(os.Stderr, hclog.Debug, 14000, 14512, executorConfig) if err != nil { - return nil, err + return nil, fmt.Errorf("failed to create executor: %v", err) } execCmd := &executor.ExecCommand{ @@ -259,7 +285,7 @@ func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, ps, err := exec.Launch(execCmd) if err != nil { pluginClient.Kill() - return nil, err + return nil, fmt.Errorf("failed to launch command with executor: %v", err) } h := &rawExecTaskHandle{ @@ -272,8 +298,6 @@ func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, logger: r.logger, } - r.tasks.Set(cfg.ID, h) - driverState := RawExecTaskState{ ReattachConfig: utils.ReattachConfigFromGoPlugin(pluginClient.ReattachConfig()), Pid: ps.Pid, @@ -285,19 +309,21 @@ func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, r.logger.Error("failed to start task, error setting driver state", "error", err) exec.Shutdown("", 0) pluginClient.Kill() - return nil, err + return nil, fmt.Errorf("failed to set driver state: %v", err) } + r.tasks.Set(cfg.ID, h) go h.run() return handle, nil } func (r *RawExecDriver) WaitTask(ctx context.Context, taskID string) (<-chan *drivers.ExitResult, error) { - ch := make(chan *drivers.ExitResult) handle, ok := r.tasks.Get(taskID) if !ok { return nil, drivers.ErrTaskNotFound } + + ch := make(chan *drivers.ExitResult) go r.handleWait(ctx, handle, ch) return ch, nil @@ -321,6 +347,8 @@ func (r *RawExecDriver) handleWait(ctx context.Context, handle *rawExecTaskHandl select { case <-ctx.Done(): return + case <-r.ctx.Done(): + return case ch <- result: } } @@ -384,7 +412,6 @@ func (r *RawExecDriver) InspectTask(taskID string) (*drivers.TaskStatus, error) DriverAttributes: map[string]string{ "pid": strconv.Itoa(handle.pid), }, - NetworkOverride: &drivers.NetworkOverride{}, } return status, nil @@ -409,6 +436,10 @@ func (r *RawExecDriver) TaskStats(taskID string) (*drivers.TaskStats, error) { }, nil } +func (r *RawExecDriver) TaskEvents(ctx context.Context) (<-chan *drivers.TaskEvent, error) { + return r.eventer.TaskEvents(ctx) +} + func (r *RawExecDriver) SignalTask(taskID string, signal string) error { handle, ok := r.tasks.Get(taskID) if !ok { @@ -417,6 +448,7 @@ func (r *RawExecDriver) SignalTask(taskID string, signal string) error { sig := os.Interrupt if s, ok := signals.SignalLookup[signal]; ok { + r.logger.Warn("signal to send to task unknown, using SIGINT", "signal", signal) sig = s } return handle.exec.Signal(sig) @@ -448,83 +480,3 @@ func (r *RawExecDriver) ExecTask(taskID string, cmd []string, timeout time.Durat }, }, nil } - -type taskStore struct { - store map[string]*rawExecTaskHandle - lock sync.RWMutex -} - -func newTaskStore() *taskStore { - return &taskStore{store: map[string]*rawExecTaskHandle{}} -} - -func (ts *taskStore) Set(id string, handle *rawExecTaskHandle) { - ts.lock.Lock() - defer ts.lock.Unlock() - ts.store[id] = handle -} - -func (ts *taskStore) Get(id string) (*rawExecTaskHandle, bool) { - ts.lock.RLock() - defer ts.lock.RUnlock() - t, ok := ts.store[id] - return t, ok -} - -func (ts *taskStore) Delete(id string) { - ts.lock.Lock() - defer ts.lock.Unlock() - delete(ts.store, id) -} - -type RawExecTaskState struct { - ReattachConfig *utils.ReattachConfig - TaskConfig *drivers.TaskConfig - Pid int - StartedAt time.Time -} - -type rawExecTaskHandle struct { - exec executor.Executor - pid int - pluginClient *plugin.Client - logger hclog.Logger - - // stateLock syncs access to all fields below - stateLock sync.RWMutex - - task *drivers.TaskConfig - procState drivers.TaskState - startedAt time.Time - completedAt time.Time - exitResult *drivers.ExitResult -} - -func (h *rawExecTaskHandle) IsRunning() bool { - return h.procState == drivers.TaskStateRunning -} - -func (h *rawExecTaskHandle) run() { - - // since run is called immediatly after the handle is created this - // ensures the exitResult is initialized so we avoid a nil pointer - // thus it does not need to be included in the lock - if h.exitResult == nil { - h.exitResult = &drivers.ExitResult{} - } - - ps, err := h.exec.Wait() - h.stateLock.Lock() - defer h.stateLock.Unlock() - - if err != nil { - h.exitResult.Err = err - h.procState = drivers.TaskStateUnknown - h.completedAt = time.Now() - return - } - h.procState = drivers.TaskStateExited - h.exitResult.ExitCode = ps.ExitCode - h.exitResult.Signal = ps.Signal - h.completedAt = ps.Time -} diff --git a/drivers/raw_exec/driver_test.go b/drivers/rawexec/driver_test.go similarity index 99% rename from drivers/raw_exec/driver_test.go rename to drivers/rawexec/driver_test.go index 0ca3caa17..1c8ed6dbf 100644 --- a/drivers/raw_exec/driver_test.go +++ b/drivers/rawexec/driver_test.go @@ -1,4 +1,4 @@ -package raw_exec +package rawexec import ( "context" diff --git a/drivers/raw_exec/driver_unix_test.go b/drivers/rawexec/driver_unix_test.go similarity index 99% rename from drivers/raw_exec/driver_unix_test.go rename to drivers/rawexec/driver_unix_test.go index e20518782..7558df44f 100644 --- a/drivers/raw_exec/driver_unix_test.go +++ b/drivers/rawexec/driver_unix_test.go @@ -1,6 +1,6 @@ // +build !windows -package raw_exec +package rawexec import ( "context" diff --git a/drivers/rawexec/state.go b/drivers/rawexec/state.go new file mode 100644 index 000000000..ecaefb87b --- /dev/null +++ b/drivers/rawexec/state.go @@ -0,0 +1,86 @@ +package rawexec + +import ( + "sync" + "time" + + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/client/driver/executor" + "github.com/hashicorp/nomad/plugins/drivers" +) + +type taskStore struct { + store map[string]*rawExecTaskHandle + lock sync.RWMutex +} + +func newTaskStore() *taskStore { + return &taskStore{store: map[string]*rawExecTaskHandle{}} +} + +func (ts *taskStore) Set(id string, handle *rawExecTaskHandle) { + ts.lock.Lock() + defer ts.lock.Unlock() + ts.store[id] = handle +} + +func (ts *taskStore) Get(id string) (*rawExecTaskHandle, bool) { + ts.lock.RLock() + defer ts.lock.RUnlock() + t, ok := ts.store[id] + return t, ok +} + +func (ts *taskStore) Delete(id string) { + ts.lock.Lock() + defer ts.lock.Unlock() + delete(ts.store, id) +} + +type rawExecTaskHandle struct { + exec executor.Executor + pid int + pluginClient *plugin.Client + logger hclog.Logger + + // stateLock syncs access to all fields below + stateLock sync.RWMutex + + task *drivers.TaskConfig + procState drivers.TaskState + startedAt time.Time + completedAt time.Time + exitResult *drivers.ExitResult +} + +func (h *rawExecTaskHandle) IsRunning() bool { + return h.procState == drivers.TaskStateRunning +} + +func (h *rawExecTaskHandle) run() { + + // since run is called immediatly after the handle is created this + // ensures the exitResult is initialized so we avoid a nil pointer + // thus it does not need to be included in the lock + if h.exitResult == nil { + h.exitResult = &drivers.ExitResult{} + } + + ps, err := h.exec.Wait() + h.stateLock.Lock() + defer h.stateLock.Unlock() + + if err != nil { + h.exitResult.Err = err + h.procState = drivers.TaskStateUnknown + h.completedAt = time.Now() + return + } + h.procState = drivers.TaskStateExited + h.exitResult.ExitCode = ps.ExitCode + h.exitResult.Signal = ps.Signal + h.completedAt = ps.Time + + // TODO: detect if the task OOMed +} diff --git a/plugins/drivers/task_handle.go b/plugins/drivers/task_handle.go index d6a7981a1..745ddea07 100644 --- a/plugins/drivers/task_handle.go +++ b/plugins/drivers/task_handle.go @@ -2,7 +2,6 @@ package drivers import ( "github.com/hashicorp/nomad/plugins/base" - "github.com/ugorji/go/codec" ) // TaskHandle is the state shared between a driver and the client. @@ -21,10 +20,10 @@ func NewTaskHandle(driver string) *TaskHandle { func (h *TaskHandle) SetDriverState(v interface{}) error { h.driverState = []byte{} - return codec.NewEncoderBytes(&h.driverState, base.MsgpackHandle).Encode(v) + return base.MsgPackEncode(&h.driverState, v) } func (h *TaskHandle) GetDriverState(v interface{}) error { - return codec.NewDecoderBytes(h.driverState, base.MsgpackHandle).Decode(v) + return base.MsgPackDecode(h.driverState, v) } diff --git a/plugins/drivers/utils/plugin_reattach_config.go b/plugins/drivers/utils/plugin_reattach_config.go index 3d5b80d33..02008279f 100644 --- a/plugins/drivers/utils/plugin_reattach_config.go +++ b/plugins/drivers/utils/plugin_reattach_config.go @@ -17,7 +17,7 @@ type ReattachConfig struct { } // ReattachConfigToGoPlugin converts a ReattachConfig wrapper struct into a go -// plugin ReattacConfig struct +// plugin ReattachConfig struct func ReattachConfigToGoPlugin(rc *ReattachConfig) (*plugin.ReattachConfig, error) { plug := &plugin.ReattachConfig{ Protocol: plugin.Protocol(rc.Protocol), From 8d04ae860da6a5239b5eecacef97da2b4571170f Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Sat, 29 Sep 2018 23:12:22 -0400 Subject: [PATCH 108/152] plugins/drivers: remove bool to track if eventLoop shutdown and use context instead --- plugins/drivers/utils/eventer.go | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/plugins/drivers/utils/eventer.go b/plugins/drivers/utils/eventer.go index 8cea3209c..45a6aa5fe 100644 --- a/plugins/drivers/utils/eventer.go +++ b/plugins/drivers/utils/eventer.go @@ -31,9 +31,6 @@ type Eventer struct { // ctx to allow control of event loop shutdown ctx context.Context - // done tracks if the event loop has stopped due to the ctx being done - done bool - logger hclog.Logger } @@ -55,7 +52,6 @@ func (e *Eventer) eventLoop() { for { select { case <-e.ctx.Done(): - e.done = true close(e.events) return case event := <-e.events: @@ -127,10 +123,14 @@ func (e *Eventer) handleConsumer(consumer *eventConsumer) { // EmitEvent can be used to broadcast a new event func (e *Eventer) EmitEvent(event *drivers.TaskEvent) error { - if e.done { - return fmt.Errorf("error sending event, context canceled") - } - e.events <- event + select { + case <-e.ctx.Done(): + return fmt.Errorf("error sending event, context canceled") + case e.events <- event: + if e.logger.IsTrace() { + e.logger.Trace("emitting event", "event", *event) + } + } return nil } From 969ecea6f04c22158bb2ac7ed297af78adc1d1ca Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Tue, 2 Oct 2018 15:36:55 -0400 Subject: [PATCH 109/152] drivers/utils: better handling of consumer cleanup in eventer --- drivers/rawexec/driver.go | 2 +- drivers/rawexec/handle.go | 58 +++++++++++++++++ drivers/rawexec/state.go | 53 --------------- plugins/drivers/utils/eventer.go | 93 +++++++++++++++------------ plugins/drivers/utils/eventer_test.go | 63 ++++++++++++------ 5 files changed, 154 insertions(+), 115 deletions(-) create mode 100644 drivers/rawexec/handle.go diff --git a/drivers/rawexec/driver.go b/drivers/rawexec/driver.go index cf1aa6b94..5d9f09096 100644 --- a/drivers/rawexec/driver.go +++ b/drivers/rawexec/driver.go @@ -448,7 +448,7 @@ func (r *RawExecDriver) SignalTask(taskID string, signal string) error { sig := os.Interrupt if s, ok := signals.SignalLookup[signal]; ok { - r.logger.Warn("signal to send to task unknown, using SIGINT", "signal", signal) + r.logger.Warn("signal to send to task unknown, using SIGINT", "signal", signal, "task_id", handle.task.ID) sig = s } return handle.exec.Signal(sig) diff --git a/drivers/rawexec/handle.go b/drivers/rawexec/handle.go new file mode 100644 index 000000000..09fe04e81 --- /dev/null +++ b/drivers/rawexec/handle.go @@ -0,0 +1,58 @@ +package rawexec + +import ( + "sync" + "time" + + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/client/driver/executor" + "github.com/hashicorp/nomad/plugins/drivers" +) + +type rawExecTaskHandle struct { + exec executor.Executor + pid int + pluginClient *plugin.Client + logger hclog.Logger + + // stateLock syncs access to all fields below + stateLock sync.RWMutex + + task *drivers.TaskConfig + procState drivers.TaskState + startedAt time.Time + completedAt time.Time + exitResult *drivers.ExitResult +} + +func (h *rawExecTaskHandle) IsRunning() bool { + return h.procState == drivers.TaskStateRunning +} + +func (h *rawExecTaskHandle) run() { + + // since run is called immediatly after the handle is created this + // ensures the exitResult is initialized so we avoid a nil pointer + // thus it does not need to be included in the lock + if h.exitResult == nil { + h.exitResult = &drivers.ExitResult{} + } + + ps, err := h.exec.Wait() + h.stateLock.Lock() + defer h.stateLock.Unlock() + + if err != nil { + h.exitResult.Err = err + h.procState = drivers.TaskStateUnknown + h.completedAt = time.Now() + return + } + h.procState = drivers.TaskStateExited + h.exitResult.ExitCode = ps.ExitCode + h.exitResult.Signal = ps.Signal + h.completedAt = ps.Time + + // TODO: detect if the task OOMed +} diff --git a/drivers/rawexec/state.go b/drivers/rawexec/state.go index ecaefb87b..3a8e28e79 100644 --- a/drivers/rawexec/state.go +++ b/drivers/rawexec/state.go @@ -2,12 +2,6 @@ package rawexec import ( "sync" - "time" - - hclog "github.com/hashicorp/go-hclog" - plugin "github.com/hashicorp/go-plugin" - "github.com/hashicorp/nomad/client/driver/executor" - "github.com/hashicorp/nomad/plugins/drivers" ) type taskStore struct { @@ -37,50 +31,3 @@ func (ts *taskStore) Delete(id string) { defer ts.lock.Unlock() delete(ts.store, id) } - -type rawExecTaskHandle struct { - exec executor.Executor - pid int - pluginClient *plugin.Client - logger hclog.Logger - - // stateLock syncs access to all fields below - stateLock sync.RWMutex - - task *drivers.TaskConfig - procState drivers.TaskState - startedAt time.Time - completedAt time.Time - exitResult *drivers.ExitResult -} - -func (h *rawExecTaskHandle) IsRunning() bool { - return h.procState == drivers.TaskStateRunning -} - -func (h *rawExecTaskHandle) run() { - - // since run is called immediatly after the handle is created this - // ensures the exitResult is initialized so we avoid a nil pointer - // thus it does not need to be included in the lock - if h.exitResult == nil { - h.exitResult = &drivers.ExitResult{} - } - - ps, err := h.exec.Wait() - h.stateLock.Lock() - defer h.stateLock.Unlock() - - if err != nil { - h.exitResult.Err = err - h.procState = drivers.TaskStateUnknown - h.completedAt = time.Now() - return - } - h.procState = drivers.TaskStateExited - h.exitResult.ExitCode = ps.ExitCode - h.exitResult.Signal = ps.Signal - h.completedAt = ps.Time - - // TODO: detect if the task OOMed -} diff --git a/plugins/drivers/utils/eventer.go b/plugins/drivers/utils/eventer.go index 45a6aa5fe..737b7f608 100644 --- a/plugins/drivers/utils/eventer.go +++ b/plugins/drivers/utils/eventer.go @@ -1,7 +1,6 @@ package utils import ( - "fmt" "sync" "time" @@ -13,20 +12,27 @@ import ( var ( // DefaultSendEventTimeout is the timeout used when publishing events to consumers DefaultSendEventTimeout = 2 * time.Second + + // ConsumerGCInterval is the interval at which garbage collection of consumers + // occures + ConsumerGCInterval = time.Minute ) // Eventer is a utility to control broadcast of TaskEvents to multiple consumers. // It also implements the TaskEvents func in the DriverPlugin interface so that // it can be embedded in a implementing driver struct. type Eventer struct { - consumersLock sync.RWMutex // events is a channel were events to be broadcasted are sent + // This channel is never closed, because it's lifetime is tied to the + // life of the driver and closing creates some subtile race conditions + // between closing it and emitting events. events chan *drivers.TaskEvent // consumers is a slice of eventConsumers to broadcast events to. // access is gaurded by consumersLock RWMutex - consumers []*eventConsumer + consumers []*eventConsumer + consumersLock sync.RWMutex // ctx to allow control of event loop shutdown ctx context.Context @@ -34,6 +40,13 @@ type Eventer struct { logger hclog.Logger } +type eventConsumer struct { + timeout time.Duration + ctx context.Context + ch chan *drivers.TaskEvent + logger hclog.Logger +} + // NewEventer returns an Eventer with a running event loop that can be stopped // by closing the given stop channel func NewEventer(ctx context.Context, logger hclog.Logger) *Eventer { @@ -52,32 +65,48 @@ func (e *Eventer) eventLoop() { for { select { case <-e.ctx.Done(): - close(e.events) + e.logger.Debug("task event loop shutdown") return case event := <-e.events: - e.consumersLock.RLock() - for _, consumer := range e.consumers { - consumer.send(event) - } - e.consumersLock.RUnlock() + e.iterateConsumers(event) + case <-time.After(ConsumerGCInterval): + e.gcConsumers() } } } -type eventConsumer struct { - timeout time.Duration - ctx context.Context - ch chan *drivers.TaskEvent - logger hclog.Logger +func (e *Eventer) iterateConsumers(event *drivers.TaskEvent) { + e.consumersLock.Lock() + filtered := e.consumers[:0] + for _, consumer := range e.consumers { + select { + case <-time.After(consumer.timeout): + filtered = append(filtered, consumer) + e.logger.Warn("timeout sending event", "task_id", event.TaskID, "message", event.Message) + case <-consumer.ctx.Done(): + // consumer context finished, filtering it out of loop + close(consumer.ch) + case consumer.ch <- event: + filtered = append(filtered, consumer) + } + } + e.consumers = filtered + e.consumersLock.Unlock() } -func (c *eventConsumer) send(event *drivers.TaskEvent) { - select { - case <-time.After(c.timeout): - c.logger.Warn("timeout sending event", "task_id", event.TaskID, "message", event.Message) - case <-c.ctx.Done(): - case c.ch <- event: +func (e *Eventer) gcConsumers() { + e.consumersLock.Lock() + filtered := e.consumers[:0] + for _, consumer := range e.consumers { + select { + case <-consumer.ctx.Done(): + // consumer context finished, filtering it out of loop + default: + filtered = append(filtered, consumer) + } } + e.consumers = filtered + e.consumersLock.Unlock() } func (e *Eventer) newConsumer(ctx context.Context) *eventConsumer { @@ -98,38 +127,18 @@ func (e *Eventer) newConsumer(ctx context.Context) *eventConsumer { // TaskEvents is an implementation of the DriverPlugin.TaskEvents function func (e *Eventer) TaskEvents(ctx context.Context) (<-chan *drivers.TaskEvent, error) { consumer := e.newConsumer(ctx) - go e.handleConsumer(consumer) return consumer.ch, nil } -func (e *Eventer) handleConsumer(consumer *eventConsumer) { - // wait for consumer or eventer ctx to finish - select { - case <-consumer.ctx.Done(): - case <-e.ctx.Done(): - } - e.consumersLock.Lock() - defer e.consumersLock.Unlock() - defer close(consumer.ch) - - filtered := e.consumers[:0] - for _, c := range e.consumers { - if c != consumer { - filtered = append(filtered, c) - } - } - e.consumers = filtered -} - // EmitEvent can be used to broadcast a new event func (e *Eventer) EmitEvent(event *drivers.TaskEvent) error { select { case <-e.ctx.Done(): - return fmt.Errorf("error sending event, context canceled") + return e.ctx.Err() case e.events <- event: if e.logger.IsTrace() { - e.logger.Trace("emitting event", "event", *event) + e.logger.Trace("emitting event", "event", event) } } return nil diff --git a/plugins/drivers/utils/eventer_test.go b/plugins/drivers/utils/eventer_test.go index 6c7972a9a..763772920 100644 --- a/plugins/drivers/utils/eventer_test.go +++ b/plugins/drivers/utils/eventer_test.go @@ -15,7 +15,7 @@ func TestEventer(t *testing.T) { t.Parallel() require := require.New(t) - ctx, cancel := context.WithCancel(context.Background()) + ctx, _ := context.WithCancel(context.Background()) e := NewEventer(ctx, testlog.HCLogger(t)) events := []*drivers.TaskEvent{ @@ -33,7 +33,7 @@ func TestEventer(t *testing.T) { }, } - ctx1, cancel1 := context.WithCancel(context.Background()) + ctx1, _ := context.WithCancel(context.Background()) consumer1, err := e.TaskEvents(ctx1) require.NoError(err) ctx2 := (context.Background()) @@ -49,8 +49,8 @@ func TestEventer(t *testing.T) { for event := range consumer1 { i++ buffer1 = append(buffer1, event) - if i == 3 { - break + if i == len(events) { + return } } }() @@ -60,8 +60,8 @@ func TestEventer(t *testing.T) { for event := range consumer2 { i++ buffer2 = append(buffer2, event) - if i == 3 { - break + if i == len(events) { + return } } }() @@ -73,20 +73,45 @@ func TestEventer(t *testing.T) { wg.Wait() require.Exactly(events, buffer1) require.Exactly(events, buffer2) - cancel1() - time.Sleep(100 * time.Millisecond) +} + +func TestEventer_iterateConsumers(t *testing.T) { + t.Parallel() + require := require.New(t) + + e := &Eventer{ + events: make(chan *drivers.TaskEvent), + ctx: context.Background(), + logger: testlog.HCLogger(t), + } + + ev := &drivers.TaskEvent{ + TaskID: "a", + Timestamp: time.Now(), + } + + ctx1, cancel1 := context.WithCancel(context.Background()) + consumer, err := e.TaskEvents(ctx1) + require.NoError(err) require.Equal(1, len(e.consumers)) - require.NoError(e.EmitEvent(&drivers.TaskEvent{})) - ev, ok := <-consumer1 - require.Nil(ev) - require.False(ok) - ev, ok = <-consumer2 - require.NotNil(ev) - require.True(ok) + var wg sync.WaitGroup + wg.Add(1) + go func() { + defer wg.Done() + ev1, ok := <-consumer + require.Exactly(ev, ev1) + require.True(ok) + }() + e.iterateConsumers(ev) + wg.Wait() - cancel() - time.Sleep(100 * time.Millisecond) - require.Zero(len(e.consumers)) - require.Error(e.EmitEvent(&drivers.TaskEvent{})) + go func() { + cancel1() + e.iterateConsumers(ev) + }() + ev1, ok := <-consumer + require.False(ok) + require.Nil(ev1) + require.Equal(0, len(e.consumers)) } From 951a0841982d299816d80c9e3c91e0b52788ca05 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Tue, 2 Oct 2018 15:40:27 -0400 Subject: [PATCH 110/152] drivers/shared: move eventer to subpackage under drivers shared package --- drivers/rawexec/driver.go | 5 +++-- {plugins/drivers/utils => drivers/shared/eventer}/eventer.go | 2 +- .../drivers/utils => drivers/shared/eventer}/eventer_test.go | 2 +- 3 files changed, 5 insertions(+), 4 deletions(-) rename {plugins/drivers/utils => drivers/shared/eventer}/eventer.go (99%) rename {plugins/drivers/utils => drivers/shared/eventer}/eventer_test.go (99%) diff --git a/drivers/rawexec/driver.go b/drivers/rawexec/driver.go index 5d9f09096..b9e04916b 100644 --- a/drivers/rawexec/driver.go +++ b/drivers/rawexec/driver.go @@ -12,6 +12,7 @@ import ( plugin "github.com/hashicorp/go-plugin" "github.com/hashicorp/nomad/client/driver/executor" dstructs "github.com/hashicorp/nomad/client/driver/structs" + "github.com/hashicorp/nomad/drivers/shared/eventer" "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/drivers" "github.com/hashicorp/nomad/plugins/drivers/utils" @@ -70,7 +71,7 @@ var ( type RawExecDriver struct { // eventer is used to handle multiplexing of TaskEvents calls such that an // event can be broadcast to all callers - eventer *utils.Eventer + eventer *eventer.Eventer // config is the driver configuration set by the SetConfig RPC config *Config @@ -122,7 +123,7 @@ func NewRawExecDriver(logger hclog.Logger) drivers.DriverPlugin { ctx, cancel := context.WithCancel(context.Background()) logger = logger.Named(pluginName) return &RawExecDriver{ - eventer: utils.NewEventer(ctx, logger), + eventer: eventer.NewEventer(ctx, logger), config: &Config{}, tasks: newTaskStore(), ctx: ctx, diff --git a/plugins/drivers/utils/eventer.go b/drivers/shared/eventer/eventer.go similarity index 99% rename from plugins/drivers/utils/eventer.go rename to drivers/shared/eventer/eventer.go index 737b7f608..b8f063b3d 100644 --- a/plugins/drivers/utils/eventer.go +++ b/drivers/shared/eventer/eventer.go @@ -1,4 +1,4 @@ -package utils +package eventer import ( "sync" diff --git a/plugins/drivers/utils/eventer_test.go b/drivers/shared/eventer/eventer_test.go similarity index 99% rename from plugins/drivers/utils/eventer_test.go rename to drivers/shared/eventer/eventer_test.go index 763772920..0b522e49c 100644 --- a/plugins/drivers/utils/eventer_test.go +++ b/drivers/shared/eventer/eventer_test.go @@ -1,4 +1,4 @@ -package utils +package eventer import ( "context" From 5fb95c1a000a2cdd34abde1379e4e2b6a1afc85f Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 3 Oct 2018 12:08:26 -0400 Subject: [PATCH 111/152] drivers/shared: added func comment to eventer --- drivers/shared/eventer/eventer.go | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/drivers/shared/eventer/eventer.go b/drivers/shared/eventer/eventer.go index b8f063b3d..a68a20162 100644 --- a/drivers/shared/eventer/eventer.go +++ b/drivers/shared/eventer/eventer.go @@ -65,7 +65,7 @@ func (e *Eventer) eventLoop() { for { select { case <-e.ctx.Done(): - e.logger.Debug("task event loop shutdown") + e.logger.Trace("task event loop shutdown") return case event := <-e.events: e.iterateConsumers(event) @@ -75,6 +75,8 @@ func (e *Eventer) eventLoop() { } } +// iterateConsumers will iterate through all consumers and broadcast the event, +// cleaning up any consumers that have closed their context func (e *Eventer) iterateConsumers(event *drivers.TaskEvent) { e.consumersLock.Lock() filtered := e.consumers[:0] From 6c9d9d517340d173d298d44e2ba62abf5e25e9c7 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Thu, 4 Oct 2018 15:45:46 -0700 Subject: [PATCH 112/152] move files around --- .../alloc_runner.go | 2 +- .../alloc_runner_health_watcher.go | 0 .../alloc_runner_test.go | 2 +- .../taskrunner/consul_template.go | 0 .../taskrunner/consul_template_test.go | 0 .../taskrunner/getters.go | 0 .../taskrunner/task_runner.go | 4 ++-- .../taskrunner/task_runner_test.go | 2 +- .../taskrunner/task_runner_unix_test.go | 0 client/{allocrunner => allocrunnerdeprecated}/testing.go | 0 client/allocrunnerv2/alloc_runner.go | 3 +-- client/allocrunnerv2/interfaces/runner.go | 8 ++++++++ client/allocrunnerv2/taskrunner/artifact_hook.go | 2 +- .../taskrunner}/getter/getter.go | 0 .../taskrunner}/getter/getter_test.go | 0 .../taskrunner}/getter/test-fixtures/archive.tar.gz | 0 .../getter/test-fixtures/archive/exist/my.config | 0 .../getter/test-fixtures/archive/new/my.config | 0 .../taskrunner}/getter/test-fixtures/archive/test.sh | 0 .../taskrunner}/getter/test-fixtures/test.sh | 0 .../taskrunner/restarts/restarts.go | 0 .../taskrunner/restarts/restarts_test.go | 0 client/allocrunnerv2/taskrunner/task_runner.go | 2 +- client/client.go | 8 ++++---- command/alloc_status.go | 2 +- 25 files changed, 21 insertions(+), 14 deletions(-) rename client/{allocrunner => allocrunnerdeprecated}/alloc_runner.go (99%) rename client/{allocrunner => allocrunnerdeprecated}/alloc_runner_health_watcher.go (100%) rename client/{allocrunner => allocrunnerdeprecated}/alloc_runner_test.go (99%) rename client/{allocrunner => allocrunnerdeprecated}/taskrunner/consul_template.go (100%) rename client/{allocrunner => allocrunnerdeprecated}/taskrunner/consul_template_test.go (100%) rename client/{allocrunner => allocrunnerdeprecated}/taskrunner/getters.go (100%) rename client/{allocrunner => allocrunnerdeprecated}/taskrunner/task_runner.go (99%) rename client/{allocrunner => allocrunnerdeprecated}/taskrunner/task_runner_test.go (99%) rename client/{allocrunner => allocrunnerdeprecated}/taskrunner/task_runner_unix_test.go (100%) rename client/{allocrunner => allocrunnerdeprecated}/testing.go (100%) rename client/{allocrunner => allocrunnerv2/taskrunner}/getter/getter.go (100%) rename client/{allocrunner => allocrunnerv2/taskrunner}/getter/getter_test.go (100%) rename client/{allocrunner => allocrunnerv2/taskrunner}/getter/test-fixtures/archive.tar.gz (100%) rename client/{allocrunner => allocrunnerv2/taskrunner}/getter/test-fixtures/archive/exist/my.config (100%) rename client/{allocrunner => allocrunnerv2/taskrunner}/getter/test-fixtures/archive/new/my.config (100%) rename client/{allocrunner => allocrunnerv2/taskrunner}/getter/test-fixtures/archive/test.sh (100%) rename client/{allocrunner => allocrunnerv2/taskrunner}/getter/test-fixtures/test.sh (100%) rename client/{allocrunner => allocrunnerv2}/taskrunner/restarts/restarts.go (100%) rename client/{allocrunner => allocrunnerv2}/taskrunner/restarts/restarts_test.go (100%) diff --git a/client/allocrunner/alloc_runner.go b/client/allocrunnerdeprecated/alloc_runner.go similarity index 99% rename from client/allocrunner/alloc_runner.go rename to client/allocrunnerdeprecated/alloc_runner.go index f436340f0..97351ccdc 100644 --- a/client/allocrunner/alloc_runner.go +++ b/client/allocrunnerdeprecated/alloc_runner.go @@ -12,7 +12,7 @@ import ( "github.com/boltdb/bolt" "github.com/hashicorp/go-multierror" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunner/taskrunner" + "github.com/hashicorp/nomad/client/allocrunnerdeprecated/taskrunner" "github.com/hashicorp/nomad/client/allocwatcher" "github.com/hashicorp/nomad/client/config" consulApi "github.com/hashicorp/nomad/client/consul" diff --git a/client/allocrunner/alloc_runner_health_watcher.go b/client/allocrunnerdeprecated/alloc_runner_health_watcher.go similarity index 100% rename from client/allocrunner/alloc_runner_health_watcher.go rename to client/allocrunnerdeprecated/alloc_runner_health_watcher.go diff --git a/client/allocrunner/alloc_runner_test.go b/client/allocrunnerdeprecated/alloc_runner_test.go similarity index 99% rename from client/allocrunner/alloc_runner_test.go rename to client/allocrunnerdeprecated/alloc_runner_test.go index f6dc39c49..38fc9927c 100644 --- a/client/allocrunner/alloc_runner_test.go +++ b/client/allocrunnerdeprecated/alloc_runner_test.go @@ -19,7 +19,7 @@ import ( "github.com/hashicorp/nomad/testutil" "github.com/stretchr/testify/assert" - "github.com/hashicorp/nomad/client/allocrunner/taskrunner" + "github.com/hashicorp/nomad/client/allocrunnerdeprecated/taskrunner" consulApi "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/state" "github.com/stretchr/testify/require" diff --git a/client/allocrunner/taskrunner/consul_template.go b/client/allocrunnerdeprecated/taskrunner/consul_template.go similarity index 100% rename from client/allocrunner/taskrunner/consul_template.go rename to client/allocrunnerdeprecated/taskrunner/consul_template.go diff --git a/client/allocrunner/taskrunner/consul_template_test.go b/client/allocrunnerdeprecated/taskrunner/consul_template_test.go similarity index 100% rename from client/allocrunner/taskrunner/consul_template_test.go rename to client/allocrunnerdeprecated/taskrunner/consul_template_test.go diff --git a/client/allocrunner/taskrunner/getters.go b/client/allocrunnerdeprecated/taskrunner/getters.go similarity index 100% rename from client/allocrunner/taskrunner/getters.go rename to client/allocrunnerdeprecated/taskrunner/getters.go diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunnerdeprecated/taskrunner/task_runner.go similarity index 99% rename from client/allocrunner/taskrunner/task_runner.go rename to client/allocrunnerdeprecated/taskrunner/task_runner.go index 386fb672a..d7054334c 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunnerdeprecated/taskrunner/task_runner.go @@ -21,8 +21,8 @@ import ( "github.com/hashicorp/go-multierror" version "github.com/hashicorp/go-version" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunner/getter" - "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/getter" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/restarts" "github.com/hashicorp/nomad/client/config" consulApi "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/driver" diff --git a/client/allocrunner/taskrunner/task_runner_test.go b/client/allocrunnerdeprecated/taskrunner/task_runner_test.go similarity index 99% rename from client/allocrunner/taskrunner/task_runner_test.go rename to client/allocrunnerdeprecated/taskrunner/task_runner_test.go index 68e082616..f5ba3063a 100644 --- a/client/allocrunner/taskrunner/task_runner_test.go +++ b/client/allocrunnerdeprecated/taskrunner/task_runner_test.go @@ -16,7 +16,7 @@ import ( "github.com/boltdb/bolt" "github.com/golang/snappy" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/restarts" "github.com/hashicorp/nomad/client/config" consulApi "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/driver/env" diff --git a/client/allocrunner/taskrunner/task_runner_unix_test.go b/client/allocrunnerdeprecated/taskrunner/task_runner_unix_test.go similarity index 100% rename from client/allocrunner/taskrunner/task_runner_unix_test.go rename to client/allocrunnerdeprecated/taskrunner/task_runner_unix_test.go diff --git a/client/allocrunner/testing.go b/client/allocrunnerdeprecated/testing.go similarity index 100% rename from client/allocrunner/testing.go rename to client/allocrunnerdeprecated/testing.go diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go index 19ff1b32b..ed77eddf5 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunnerv2/alloc_runner.go @@ -9,7 +9,6 @@ import ( log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunner" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" "github.com/hashicorp/nomad/client/allocrunnerv2/state" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" @@ -501,7 +500,7 @@ func (ar *allocRunner) IsMigrating() bool { return ar.prevAllocWatcher.IsMigrating() } -func (ar *allocRunner) StatsReporter() allocrunner.AllocStatsReporter { +func (ar *allocRunner) StatsReporter() interfaces.AllocStatsReporter { return ar } diff --git a/client/allocrunnerv2/interfaces/runner.go b/client/allocrunnerv2/interfaces/runner.go index 3925e2384..4658e3aa3 100644 --- a/client/allocrunnerv2/interfaces/runner.go +++ b/client/allocrunnerv2/interfaces/runner.go @@ -3,6 +3,8 @@ package interfaces import ( "github.com/hashicorp/nomad/client/allocrunnerv2/state" "github.com/hashicorp/nomad/nomad/structs" + + cstructs "github.com/hashicorp/nomad/client/structs" ) // AllocRunner is the interface for an allocation runner. @@ -25,3 +27,9 @@ type TaskStateHandler interface { // TaskStateUpdated is used to emit updated task state TaskStateUpdated(task string, state *structs.TaskState) } + +// AllocStatsReporter gives acess to the latest resource usage from the +// allocation +type AllocStatsReporter interface { + LatestAllocStats(taskFilter string) (*cstructs.AllocResourceUsage, error) +} diff --git a/client/allocrunnerv2/taskrunner/artifact_hook.go b/client/allocrunnerv2/taskrunner/artifact_hook.go index 2253fdfe3..de8e1703e 100644 --- a/client/allocrunnerv2/taskrunner/artifact_hook.go +++ b/client/allocrunnerv2/taskrunner/artifact_hook.go @@ -5,8 +5,8 @@ import ( "fmt" log "github.com/hashicorp/go-hclog" - "github.com/hashicorp/nomad/client/allocrunner/getter" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/getter" ti "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" "github.com/hashicorp/nomad/nomad/structs" ) diff --git a/client/allocrunner/getter/getter.go b/client/allocrunnerv2/taskrunner/getter/getter.go similarity index 100% rename from client/allocrunner/getter/getter.go rename to client/allocrunnerv2/taskrunner/getter/getter.go diff --git a/client/allocrunner/getter/getter_test.go b/client/allocrunnerv2/taskrunner/getter/getter_test.go similarity index 100% rename from client/allocrunner/getter/getter_test.go rename to client/allocrunnerv2/taskrunner/getter/getter_test.go diff --git a/client/allocrunner/getter/test-fixtures/archive.tar.gz b/client/allocrunnerv2/taskrunner/getter/test-fixtures/archive.tar.gz similarity index 100% rename from client/allocrunner/getter/test-fixtures/archive.tar.gz rename to client/allocrunnerv2/taskrunner/getter/test-fixtures/archive.tar.gz diff --git a/client/allocrunner/getter/test-fixtures/archive/exist/my.config b/client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/exist/my.config similarity index 100% rename from client/allocrunner/getter/test-fixtures/archive/exist/my.config rename to client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/exist/my.config diff --git a/client/allocrunner/getter/test-fixtures/archive/new/my.config b/client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/new/my.config similarity index 100% rename from client/allocrunner/getter/test-fixtures/archive/new/my.config rename to client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/new/my.config diff --git a/client/allocrunner/getter/test-fixtures/archive/test.sh b/client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/test.sh similarity index 100% rename from client/allocrunner/getter/test-fixtures/archive/test.sh rename to client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/test.sh diff --git a/client/allocrunner/getter/test-fixtures/test.sh b/client/allocrunnerv2/taskrunner/getter/test-fixtures/test.sh similarity index 100% rename from client/allocrunner/getter/test-fixtures/test.sh rename to client/allocrunnerv2/taskrunner/getter/test-fixtures/test.sh diff --git a/client/allocrunner/taskrunner/restarts/restarts.go b/client/allocrunnerv2/taskrunner/restarts/restarts.go similarity index 100% rename from client/allocrunner/taskrunner/restarts/restarts.go rename to client/allocrunnerv2/taskrunner/restarts/restarts.go diff --git a/client/allocrunner/taskrunner/restarts/restarts_test.go b/client/allocrunnerv2/taskrunner/restarts/restarts_test.go similarity index 100% rename from client/allocrunner/taskrunner/restarts/restarts_test.go rename to client/allocrunnerv2/taskrunner/restarts/restarts_test.go diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go index cf541e720..1171e8aaa 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -9,8 +9,8 @@ import ( metrics "github.com/armon/go-metrics" log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/restarts" "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/consul" diff --git a/client/client.go b/client/client.go index 2fbeaf7c5..837708bb9 100644 --- a/client/client.go +++ b/client/client.go @@ -18,6 +18,7 @@ import ( consulapi "github.com/hashicorp/consul/api" hclog "github.com/hashicorp/go-hclog" multierror "github.com/hashicorp/go-multierror" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" arstate "github.com/hashicorp/nomad/client/allocrunnerv2/state" consulApi "github.com/hashicorp/nomad/client/consul" cstructs "github.com/hashicorp/nomad/client/structs" @@ -27,7 +28,6 @@ import ( "github.com/hashicorp/consul/lib" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunner" "github.com/hashicorp/nomad/client/allocrunnerv2" "github.com/hashicorp/nomad/client/allocwatcher" "github.com/hashicorp/nomad/client/config" @@ -94,7 +94,7 @@ const ( type ClientStatsReporter interface { // GetAllocStats returns the AllocStatsReporter for the passed allocation. // If it does not exist an error is reported. - GetAllocStats(allocID string) (allocrunner.AllocStatsReporter, error) + GetAllocStats(allocID string) (interfaces.AllocStatsReporter, error) // LatestHostStats returns the latest resource usage stats for the host LatestHostStats() *stats.HostStats @@ -113,7 +113,7 @@ type AllocRunner interface { Listener() *cstructs.AllocListener Restore() error Run() - StatsReporter() allocrunner.AllocStatsReporter + StatsReporter() interfaces.AllocStatsReporter Update(*structs.Allocation) WaitCh() <-chan struct{} } @@ -600,7 +600,7 @@ func (c *Client) StatsReporter() ClientStatsReporter { return c } -func (c *Client) GetAllocStats(allocID string) (allocrunner.AllocStatsReporter, error) { +func (c *Client) GetAllocStats(allocID string) (interfaces.AllocStatsReporter, error) { c.allocLock.RLock() defer c.allocLock.RUnlock() ar, ok := c.allocs[allocID] diff --git a/command/alloc_status.go b/command/alloc_status.go index 773791c59..c4373fd7c 100644 --- a/command/alloc_status.go +++ b/command/alloc_status.go @@ -12,7 +12,7 @@ import ( "github.com/hashicorp/nomad/api" "github.com/hashicorp/nomad/api/contexts" - "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/restarts" "github.com/posener/complete" ) From 7d9c069f09fd59f7ddf442ccae03d573a1300d7a Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Thu, 4 Oct 2018 15:48:27 -0700 Subject: [PATCH 113/152] skip building deprecated files --- client/allocrunnerdeprecated/alloc_runner.go | 2 ++ client/allocrunnerdeprecated/alloc_runner_health_watcher.go | 2 ++ client/allocrunnerdeprecated/alloc_runner_test.go | 2 ++ client/allocrunnerdeprecated/taskrunner/consul_template.go | 2 ++ client/allocrunnerdeprecated/taskrunner/consul_template_test.go | 2 ++ client/allocrunnerdeprecated/taskrunner/getters.go | 2 ++ client/allocrunnerdeprecated/taskrunner/task_runner.go | 2 ++ client/allocrunnerdeprecated/taskrunner/task_runner_test.go | 2 ++ .../allocrunnerdeprecated/taskrunner/task_runner_unix_test.go | 2 +- client/allocrunnerdeprecated/testing.go | 2 ++ 10 files changed, 19 insertions(+), 1 deletion(-) diff --git a/client/allocrunnerdeprecated/alloc_runner.go b/client/allocrunnerdeprecated/alloc_runner.go index 97351ccdc..0fb952fe6 100644 --- a/client/allocrunnerdeprecated/alloc_runner.go +++ b/client/allocrunnerdeprecated/alloc_runner.go @@ -1,3 +1,5 @@ +// +build deprecated + package allocrunner import ( diff --git a/client/allocrunnerdeprecated/alloc_runner_health_watcher.go b/client/allocrunnerdeprecated/alloc_runner_health_watcher.go index 81c726f73..a457256e6 100644 --- a/client/allocrunnerdeprecated/alloc_runner_health_watcher.go +++ b/client/allocrunnerdeprecated/alloc_runner_health_watcher.go @@ -1,3 +1,5 @@ +// +build deprecated + package allocrunner import ( diff --git a/client/allocrunnerdeprecated/alloc_runner_test.go b/client/allocrunnerdeprecated/alloc_runner_test.go index 38fc9927c..7b4c56bb4 100644 --- a/client/allocrunnerdeprecated/alloc_runner_test.go +++ b/client/allocrunnerdeprecated/alloc_runner_test.go @@ -1,3 +1,5 @@ +// +build deprecated + package allocrunner import ( diff --git a/client/allocrunnerdeprecated/taskrunner/consul_template.go b/client/allocrunnerdeprecated/taskrunner/consul_template.go index ac7aed79c..25f1e4d5f 100644 --- a/client/allocrunnerdeprecated/taskrunner/consul_template.go +++ b/client/allocrunnerdeprecated/taskrunner/consul_template.go @@ -1,3 +1,5 @@ +// +build deprecated + package taskrunner import ( diff --git a/client/allocrunnerdeprecated/taskrunner/consul_template_test.go b/client/allocrunnerdeprecated/taskrunner/consul_template_test.go index d07744493..5f722cbb0 100644 --- a/client/allocrunnerdeprecated/taskrunner/consul_template_test.go +++ b/client/allocrunnerdeprecated/taskrunner/consul_template_test.go @@ -1,3 +1,5 @@ +// +build deprecated + package taskrunner import ( diff --git a/client/allocrunnerdeprecated/taskrunner/getters.go b/client/allocrunnerdeprecated/taskrunner/getters.go index 0a4bca06f..6fb75bc01 100644 --- a/client/allocrunnerdeprecated/taskrunner/getters.go +++ b/client/allocrunnerdeprecated/taskrunner/getters.go @@ -1,3 +1,5 @@ +// +build deprecated + package taskrunner // Name returns the name of the task diff --git a/client/allocrunnerdeprecated/taskrunner/task_runner.go b/client/allocrunnerdeprecated/taskrunner/task_runner.go index d7054334c..d8b58e480 100644 --- a/client/allocrunnerdeprecated/taskrunner/task_runner.go +++ b/client/allocrunnerdeprecated/taskrunner/task_runner.go @@ -1,3 +1,5 @@ +// +build deprecated + package taskrunner import ( diff --git a/client/allocrunnerdeprecated/taskrunner/task_runner_test.go b/client/allocrunnerdeprecated/taskrunner/task_runner_test.go index f5ba3063a..50dcc3ad3 100644 --- a/client/allocrunnerdeprecated/taskrunner/task_runner_test.go +++ b/client/allocrunnerdeprecated/taskrunner/task_runner_test.go @@ -1,3 +1,5 @@ +// +build deprecated + package taskrunner import ( diff --git a/client/allocrunnerdeprecated/taskrunner/task_runner_unix_test.go b/client/allocrunnerdeprecated/taskrunner/task_runner_unix_test.go index 1bb397f46..f911767f6 100644 --- a/client/allocrunnerdeprecated/taskrunner/task_runner_unix_test.go +++ b/client/allocrunnerdeprecated/taskrunner/task_runner_unix_test.go @@ -1,4 +1,4 @@ -// +build !windows +// +build deprecated,!windows package taskrunner diff --git a/client/allocrunnerdeprecated/testing.go b/client/allocrunnerdeprecated/testing.go index 7c4853623..87c0eaae7 100644 --- a/client/allocrunnerdeprecated/testing.go +++ b/client/allocrunnerdeprecated/testing.go @@ -1,3 +1,5 @@ +// +build deprecated + package allocrunner import ( From 9baa7402efbfa3a213b6f8d60634160c69a364fb Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Thu, 4 Oct 2018 16:16:08 -0700 Subject: [PATCH 114/152] fix test compiling --- client/allocwatcher/alloc_watcher_test.go | 2 +- client/servers/manager_internal_test.go | 6 +++--- client/servers/manager_test.go | 6 +++--- client/stats/cpu_test.go | 2 +- client/vaultclient/vaultclient_test.go | 8 ++++---- 5 files changed, 12 insertions(+), 12 deletions(-) diff --git a/client/allocwatcher/alloc_watcher_test.go b/client/allocwatcher/alloc_watcher_test.go index 64509b03f..b8a3c8581 100644 --- a/client/allocwatcher/alloc_watcher_test.go +++ b/client/allocwatcher/alloc_watcher_test.go @@ -45,7 +45,7 @@ func newFakeAllocRunner(t *testing.T, logger hclog.Logger) *fakeAllocRunner { return &fakeAllocRunner{ alloc: alloc, AllocDir: allocdir.NewAllocDir(logger, path), - Broadcaster: cstructs.NewAllocBroadcaster(alloc), + Broadcaster: cstructs.NewAllocBroadcaster(), } } diff --git a/client/servers/manager_internal_test.go b/client/servers/manager_internal_test.go index dc17be1f9..f28074dea 100644 --- a/client/servers/manager_internal_test.go +++ b/client/servers/manager_internal_test.go @@ -36,14 +36,14 @@ func (cp *fauxConnPool) Ping(net.Addr) error { } func testManager(t *testing.T) (m *Manager) { - logger := testlog.Logger(t) + logger := testlog.HCLogger(t) shutdownCh := make(chan struct{}) m = New(logger, shutdownCh, &fauxConnPool{}) return m } func testManagerFailProb(t *testing.T, failPct float64) (m *Manager) { - logger := testlog.Logger(t) + logger := testlog.HCLogger(t) shutdownCh := make(chan struct{}) m = New(logger, shutdownCh, &fauxConnPool{failPct: failPct}) return m @@ -136,7 +136,7 @@ func TestManagerInternal_refreshServerRebalanceTimer(t *testing.T) { {1000000, 19, 10 * time.Minute}, } - logger := testlog.Logger(t) + logger := testlog.HCLogger(t) shutdownCh := make(chan struct{}) for _, s := range clusters { diff --git a/client/servers/manager_test.go b/client/servers/manager_test.go index 235064797..e33ab4119 100644 --- a/client/servers/manager_test.go +++ b/client/servers/manager_test.go @@ -33,14 +33,14 @@ func (cp *fauxConnPool) Ping(net.Addr) error { } func testManager(t *testing.T) (m *servers.Manager) { - logger := testlog.Logger(t) + logger := testlog.HCLogger(t) shutdownCh := make(chan struct{}) m = servers.New(logger, shutdownCh, &fauxConnPool{}) return m } func testManagerFailProb(t *testing.T, failPct float64) (m *servers.Manager) { - logger := testlog.Logger(t) + logger := testlog.HCLogger(t) shutdownCh := make(chan struct{}) m = servers.New(logger, shutdownCh, &fauxConnPool{failPct: failPct}) return m @@ -113,7 +113,7 @@ func TestServers_FindServer(t *testing.T) { } func TestServers_New(t *testing.T) { - logger := testlog.Logger(t) + logger := testlog.HCLogger(t) shutdownCh := make(chan struct{}) m := servers.New(logger, shutdownCh, &fauxConnPool{}) if m == nil { diff --git a/client/stats/cpu_test.go b/client/stats/cpu_test.go index b963d91a6..eedd66e47 100644 --- a/client/stats/cpu_test.go +++ b/client/stats/cpu_test.go @@ -26,7 +26,7 @@ func TestHostStats_CPU(t *testing.T) { assert := assert.New(t) assert.Nil(shelpers.Init()) - logger := testlog.Logger(t) + logger := testlog.HCLogger(t) cwd, err := os.Getwd() assert.Nil(err) hs := NewHostStatsCollector(logger, cwd) diff --git a/client/vaultclient/vaultclient_test.go b/client/vaultclient/vaultclient_test.go index faa9c6695..d9d334d5d 100644 --- a/client/vaultclient/vaultclient_test.go +++ b/client/vaultclient/vaultclient_test.go @@ -16,7 +16,7 @@ func TestVaultClient_TokenRenewals(t *testing.T) { v := testutil.NewTestVault(t) defer v.Stop() - logger := testlog.Logger(t) + logger := testlog.HCLogger(t) v.Config.ConnectionRetryIntv = 100 * time.Millisecond v.Config.TaskTokenTTL = "4s" c, err := NewVaultClient(v.Config, logger, nil) @@ -100,7 +100,7 @@ func TestVaultClient_Heap(t *testing.T) { conf.VaultConfig.Token = "testvaulttoken" conf.VaultConfig.TaskTokenTTL = "10s" - logger := testlog.Logger(t) + logger := testlog.HCLogger(t) c, err := NewVaultClient(conf.VaultConfig, logger, nil) if err != nil { t.Fatal(err) @@ -203,7 +203,7 @@ func TestVaultClient_RenewNonRenewableLease(t *testing.T) { v := testutil.NewTestVault(t) defer v.Stop() - logger := testlog.Logger(t) + logger := testlog.HCLogger(t) v.Config.ConnectionRetryIntv = 100 * time.Millisecond v.Config.TaskTokenTTL = "4s" c, err := NewVaultClient(v.Config, logger, nil) @@ -252,7 +252,7 @@ func TestVaultClient_RenewNonexistentLease(t *testing.T) { v := testutil.NewTestVault(t) defer v.Stop() - logger := testlog.Logger(t) + logger := testlog.HCLogger(t) v.Config.ConnectionRetryIntv = 100 * time.Millisecond v.Config.TaskTokenTTL = "4s" c, err := NewVaultClient(v.Config, logger, nil) From 45e41cca03b6dc37bc48f7dc199c6c914e493018 Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Thu, 4 Oct 2018 16:22:01 -0700 Subject: [PATCH 115/152] allocrunnerv2 -> allocrunner --- .../alloc_runner.go | 8 ++++---- .../alloc_runner_hooks.go | 4 ++-- .../alloc_runner_test.go | 4 ++-- .../allocdir_hook.go | 2 +- client/{allocrunnerv2 => allocrunner}/config.go | 2 +- .../health_hook.go | 4 ++-- .../health_hook_test.go | 4 ++-- .../interfaces/runner.go | 2 +- .../interfaces/runner_lifecycle.go | 2 +- .../interfaces/task_lifecycle.go | 0 .../migrate_hook.go | 2 +- .../state/state.go | 0 .../taskrunner/artifact_hook.go | 6 +++--- .../taskrunner/errors.go | 0 .../taskrunner/getter/getter.go | 0 .../taskrunner/getter/getter_test.go | 0 .../getter/test-fixtures/archive.tar.gz | 0 .../getter/test-fixtures/archive/exist/my.config | 0 .../getter/test-fixtures/archive/new/my.config | 0 .../getter/test-fixtures/archive/test.sh | 0 .../taskrunner/getter/test-fixtures/test.sh | 0 .../taskrunner/handleproxy.go | 0 .../taskrunner/handleproxy_test.go | 0 .../taskrunner/interfaces/events.go | 0 .../taskrunner/interfaces/lifecycle.go | 0 .../taskrunner/lifecycle.go | 0 .../taskrunner/logmon_hook.go | 2 +- .../taskrunner/logmon_hook_test.go | 2 +- .../taskrunner/restarts/restarts.go | 0 .../taskrunner/restarts/restarts_test.go | 0 .../taskrunner/service_hook.go | 2 +- .../taskrunner/shutdown_delay_hook.go | 2 +- .../taskrunner/state/state.go | 0 .../taskrunner/stats_hook.go | 2 +- .../taskrunner/stats_hook_test.go | 2 +- .../taskrunner/task_dir_hook.go | 2 +- .../taskrunner/task_runner.go | 6 +++--- .../taskrunner/task_runner_getters.go | 0 .../taskrunner/task_runner_hooks.go | 4 ++-- .../taskrunner/template/template.go | 2 +- .../taskrunner/template/template_test.go | 0 .../taskrunner/template_hook.go | 6 +++--- .../taskrunner/validate_hook.go | 2 +- .../taskrunner/validate_hook_test.go | 0 .../taskrunner/vault_hook.go | 4 ++-- client/allocrunner/util.go | 1 + client/allocrunnerdeprecated/alloc_runner.go | 6 +++--- .../taskrunner/task_runner.go | 16 ++++++++-------- .../taskrunner/task_runner_test.go | 4 ++-- client/allocrunnerv2/util.go | 1 - client/client.go | 14 +++++++------- client/state/interface.go | 2 +- client/state/noopdb.go | 2 +- client/state/state_database.go | 2 +- command/alloc_status.go | 2 +- 55 files changed, 64 insertions(+), 64 deletions(-) rename client/{allocrunnerv2 => allocrunner}/alloc_runner.go (98%) rename client/{allocrunnerv2 => allocrunner}/alloc_runner_hooks.go (98%) rename client/{allocrunnerv2 => allocrunner}/alloc_runner_test.go (97%) rename client/{allocrunnerv2 => allocrunner}/allocdir_hook.go (96%) rename client/{allocrunnerv2 => allocrunner}/config.go (98%) rename client/{allocrunnerv2 => allocrunner}/health_hook.go (98%) rename client/{allocrunnerv2 => allocrunner}/health_hook_test.go (99%) rename client/{allocrunnerv2 => allocrunner}/interfaces/runner.go (94%) rename client/{allocrunnerv2 => allocrunner}/interfaces/runner_lifecycle.go (93%) rename client/{allocrunnerv2 => allocrunner}/interfaces/task_lifecycle.go (100%) rename client/{allocrunnerv2 => allocrunner}/migrate_hook.go (98%) rename client/{allocrunnerv2 => allocrunner}/state/state.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/artifact_hook.go (86%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/errors.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/getter/getter.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/getter/getter_test.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/getter/test-fixtures/archive.tar.gz (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/getter/test-fixtures/archive/exist/my.config (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/getter/test-fixtures/archive/new/my.config (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/getter/test-fixtures/archive/test.sh (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/getter/test-fixtures/test.sh (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/handleproxy.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/handleproxy_test.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/interfaces/events.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/interfaces/lifecycle.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/lifecycle.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/logmon_hook.go (97%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/logmon_hook_test.go (74%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/restarts/restarts.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/restarts/restarts_test.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/service_hook.go (98%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/shutdown_delay_hook.go (92%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/state/state.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/stats_hook.go (97%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/stats_hook_test.go (98%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/task_dir_hook.go (95%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/task_runner.go (99%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/task_runner_getters.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/task_runner_hooks.go (98%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/template/template.go (99%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/template/template_test.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/template_hook.go (94%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/validate_hook.go (96%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/validate_hook_test.go (100%) rename client/{allocrunnerv2 => allocrunner}/taskrunner/vault_hook.go (98%) create mode 100644 client/allocrunner/util.go delete mode 100644 client/allocrunnerv2/util.go diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunner/alloc_runner.go similarity index 98% rename from client/allocrunnerv2/alloc_runner.go rename to client/allocrunner/alloc_runner.go index ed77eddf5..f0e09b5cd 100644 --- a/client/allocrunnerv2/alloc_runner.go +++ b/client/allocrunner/alloc_runner.go @@ -1,4 +1,4 @@ -package allocrunnerv2 +package allocrunner import ( "context" @@ -9,9 +9,9 @@ import ( log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - "github.com/hashicorp/nomad/client/allocrunnerv2/state" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/state" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner" "github.com/hashicorp/nomad/client/allocwatcher" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/consul" diff --git a/client/allocrunnerv2/alloc_runner_hooks.go b/client/allocrunner/alloc_runner_hooks.go similarity index 98% rename from client/allocrunnerv2/alloc_runner_hooks.go rename to client/allocrunner/alloc_runner_hooks.go index 6936fd654..db0d9a506 100644 --- a/client/allocrunnerv2/alloc_runner_hooks.go +++ b/client/allocrunner/alloc_runner_hooks.go @@ -1,4 +1,4 @@ -package allocrunnerv2 +package allocrunner import ( "context" @@ -6,7 +6,7 @@ import ( "time" multierror "github.com/hashicorp/go-multierror" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" "github.com/hashicorp/nomad/nomad/structs" ) diff --git a/client/allocrunnerv2/alloc_runner_test.go b/client/allocrunner/alloc_runner_test.go similarity index 97% rename from client/allocrunnerv2/alloc_runner_test.go rename to client/allocrunner/alloc_runner_test.go index 99460f758..925375ab3 100644 --- a/client/allocrunnerv2/alloc_runner_test.go +++ b/client/allocrunner/alloc_runner_test.go @@ -1,4 +1,4 @@ -package allocrunnerv2 +package allocrunner import ( "testing" @@ -43,7 +43,7 @@ func TestAllocRunner_AllocState_Initialized(t *testing.T) { import ( "testing" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" clientconfig "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/nomad/mock" diff --git a/client/allocrunnerv2/allocdir_hook.go b/client/allocrunner/allocdir_hook.go similarity index 96% rename from client/allocrunnerv2/allocdir_hook.go rename to client/allocrunner/allocdir_hook.go index f4331a9da..bfab569ec 100644 --- a/client/allocrunnerv2/allocdir_hook.go +++ b/client/allocrunner/allocdir_hook.go @@ -1,4 +1,4 @@ -package allocrunnerv2 +package allocrunner import ( "context" diff --git a/client/allocrunnerv2/config.go b/client/allocrunner/config.go similarity index 98% rename from client/allocrunnerv2/config.go rename to client/allocrunner/config.go index 50b8a672a..f15ac255d 100644 --- a/client/allocrunnerv2/config.go +++ b/client/allocrunner/config.go @@ -1,4 +1,4 @@ -package allocrunnerv2 +package allocrunner import ( log "github.com/hashicorp/go-hclog" diff --git a/client/allocrunnerv2/health_hook.go b/client/allocrunner/health_hook.go similarity index 98% rename from client/allocrunnerv2/health_hook.go rename to client/allocrunner/health_hook.go index 9aa5e2754..ff950469e 100644 --- a/client/allocrunnerv2/health_hook.go +++ b/client/allocrunner/health_hook.go @@ -1,4 +1,4 @@ -package allocrunnerv2 +package allocrunner import ( "context" @@ -8,7 +8,7 @@ import ( log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allochealth" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" "github.com/hashicorp/nomad/client/consul" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" diff --git a/client/allocrunnerv2/health_hook_test.go b/client/allocrunner/health_hook_test.go similarity index 99% rename from client/allocrunnerv2/health_hook_test.go rename to client/allocrunner/health_hook_test.go index c555297a6..45866ec73 100644 --- a/client/allocrunnerv2/health_hook_test.go +++ b/client/allocrunner/health_hook_test.go @@ -1,4 +1,4 @@ -package allocrunnerv2 +package allocrunner import ( "context" @@ -7,7 +7,7 @@ import ( "time" consulapi "github.com/hashicorp/consul/api" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" "github.com/hashicorp/nomad/client/consul" cstructs "github.com/hashicorp/nomad/client/structs" agentconsul "github.com/hashicorp/nomad/command/agent/consul" diff --git a/client/allocrunnerv2/interfaces/runner.go b/client/allocrunner/interfaces/runner.go similarity index 94% rename from client/allocrunnerv2/interfaces/runner.go rename to client/allocrunner/interfaces/runner.go index 4658e3aa3..e488d932a 100644 --- a/client/allocrunnerv2/interfaces/runner.go +++ b/client/allocrunner/interfaces/runner.go @@ -1,7 +1,7 @@ package interfaces import ( - "github.com/hashicorp/nomad/client/allocrunnerv2/state" + "github.com/hashicorp/nomad/client/allocrunner/state" "github.com/hashicorp/nomad/nomad/structs" cstructs "github.com/hashicorp/nomad/client/structs" diff --git a/client/allocrunnerv2/interfaces/runner_lifecycle.go b/client/allocrunner/interfaces/runner_lifecycle.go similarity index 93% rename from client/allocrunnerv2/interfaces/runner_lifecycle.go rename to client/allocrunner/interfaces/runner_lifecycle.go index f3150e60c..b19f284a9 100644 --- a/client/allocrunnerv2/interfaces/runner_lifecycle.go +++ b/client/allocrunner/interfaces/runner_lifecycle.go @@ -3,7 +3,7 @@ package interfaces import ( "context" - "github.com/hashicorp/nomad/client/allocrunnerv2/state" + "github.com/hashicorp/nomad/client/allocrunner/state" "github.com/hashicorp/nomad/nomad/structs" ) diff --git a/client/allocrunnerv2/interfaces/task_lifecycle.go b/client/allocrunner/interfaces/task_lifecycle.go similarity index 100% rename from client/allocrunnerv2/interfaces/task_lifecycle.go rename to client/allocrunner/interfaces/task_lifecycle.go diff --git a/client/allocrunnerv2/migrate_hook.go b/client/allocrunner/migrate_hook.go similarity index 98% rename from client/allocrunnerv2/migrate_hook.go rename to client/allocrunner/migrate_hook.go index c442b693c..ca9efcc25 100644 --- a/client/allocrunnerv2/migrate_hook.go +++ b/client/allocrunner/migrate_hook.go @@ -1,4 +1,4 @@ -package allocrunnerv2 +package allocrunner import ( "context" diff --git a/client/allocrunnerv2/state/state.go b/client/allocrunner/state/state.go similarity index 100% rename from client/allocrunnerv2/state/state.go rename to client/allocrunner/state/state.go diff --git a/client/allocrunnerv2/taskrunner/artifact_hook.go b/client/allocrunner/taskrunner/artifact_hook.go similarity index 86% rename from client/allocrunnerv2/taskrunner/artifact_hook.go rename to client/allocrunner/taskrunner/artifact_hook.go index de8e1703e..6860a7a8b 100644 --- a/client/allocrunnerv2/taskrunner/artifact_hook.go +++ b/client/allocrunner/taskrunner/artifact_hook.go @@ -5,9 +5,9 @@ import ( "fmt" log "github.com/hashicorp/go-hclog" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/getter" - ti "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/getter" + ti "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/nomad/structs" ) diff --git a/client/allocrunnerv2/taskrunner/errors.go b/client/allocrunner/taskrunner/errors.go similarity index 100% rename from client/allocrunnerv2/taskrunner/errors.go rename to client/allocrunner/taskrunner/errors.go diff --git a/client/allocrunnerv2/taskrunner/getter/getter.go b/client/allocrunner/taskrunner/getter/getter.go similarity index 100% rename from client/allocrunnerv2/taskrunner/getter/getter.go rename to client/allocrunner/taskrunner/getter/getter.go diff --git a/client/allocrunnerv2/taskrunner/getter/getter_test.go b/client/allocrunner/taskrunner/getter/getter_test.go similarity index 100% rename from client/allocrunnerv2/taskrunner/getter/getter_test.go rename to client/allocrunner/taskrunner/getter/getter_test.go diff --git a/client/allocrunnerv2/taskrunner/getter/test-fixtures/archive.tar.gz b/client/allocrunner/taskrunner/getter/test-fixtures/archive.tar.gz similarity index 100% rename from client/allocrunnerv2/taskrunner/getter/test-fixtures/archive.tar.gz rename to client/allocrunner/taskrunner/getter/test-fixtures/archive.tar.gz diff --git a/client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/exist/my.config b/client/allocrunner/taskrunner/getter/test-fixtures/archive/exist/my.config similarity index 100% rename from client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/exist/my.config rename to client/allocrunner/taskrunner/getter/test-fixtures/archive/exist/my.config diff --git a/client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/new/my.config b/client/allocrunner/taskrunner/getter/test-fixtures/archive/new/my.config similarity index 100% rename from client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/new/my.config rename to client/allocrunner/taskrunner/getter/test-fixtures/archive/new/my.config diff --git a/client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/test.sh b/client/allocrunner/taskrunner/getter/test-fixtures/archive/test.sh similarity index 100% rename from client/allocrunnerv2/taskrunner/getter/test-fixtures/archive/test.sh rename to client/allocrunner/taskrunner/getter/test-fixtures/archive/test.sh diff --git a/client/allocrunnerv2/taskrunner/getter/test-fixtures/test.sh b/client/allocrunner/taskrunner/getter/test-fixtures/test.sh similarity index 100% rename from client/allocrunnerv2/taskrunner/getter/test-fixtures/test.sh rename to client/allocrunner/taskrunner/getter/test-fixtures/test.sh diff --git a/client/allocrunnerv2/taskrunner/handleproxy.go b/client/allocrunner/taskrunner/handleproxy.go similarity index 100% rename from client/allocrunnerv2/taskrunner/handleproxy.go rename to client/allocrunner/taskrunner/handleproxy.go diff --git a/client/allocrunnerv2/taskrunner/handleproxy_test.go b/client/allocrunner/taskrunner/handleproxy_test.go similarity index 100% rename from client/allocrunnerv2/taskrunner/handleproxy_test.go rename to client/allocrunner/taskrunner/handleproxy_test.go diff --git a/client/allocrunnerv2/taskrunner/interfaces/events.go b/client/allocrunner/taskrunner/interfaces/events.go similarity index 100% rename from client/allocrunnerv2/taskrunner/interfaces/events.go rename to client/allocrunner/taskrunner/interfaces/events.go diff --git a/client/allocrunnerv2/taskrunner/interfaces/lifecycle.go b/client/allocrunner/taskrunner/interfaces/lifecycle.go similarity index 100% rename from client/allocrunnerv2/taskrunner/interfaces/lifecycle.go rename to client/allocrunner/taskrunner/interfaces/lifecycle.go diff --git a/client/allocrunnerv2/taskrunner/lifecycle.go b/client/allocrunner/taskrunner/lifecycle.go similarity index 100% rename from client/allocrunnerv2/taskrunner/lifecycle.go rename to client/allocrunner/taskrunner/lifecycle.go diff --git a/client/allocrunnerv2/taskrunner/logmon_hook.go b/client/allocrunner/taskrunner/logmon_hook.go similarity index 97% rename from client/allocrunnerv2/taskrunner/logmon_hook.go rename to client/allocrunner/taskrunner/logmon_hook.go index 2c47142ca..dad193a93 100644 --- a/client/allocrunnerv2/taskrunner/logmon_hook.go +++ b/client/allocrunner/taskrunner/logmon_hook.go @@ -8,7 +8,7 @@ import ( hclog "github.com/hashicorp/go-hclog" plugin "github.com/hashicorp/go-plugin" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" "github.com/hashicorp/nomad/client/logmon" "github.com/hashicorp/nomad/helper/uuid" ) diff --git a/client/allocrunnerv2/taskrunner/logmon_hook_test.go b/client/allocrunner/taskrunner/logmon_hook_test.go similarity index 74% rename from client/allocrunnerv2/taskrunner/logmon_hook_test.go rename to client/allocrunner/taskrunner/logmon_hook_test.go index b7df170e9..075397d79 100644 --- a/client/allocrunnerv2/taskrunner/logmon_hook_test.go +++ b/client/allocrunner/taskrunner/logmon_hook_test.go @@ -1,6 +1,6 @@ package taskrunner -import "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" +import "github.com/hashicorp/nomad/client/allocrunner/interfaces" // Statically assert the logmon hook implements the expected interfaces var _ interfaces.TaskPrestartHook = (*logmonHook)(nil) diff --git a/client/allocrunnerv2/taskrunner/restarts/restarts.go b/client/allocrunner/taskrunner/restarts/restarts.go similarity index 100% rename from client/allocrunnerv2/taskrunner/restarts/restarts.go rename to client/allocrunner/taskrunner/restarts/restarts.go diff --git a/client/allocrunnerv2/taskrunner/restarts/restarts_test.go b/client/allocrunner/taskrunner/restarts/restarts_test.go similarity index 100% rename from client/allocrunnerv2/taskrunner/restarts/restarts_test.go rename to client/allocrunner/taskrunner/restarts/restarts_test.go diff --git a/client/allocrunnerv2/taskrunner/service_hook.go b/client/allocrunner/taskrunner/service_hook.go similarity index 98% rename from client/allocrunnerv2/taskrunner/service_hook.go rename to client/allocrunner/taskrunner/service_hook.go index 28ed0c8aa..7fdbba50a 100644 --- a/client/allocrunnerv2/taskrunner/service_hook.go +++ b/client/allocrunner/taskrunner/service_hook.go @@ -6,7 +6,7 @@ import ( "sync" log "github.com/hashicorp/go-hclog" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" diff --git a/client/allocrunnerv2/taskrunner/shutdown_delay_hook.go b/client/allocrunner/taskrunner/shutdown_delay_hook.go similarity index 92% rename from client/allocrunnerv2/taskrunner/shutdown_delay_hook.go rename to client/allocrunner/taskrunner/shutdown_delay_hook.go index 4bad97136..15f76ebd1 100644 --- a/client/allocrunnerv2/taskrunner/shutdown_delay_hook.go +++ b/client/allocrunner/taskrunner/shutdown_delay_hook.go @@ -5,7 +5,7 @@ import ( "time" log "github.com/hashicorp/go-hclog" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" ) // shutdownDelayHook delays shutting down a task between deregistering it from diff --git a/client/allocrunnerv2/taskrunner/state/state.go b/client/allocrunner/taskrunner/state/state.go similarity index 100% rename from client/allocrunnerv2/taskrunner/state/state.go rename to client/allocrunner/taskrunner/state/state.go diff --git a/client/allocrunnerv2/taskrunner/stats_hook.go b/client/allocrunner/taskrunner/stats_hook.go similarity index 97% rename from client/allocrunnerv2/taskrunner/stats_hook.go rename to client/allocrunner/taskrunner/stats_hook.go index 46dbee926..84cbc62aa 100644 --- a/client/allocrunnerv2/taskrunner/stats_hook.go +++ b/client/allocrunner/taskrunner/stats_hook.go @@ -7,7 +7,7 @@ import ( "time" hclog "github.com/hashicorp/go-hclog" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" "github.com/hashicorp/nomad/client/driver" cstructs "github.com/hashicorp/nomad/client/structs" ) diff --git a/client/allocrunnerv2/taskrunner/stats_hook_test.go b/client/allocrunner/taskrunner/stats_hook_test.go similarity index 98% rename from client/allocrunnerv2/taskrunner/stats_hook_test.go rename to client/allocrunner/taskrunner/stats_hook_test.go index 98c9854c3..180684299 100644 --- a/client/allocrunnerv2/taskrunner/stats_hook_test.go +++ b/client/allocrunner/taskrunner/stats_hook_test.go @@ -5,7 +5,7 @@ import ( "testing" "time" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" "github.com/hashicorp/nomad/client/driver" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/helper/testlog" diff --git a/client/allocrunnerv2/taskrunner/task_dir_hook.go b/client/allocrunner/taskrunner/task_dir_hook.go similarity index 95% rename from client/allocrunnerv2/taskrunner/task_dir_hook.go rename to client/allocrunner/taskrunner/task_dir_hook.go index 0f843eec0..4053f54ad 100644 --- a/client/allocrunnerv2/taskrunner/task_dir_hook.go +++ b/client/allocrunner/taskrunner/task_dir_hook.go @@ -4,7 +4,7 @@ import ( "context" log "github.com/hashicorp/go-hclog" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" cconfig "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/nomad/structs" diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go similarity index 99% rename from client/allocrunnerv2/taskrunner/task_runner.go rename to client/allocrunner/taskrunner/task_runner.go index 1171e8aaa..e93456873 100644 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -9,9 +9,9 @@ import ( metrics "github.com/armon/go-metrics" log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/restarts" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/state" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/driver" diff --git a/client/allocrunnerv2/taskrunner/task_runner_getters.go b/client/allocrunner/taskrunner/task_runner_getters.go similarity index 100% rename from client/allocrunnerv2/taskrunner/task_runner_getters.go rename to client/allocrunner/taskrunner/task_runner_getters.go diff --git a/client/allocrunnerv2/taskrunner/task_runner_hooks.go b/client/allocrunner/taskrunner/task_runner_hooks.go similarity index 98% rename from client/allocrunnerv2/taskrunner/task_runner_hooks.go rename to client/allocrunner/taskrunner/task_runner_hooks.go index 14caba832..7d9b4196f 100644 --- a/client/allocrunnerv2/taskrunner/task_runner_hooks.go +++ b/client/allocrunner/taskrunner/task_runner_hooks.go @@ -6,8 +6,8 @@ import ( "time" multierror "github.com/hashicorp/go-multierror" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/state" "github.com/hashicorp/nomad/nomad/structs" ) diff --git a/client/allocrunnerv2/taskrunner/template/template.go b/client/allocrunner/taskrunner/template/template.go similarity index 99% rename from client/allocrunnerv2/taskrunner/template/template.go rename to client/allocrunner/taskrunner/template/template.go index 0685d1459..0d90306ec 100644 --- a/client/allocrunnerv2/taskrunner/template/template.go +++ b/client/allocrunner/taskrunner/template/template.go @@ -17,7 +17,7 @@ import ( "github.com/hashicorp/consul-template/signals" envparse "github.com/hashicorp/go-envparse" multierror "github.com/hashicorp/go-multierror" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver/env" "github.com/hashicorp/nomad/helper" diff --git a/client/allocrunnerv2/taskrunner/template/template_test.go b/client/allocrunner/taskrunner/template/template_test.go similarity index 100% rename from client/allocrunnerv2/taskrunner/template/template_test.go rename to client/allocrunner/taskrunner/template/template_test.go diff --git a/client/allocrunnerv2/taskrunner/template_hook.go b/client/allocrunner/taskrunner/template_hook.go similarity index 94% rename from client/allocrunnerv2/taskrunner/template_hook.go rename to client/allocrunner/taskrunner/template_hook.go index 4b3073d7c..3b13443fc 100644 --- a/client/allocrunnerv2/taskrunner/template_hook.go +++ b/client/allocrunner/taskrunner/template_hook.go @@ -6,9 +6,9 @@ import ( "sync" log "github.com/hashicorp/go-hclog" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - ti "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/template" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" + ti "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/template" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver/env" "github.com/hashicorp/nomad/nomad/structs" diff --git a/client/allocrunnerv2/taskrunner/validate_hook.go b/client/allocrunner/taskrunner/validate_hook.go similarity index 96% rename from client/allocrunnerv2/taskrunner/validate_hook.go rename to client/allocrunner/taskrunner/validate_hook.go index c17544875..150604802 100644 --- a/client/allocrunnerv2/taskrunner/validate_hook.go +++ b/client/allocrunner/taskrunner/validate_hook.go @@ -6,7 +6,7 @@ import ( log "github.com/hashicorp/go-hclog" multierror "github.com/hashicorp/go-multierror" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver/env" "github.com/hashicorp/nomad/nomad/structs" diff --git a/client/allocrunnerv2/taskrunner/validate_hook_test.go b/client/allocrunner/taskrunner/validate_hook_test.go similarity index 100% rename from client/allocrunnerv2/taskrunner/validate_hook_test.go rename to client/allocrunner/taskrunner/validate_hook_test.go diff --git a/client/allocrunnerv2/taskrunner/vault_hook.go b/client/allocrunner/taskrunner/vault_hook.go similarity index 98% rename from client/allocrunnerv2/taskrunner/vault_hook.go rename to client/allocrunner/taskrunner/vault_hook.go index 479479e71..1329e6cfe 100644 --- a/client/allocrunnerv2/taskrunner/vault_hook.go +++ b/client/allocrunner/taskrunner/vault_hook.go @@ -13,8 +13,8 @@ import ( log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - ti "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" + ti "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" ) diff --git a/client/allocrunner/util.go b/client/allocrunner/util.go new file mode 100644 index 000000000..cc0b24938 --- /dev/null +++ b/client/allocrunner/util.go @@ -0,0 +1 @@ +package allocrunner diff --git a/client/allocrunnerdeprecated/alloc_runner.go b/client/allocrunnerdeprecated/alloc_runner.go index 0fb952fe6..f5636802b 100644 --- a/client/allocrunnerdeprecated/alloc_runner.go +++ b/client/allocrunnerdeprecated/alloc_runner.go @@ -201,7 +201,7 @@ func (r *AllocRunner) pre060StateFilePath() string { // RestoreState is used to restore the state of the alloc runner func (r *AllocRunner) RestoreState() error { - //XXX Deprecated: see allocrunnerv2 + //XXX Deprecated: see allocrunner //err := r.stateDB.View(func(tx *bolt.Tx) error { // bkt, err := state.GetAllocationBucket(tx, r.allocID) // if err != nil { @@ -346,7 +346,7 @@ func (r *AllocRunner) saveAllocRunnerState() error { return nil } - //XXX Deprecated: see allocrunnerv2 + //XXX Deprecated: see allocrunner return nil //// Grab all the relevant data @@ -443,7 +443,7 @@ func (r *AllocRunner) DestroyState() error { // } // return nil //}) - panic("deprecated: use allocrunnerv2") + panic("deprecated: use allocrunner") } // DestroyContext is used to destroy the context diff --git a/client/allocrunnerdeprecated/taskrunner/task_runner.go b/client/allocrunnerdeprecated/taskrunner/task_runner.go index d8b58e480..e34903532 100644 --- a/client/allocrunnerdeprecated/taskrunner/task_runner.go +++ b/client/allocrunnerdeprecated/taskrunner/task_runner.go @@ -23,8 +23,8 @@ import ( "github.com/hashicorp/go-multierror" version "github.com/hashicorp/go-version" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/getter" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/restarts" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/getter" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" "github.com/hashicorp/nomad/client/config" consulApi "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/driver" @@ -354,7 +354,7 @@ func (r *TaskRunner) pre060StateFilePath() string { // executor. func (r *TaskRunner) RestoreState() (string, error) { var snap taskRunnerState - //XXX Deprecated: see allocrunnerv2 + //XXX Deprecated: see allocrunner //err := r.stateDB.View(func(tx *bolt.Tx) error { // bkt, err := state.GetTaskBucket(tx, r.alloc.ID, r.task.Name) // if err != nil { @@ -513,7 +513,7 @@ func (r *TaskRunner) SaveState() error { } // Start the transaction. - //XXX Deprecated: see allocrunnerv2 + //XXX Deprecated: see allocrunner return nil //return r.stateDB.Batch(func(tx *bolt.Tx) error { // // Grab the task bucket @@ -546,7 +546,7 @@ func (r *TaskRunner) DestroyState() error { // } // return nil //}) - //XXX Deprecated: see allocrunnerv2 + //XXX Deprecated: see allocrunner panic("deprecated") } @@ -1508,7 +1508,7 @@ func (r *TaskRunner) registerServices(d driver.Driver, h driver.DriverHandle, n //} //interpolatedTask := interpolateServices(r.envBuilder.Build(), r.task) //taskServices := consul.NewTaskServices(r.alloc, interpolatedTask, r, exec, n) - panic("XXX broken during transition to allocrunnerv2") + panic("XXX broken during transition to allocrunner") return r.consul.RegisterTask(nil) } @@ -1723,7 +1723,7 @@ func (r *TaskRunner) updateServices(d driver.Driver, h driver.ScriptExecutor, //r.driverNetLock.Unlock() //oldTaskServices := consul.NewTaskServices(oldAlloc, oldTask, r, exec, net) //newTaskServices := consul.NewTaskServices(newAlloc, newTask, r, exec, net) - panic("XXX broken during transition to allocrunnerv2") + panic("XXX broken during transition to allocrunner") //return r.consul.UpdateTask(oldTaskServices, newTaskServices) return r.consul.UpdateTask(nil, nil) } @@ -1732,7 +1732,7 @@ func (r *TaskRunner) updateServices(d driver.Driver, h driver.ScriptExecutor, // Canary=true and Canary=false versions in case Canary=false is set at the // same time as the alloc is stopped. func (r *TaskRunner) removeServices() { - panic("XXX broken during transition to allocrunnerv2") + panic("XXX broken during transition to allocrunner") //interpTask := interpolateServices(r.envBuilder.Build(), r.task) //taskServices := consul.NewTaskServices(r.alloc, interpTask, r, nil, nil) //r.consul.RemoveTask(taskServices) diff --git a/client/allocrunnerdeprecated/taskrunner/task_runner_test.go b/client/allocrunnerdeprecated/taskrunner/task_runner_test.go index 50dcc3ad3..8bd03976b 100644 --- a/client/allocrunnerdeprecated/taskrunner/task_runner_test.go +++ b/client/allocrunnerdeprecated/taskrunner/task_runner_test.go @@ -18,7 +18,7 @@ import ( "github.com/boltdb/bolt" "github.com/golang/snappy" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/restarts" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" "github.com/hashicorp/nomad/client/config" consulApi "github.com/hashicorp/nomad/client/consul" "github.com/hashicorp/nomad/client/driver/env" @@ -666,7 +666,7 @@ func TestTaskRunner_UnregisterConsul_Retries(t *testing.T) { } } -//XXX Ported to allocrunnerv2/task_runner/validate_hook_test.go +//XXX Ported to allocrunner/task_runner/validate_hook_test.go func TestTaskRunner_Validate_UserEnforcement(t *testing.T) { t.Parallel() ctx := testTaskRunner(t, false) diff --git a/client/allocrunnerv2/util.go b/client/allocrunnerv2/util.go deleted file mode 100644 index b833b6c2d..000000000 --- a/client/allocrunnerv2/util.go +++ /dev/null @@ -1 +0,0 @@ -package allocrunnerv2 diff --git a/client/client.go b/client/client.go index 837708bb9..4f4b6f7e6 100644 --- a/client/client.go +++ b/client/client.go @@ -18,8 +18,8 @@ import ( consulapi "github.com/hashicorp/consul/api" hclog "github.com/hashicorp/go-hclog" multierror "github.com/hashicorp/go-multierror" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - arstate "github.com/hashicorp/nomad/client/allocrunnerv2/state" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" + arstate "github.com/hashicorp/nomad/client/allocrunner/state" consulApi "github.com/hashicorp/nomad/client/consul" cstructs "github.com/hashicorp/nomad/client/structs" hstats "github.com/hashicorp/nomad/helper/stats" @@ -28,7 +28,7 @@ import ( "github.com/hashicorp/consul/lib" "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunnerv2" + "github.com/hashicorp/nomad/client/allocrunner" "github.com/hashicorp/nomad/client/allocwatcher" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/servers" @@ -761,7 +761,7 @@ func (c *Client) restoreState() error { for _, alloc := range allocs { c.configLock.RLock() - arConf := &allocrunnerv2.Config{ + arConf := &allocrunner.Config{ Alloc: alloc, Logger: c.logger, ClientConfig: c.config, @@ -772,7 +772,7 @@ func (c *Client) restoreState() error { } c.configLock.RUnlock() - ar, err := allocrunnerv2.NewAllocRunner(arConf) + ar, err := allocrunner.NewAllocRunner(arConf) if err != nil { c.logger.Error("error running alloc", "error", err, "alloc_id", alloc.ID) mErr.Errors = append(mErr.Errors, err) @@ -1931,7 +1931,7 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error // The long term fix is to pass in the config and node separately and then // we don't have to do a copy. c.configLock.RLock() - arConf := &allocrunnerv2.Config{ + arConf := &allocrunner.Config{ Alloc: alloc, Logger: c.logger, ClientConfig: c.config, @@ -1943,7 +1943,7 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error } c.configLock.RUnlock() - ar, err := allocrunnerv2.NewAllocRunner(arConf) + ar, err := allocrunner.NewAllocRunner(arConf) if err != nil { return err } diff --git a/client/state/interface.go b/client/state/interface.go index 952620422..6f571659c 100644 --- a/client/state/interface.go +++ b/client/state/interface.go @@ -1,7 +1,7 @@ package state import ( - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/state" "github.com/hashicorp/nomad/nomad/structs" ) diff --git a/client/state/noopdb.go b/client/state/noopdb.go index a98461e88..05ee2b562 100644 --- a/client/state/noopdb.go +++ b/client/state/noopdb.go @@ -1,7 +1,7 @@ package state import ( - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/state" "github.com/hashicorp/nomad/nomad/structs" ) diff --git a/client/state/state_database.go b/client/state/state_database.go index ba90da2d0..b63355b76 100644 --- a/client/state/state_database.go +++ b/client/state/state_database.go @@ -4,7 +4,7 @@ import ( "fmt" "path/filepath" - trstate "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + trstate "github.com/hashicorp/nomad/client/allocrunner/taskrunner/state" "github.com/hashicorp/nomad/helper/boltdd" "github.com/hashicorp/nomad/nomad/structs" ) diff --git a/command/alloc_status.go b/command/alloc_status.go index c4373fd7c..773791c59 100644 --- a/command/alloc_status.go +++ b/command/alloc_status.go @@ -12,7 +12,7 @@ import ( "github.com/hashicorp/nomad/api" "github.com/hashicorp/nomad/api/contexts" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/restarts" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" "github.com/posener/complete" ) From ad4fac526cd7b116a1b6ae88c8050a74e64cf04f Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Thu, 4 Oct 2018 16:32:14 -0700 Subject: [PATCH 116/152] more test fixes --- client/driver/logging/rotator_test.go | 18 +++++++++--------- .../driver/logging/syslog_parser_unix_test.go | 4 ++-- .../driver/logging/syslog_server_unix_test.go | 4 +++- client/testing.go | 1 + 4 files changed, 15 insertions(+), 12 deletions(-) diff --git a/client/driver/logging/rotator_test.go b/client/driver/logging/rotator_test.go index d988c6493..8f06cec70 100644 --- a/client/driver/logging/rotator_test.go +++ b/client/driver/logging/rotator_test.go @@ -19,7 +19,7 @@ var ( func TestFileRotator_IncorrectPath(t *testing.T) { t.Parallel() - if _, err := NewFileRotator("/foo", baseFileName, 10, 10, testlog.Logger(t)); err == nil { + if _, err := NewFileRotator("/foo", baseFileName, 10, 10, testlog.HCLogger(t)); err == nil { t.Fatalf("expected error") } } @@ -33,7 +33,7 @@ func TestFileRotator_CreateNewFile(t *testing.T) { } defer os.RemoveAll(path) - _, err = NewFileRotator(path, baseFileName, 10, 10, testlog.Logger(t)) + _, err = NewFileRotator(path, baseFileName, 10, 10, testlog.HCLogger(t)) if err != nil { t.Fatalf("test setup err: %v", err) } @@ -61,7 +61,7 @@ func TestFileRotator_OpenLastFile(t *testing.T) { t.Fatalf("test setup failure: %v", err) } - fr, err := NewFileRotator(path, baseFileName, 10, 10, testlog.Logger(t)) + fr, err := NewFileRotator(path, baseFileName, 10, 10, testlog.HCLogger(t)) if err != nil { t.Fatalf("test setup err: %v", err) } @@ -85,7 +85,7 @@ func TestFileRotator_WriteToCurrentFile(t *testing.T) { t.Fatalf("test setup failure: %v", err) } - fr, err := NewFileRotator(path, baseFileName, 10, 5, testlog.Logger(t)) + fr, err := NewFileRotator(path, baseFileName, 10, 5, testlog.HCLogger(t)) if err != nil { t.Fatalf("test setup err: %v", err) } @@ -118,7 +118,7 @@ func TestFileRotator_RotateFiles(t *testing.T) { } defer os.RemoveAll(path) - fr, err := NewFileRotator(path, baseFileName, 10, 5, testlog.Logger(t)) + fr, err := NewFileRotator(path, baseFileName, 10, 5, testlog.HCLogger(t)) if err != nil { t.Fatalf("test setup err: %v", err) } @@ -177,7 +177,7 @@ func TestFileRotator_RotateFiles_Boundary(t *testing.T) { } defer os.RemoveAll(path) - fr, err := NewFileRotator(path, baseFileName, 10, 5, testlog.Logger(t)) + fr, err := NewFileRotator(path, baseFileName, 10, 5, testlog.HCLogger(t)) if err != nil { t.Fatalf("test setup err: %v", err) } @@ -243,7 +243,7 @@ func TestFileRotator_WriteRemaining(t *testing.T) { t.Fatalf("test setup failure: %v", err) } - fr, err := NewFileRotator(path, baseFileName, 10, 5, testlog.Logger(t)) + fr, err := NewFileRotator(path, baseFileName, 10, 5, testlog.HCLogger(t)) if err != nil { t.Fatalf("test setup err: %v", err) } @@ -316,7 +316,7 @@ func TestFileRotator_PurgeOldFiles(t *testing.T) { } defer os.RemoveAll(path) - fr, err := NewFileRotator(path, baseFileName, 2, 2, testlog.Logger(t)) + fr, err := NewFileRotator(path, baseFileName, 2, 2, testlog.HCLogger(t)) if err != nil { t.Fatalf("test setup err: %v", err) } @@ -366,7 +366,7 @@ func benchmarkRotatorWithInputSize(size int, b *testing.B) { } defer os.RemoveAll(path) - fr, err := NewFileRotator(path, baseFileName, 5, 1024*1024, testlog.Logger(b)) + fr, err := NewFileRotator(path, baseFileName, 5, 1024*1024, testlog.HCLogger(b)) if err != nil { b.Fatalf("test setup err: %v", err) } diff --git a/client/driver/logging/syslog_parser_unix_test.go b/client/driver/logging/syslog_parser_unix_test.go index 38d0b6b8a..ebb337fb1 100644 --- a/client/driver/logging/syslog_parser_unix_test.go +++ b/client/driver/logging/syslog_parser_unix_test.go @@ -13,7 +13,7 @@ import ( func TestLogParser_Priority(t *testing.T) { t.Parallel() line := []byte("<30>2016-02-10T10:16:43-08:00 d-thinkpad docker/e2a1e3ebd3a3[22950]: 1:C 10 Feb 18:16:43.391 # Warning: no config file specified, using the default config. In order to specify a config file use redis-server /path/to/redis.conf") - d := NewDockerLogParser(testlog.Logger(t)) + d := NewDockerLogParser(testlog.HCLogger(t)) p, _, err := d.parsePriority(line) if err != nil { t.Fatalf("got an err: %v", err) @@ -32,7 +32,7 @@ func TestLogParser_Priority(t *testing.T) { func TestLogParser_Priority_UnixFormatter(t *testing.T) { t.Parallel() line := []byte("<30>Feb 6, 10:16:43 docker/e2a1e3ebd3a3[22950]: 1:C 10 Feb 18:16:43.391 # Warning: no config file specified, using the default config. In order to specify a config file use redis-server /path/to/redis.conf") - d := NewDockerLogParser(testlog.Logger(t)) + d := NewDockerLogParser(testlog.HCLogger(t)) p, _, err := d.parsePriority(line) if err != nil { t.Fatalf("got an err: %v", err) diff --git a/client/driver/logging/syslog_server_unix_test.go b/client/driver/logging/syslog_server_unix_test.go index 2540d9100..d71e32ee9 100644 --- a/client/driver/logging/syslog_server_unix_test.go +++ b/client/driver/logging/syslog_server_unix_test.go @@ -7,6 +7,8 @@ import ( "path" "testing" "time" + + "github.com/hashicorp/nomad/helper/testlog" ) func TestSyslogServer_Start_Shutdown(t *testing.T) { @@ -24,7 +26,7 @@ func TestSyslogServer_Start_Shutdown(t *testing.T) { t.Fatalf("Failed to listen unix socket: %v", err) } - s := NewSyslogServer(l, make(chan *SyslogMessage, 2048), nil) + s := NewSyslogServer(l, make(chan *SyslogMessage, 2048), testlog.HCLogger(t)) go s.Start() if s.done { diff --git a/client/testing.go b/client/testing.go index bfeef18e4..35d17b8d9 100644 --- a/client/testing.go +++ b/client/testing.go @@ -23,6 +23,7 @@ func TestClient(t testing.T, cb func(c *config.Config)) *Client { conf.Options[fingerprint.TightenNetworkTimeoutsConfig] = "true" logger := testlog.HCLogger(t) + conf.Logger = logger // Set the plugin loaders conf.PluginLoader = catalog.TestPluginLoader(t) From 89dafaaea9b480967b3885e20ca23da217d6e74d Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Thu, 4 Oct 2018 16:41:40 -0700 Subject: [PATCH 117/152] compile on windows --- client/driver/executor/executor_windows.go | 6 +++--- client/fingerprint/network_windows.go | 6 +++--- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/client/driver/executor/executor_windows.go b/client/driver/executor/executor_windows.go index 3b508184f..f7e05b8fa 100644 --- a/client/driver/executor/executor_windows.go +++ b/client/driver/executor/executor_windows.go @@ -9,10 +9,10 @@ import ( // configure new process group for child process func (e *UniversalExecutor) setNewProcessGroup() error { // We need to check that as build flags includes windows for this file - if e.cmd.SysProcAttr == nil { - e.cmd.SysProcAttr = &syscall.SysProcAttr{} + if e.childCmd.SysProcAttr == nil { + e.childCmd.SysProcAttr = &syscall.SysProcAttr{} } - e.cmd.SysProcAttr.CreationFlags = syscall.CREATE_NEW_PROCESS_GROUP + e.childCmd.SysProcAttr.CreationFlags = syscall.CREATE_NEW_PROCESS_GROUP return nil } diff --git a/client/fingerprint/network_windows.go b/client/fingerprint/network_windows.go index 8617a65a8..e71ce0ce9 100644 --- a/client/fingerprint/network_windows.go +++ b/client/fingerprint/network_windows.go @@ -14,7 +14,7 @@ func (f *NetworkFingerprint) linkSpeed(device string) int { outBytes, err := exec.Command(path, command).Output() if err != nil { - f.logger.Printf("[WARN] fingerprint.network: Error calling %s (%s): %v", path, command, err) + f.logger.Warn("failed to detect link speed", "path", path, "command", command, "error", err) return 0 } @@ -26,14 +26,14 @@ func (f *NetworkFingerprint) linkSpeed(device string) int { func (f *NetworkFingerprint) parseLinkSpeed(commandOutput string) int { args := strings.Split(commandOutput, " ") if len(args) != 2 { - f.logger.Printf("[WARN] fingerprint.network: Couldn't split LinkSpeed (%s)", commandOutput) + f.logger.Warn("couldn't split LinkSpeed output", "output", commandOutput) return 0 } unit := strings.Replace(args[1], "\r\n", "", -1) value, err := strconv.Atoi(args[0]) if err != nil { - f.logger.Printf("[WARN] fingerprint.network: Unable to parse LinkSpeed value (%s)", commandOutput) + f.logger.Warn("unable to parse LinkSpeed value", "value", commandOutput) return 0 } From 7946a14aa8779b8db30f706d2199e648df0d7a7e Mon Sep 17 00:00:00 2001 From: Alex Dadgar Date: Thu, 4 Oct 2018 19:36:40 -0700 Subject: [PATCH 118/152] Fix lints --- client/allocrunner/alloc_runner_hooks.go | 2 +- .../taskrunner/handleproxy_test.go | 2 +- client/allocrunner/taskrunner/task_runner.go | 6 +- .../taskrunner/task_runner_hooks.go | 2 +- client/allocrunnerdeprecated/alloc_runner.go | 36 ++++----- .../taskrunner/task_runner.go | 6 +- client/allocwatcher/alloc_watcher_test.go | 2 +- client/fingerprint/cgroup_test.go | 8 +- command/agent/consul/check_watcher_test.go | 12 ++- command/agent/consul/int_test.go | 76 +++++++++++-------- command/agent/consul/unit_test.go | 3 +- drivers/rawexec/driver_test.go | 2 + drivers/rawexec/handle.go | 2 +- plugins/drivers/driver.go | 4 +- 14 files changed, 92 insertions(+), 71 deletions(-) diff --git a/client/allocrunner/alloc_runner_hooks.go b/client/allocrunner/alloc_runner_hooks.go index db0d9a506..02372f9a8 100644 --- a/client/allocrunner/alloc_runner_hooks.go +++ b/client/allocrunner/alloc_runner_hooks.go @@ -76,7 +76,7 @@ func (ar *allocRunner) initRunnerHooks() { hs := &allocHealthSetter{ar} // Create the alloc directory hook. This is run first to ensure the - // directoy path exists for other hooks. + // directory path exists for other hooks. ar.runnerHooks = []interfaces.RunnerHook{ newAllocDirHook(hookLogger, ar.allocDir), newDiskMigrationHook(hookLogger, ar.prevAllocWatcher, ar.allocDir), diff --git a/client/allocrunner/taskrunner/handleproxy_test.go b/client/allocrunner/taskrunner/handleproxy_test.go index 5231d5297..b43029407 100644 --- a/client/allocrunner/taskrunner/handleproxy_test.go +++ b/client/allocrunner/taskrunner/handleproxy_test.go @@ -20,7 +20,7 @@ func TestHandleResult_Wait_Result(t *testing.T) { outCh1 := make(chan *structs.WaitResult) outCh2 := make(chan *structs.WaitResult) - // Create two recievers + // Create two receivers go func() { outCh1 <- h.Wait(context.Background()) }() diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index e93456873..3f93313ca 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -69,10 +69,6 @@ type TaskRunner struct { // stateDB is for persisting localState and taskState stateDB cstate.StateDB - // persistedHash is the hash of the last persisted state for skipping - // unnecessary writes - persistedHash []byte - // ctx is the task runner's context representing the tasks's lifecycle. // Canceling the context will cause the task to be destroyed. ctx context.Context @@ -632,7 +628,7 @@ func (tr *TaskRunner) appendEvent(event *structs.TaskEvent) error { // Ensure the event is populated with human readable strings event.PopulateEventDisplayMessage() - // Propogate failure from event to task state + // Propagate failure from event to task state if event.FailsTask { tr.state.Failed = true } diff --git a/client/allocrunner/taskrunner/task_runner_hooks.go b/client/allocrunner/taskrunner/task_runner_hooks.go index 7d9b4196f..7ad89f64b 100644 --- a/client/allocrunner/taskrunner/task_runner_hooks.go +++ b/client/allocrunner/taskrunner/task_runner_hooks.go @@ -19,7 +19,7 @@ func (tr *TaskRunner) initHooks() { tr.logmonHookConfig = newLogMonHookConfig(task.Name, tr.taskDir.LogDir) // Create the task directory hook. This is run first to ensure the - // directoy path exists for other hooks. + // directory path exists for other hooks. tr.runnerHooks = []interfaces.TaskHook{ newValidateHook(tr.clientConfig, hookLogger), newTaskDirHook(tr, hookLogger), diff --git a/client/allocrunnerdeprecated/alloc_runner.go b/client/allocrunnerdeprecated/alloc_runner.go index f5636802b..69912380d 100644 --- a/client/allocrunnerdeprecated/alloc_runner.go +++ b/client/allocrunnerdeprecated/alloc_runner.go @@ -26,11 +26,11 @@ import ( ) var ( - // The following are the key paths written to the state database - allocRunnerStateAllocKey = []byte("alloc") - allocRunnerStateImmutableKey = []byte("immutable") - allocRunnerStateMutableKey = []byte("mutable") - allocRunnerStateAllocDirKey = []byte("alloc-dir") +// The following are the key paths written to the state database +//allocRunnerStateAllocKey = []byte("alloc") +//allocRunnerStateImmutableKey = []byte("immutable") +//allocRunnerStateMutableKey = []byte("mutable") +//allocRunnerStateAllocDirKey = []byte("alloc-dir") ) // AllocStateUpdater is used to update the status of an allocation @@ -115,23 +115,23 @@ type AllocRunner struct { // allocRunnerAllocState is state that only has to be written when the alloc // changes. -type allocRunnerAllocState struct { - Alloc *structs.Allocation -} +//type allocRunnerAllocState struct { +//Alloc *structs.Allocation +//} -// allocRunnerImmutableState is state that only has to be written once. -type allocRunnerImmutableState struct { - Version string -} +//// allocRunnerImmutableState is state that only has to be written once. +//type allocRunnerImmutableState struct { +//Version string +//} // allocRunnerMutableState is state that has to be written on each save as it // changes over the life-cycle of the alloc_runner. -type allocRunnerMutableState struct { - AllocClientStatus string - AllocClientDescription string - TaskStates map[string]*structs.TaskState - DeploymentStatus *structs.AllocDeploymentStatus -} +//type allocRunnerMutableState struct { +//AllocClientStatus string +//AllocClientDescription string +//TaskStates map[string]*structs.TaskState +//DeploymentStatus *structs.AllocDeploymentStatus +//} // NewAllocRunner is used to create a new allocation context func NewAllocRunner(logger *log.Logger, config *config.Config, stateDB *bolt.DB, updater AllocStateUpdater, diff --git a/client/allocrunnerdeprecated/taskrunner/task_runner.go b/client/allocrunnerdeprecated/taskrunner/task_runner.go index e34903532..9642591b0 100644 --- a/client/allocrunnerdeprecated/taskrunner/task_runner.go +++ b/client/allocrunnerdeprecated/taskrunner/task_runner.go @@ -64,9 +64,9 @@ const ( ) var ( - // taskRunnerStateAllKey holds all the task runners state. At the moment - // there is no need to split it - taskRunnerStateAllKey = []byte("simple-all") +// taskRunnerStateAllKey holds all the task runners state. At the moment +// there is no need to split it +//taskRunnerStateAllKey = []byte("simple-all") ) // taskRestartEvent wraps a TaskEvent with additional metadata to control diff --git a/client/allocwatcher/alloc_watcher_test.go b/client/allocwatcher/alloc_watcher_test.go index b8a3c8581..6eb7bcc3d 100644 --- a/client/allocwatcher/alloc_watcher_test.go +++ b/client/allocwatcher/alloc_watcher_test.go @@ -39,7 +39,7 @@ func newFakeAllocRunner(t *testing.T, logger hclog.Logger) *fakeAllocRunner { alloc.Job.TaskGroups[0].EphemeralDisk.Sticky = true alloc.Job.TaskGroups[0].EphemeralDisk.Migrate = true - path, err := ioutil.TempDir("", "nomad_test_wathcer") + path, err := ioutil.TempDir("", "nomad_test_watcher") require.NoError(t, err) return &fakeAllocRunner{ diff --git a/client/fingerprint/cgroup_test.go b/client/fingerprint/cgroup_test.go index b3894d2d2..5016cd855 100644 --- a/client/fingerprint/cgroup_test.go +++ b/client/fingerprint/cgroup_test.go @@ -43,7 +43,7 @@ func (m *MountPointDetectorEmptyMountPoint) MountPoint() (string, error) { func TestCGroupFingerprint(t *testing.T) { { f := &CGroupFingerprint{ - logger: testlog.Logger(t), + logger: testlog.HCLogger(t), lastState: cgroupUnavailable, mountPointDetector: &MountPointDetectorMountPointFail{}, } @@ -66,7 +66,7 @@ func TestCGroupFingerprint(t *testing.T) { { f := &CGroupFingerprint{ - logger: testlog.Logger(t), + logger: testlog.HCLogger(t), lastState: cgroupUnavailable, mountPointDetector: &MountPointDetectorValidMountPoint{}, } @@ -88,7 +88,7 @@ func TestCGroupFingerprint(t *testing.T) { { f := &CGroupFingerprint{ - logger: testlog.Logger(t), + logger: testlog.HCLogger(t), lastState: cgroupUnavailable, mountPointDetector: &MountPointDetectorEmptyMountPoint{}, } @@ -109,7 +109,7 @@ func TestCGroupFingerprint(t *testing.T) { } { f := &CGroupFingerprint{ - logger: testlog.Logger(t), + logger: testlog.HCLogger(t), lastState: cgroupAvailable, mountPointDetector: &MountPointDetectorValidMountPoint{}, } diff --git a/command/agent/consul/check_watcher_test.go b/command/agent/consul/check_watcher_test.go index cb5eb5cba..fd8d578ed 100644 --- a/command/agent/consul/check_watcher_test.go +++ b/command/agent/consul/check_watcher_test.go @@ -51,11 +51,19 @@ func newFakeCheckRestarter(w *checkWatcher, allocID, taskName, checkName string, // watching and is normally fulfilled by a TaskRunner. // // Restarts are recorded in the []restarts field and re-Watch the check. -func (c *fakeCheckRestarter) Restart(source, reason string, failure bool) { - c.restarts = append(c.restarts, checkRestartRecord{time.Now(), source, reason, failure}) +//func (c *fakeCheckRestarter) Restart(source, reason string, failure bool) { +func (c *fakeCheckRestarter) Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error { + restart := checkRestartRecord{ + timestamp: time.Now(), + source: event.Type, + reason: event.DisplayMessage, + failure: failure, + } + c.restarts = append(c.restarts, restart) // Re-Watch the check just like TaskRunner c.watcher.Watch(c.allocID, c.taskName, c.checkName, c.check, c) + return nil } // String for debugging diff --git a/command/agent/consul/int_test.go b/command/agent/consul/int_test.go index 7239d1904..c279131b9 100644 --- a/command/agent/consul/int_test.go +++ b/command/agent/consul/int_test.go @@ -1,33 +1,44 @@ package consul_test import ( + "context" "io/ioutil" "os" "path/filepath" "testing" "time" - "github.com/boltdb/bolt" consulapi "github.com/hashicorp/consul/api" "github.com/hashicorp/consul/testutil" + log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner/taskrunner" "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/command/agent/consul" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/nomad/mock" "github.com/hashicorp/nomad/nomad/structs" - "github.com/stretchr/testify/assert" + "github.com/stretchr/testify/require" ) +type mockUpdater struct { + logger log.Logger +} + +func (m *mockUpdater) TaskStateUpdated(task string, state *structs.TaskState) { + m.logger.Named("test.updater").Debug("update", "task", task, "state", state) +} + +// TODO Fix // TestConsul_Integration asserts TaskRunner properly registers and deregisters // services and checks with Consul using an embedded Consul agent. func TestConsul_Integration(t *testing.T) { if testing.Short() { t.Skip("-short set; skipping") } - assert := assert.New(t) + require := require.New(t) // Create an embedded Consul server testconsul, err := testutil.NewTestServerConfig(func(c *testutil.TestServerConfig) { @@ -61,15 +72,6 @@ func TestConsul_Integration(t *testing.T) { } defer os.RemoveAll(conf.AllocDir) - tmp, err := ioutil.TempFile("", "state-db") - if err != nil { - t.Fatalf("error creating state db file: %v", err) - } - db, err := bolt.Open(tmp.Name(), 0600, nil) - if err != nil { - t.Fatalf("error creating state db: %v", err) - } - alloc := mock.Alloc() task := alloc.Job.TaskGroups[0].Tasks[0] task.Driver = "mock_driver" @@ -121,10 +123,8 @@ func TestConsul_Integration(t *testing.T) { }, } - logger := testlog.Logger(t) - logUpdate := func(name, state string, event *structs.TaskEvent, lazySync bool) { - logger.Printf("[TEST] test.updater: name=%q state=%q event=%v", name, state, event) - } + logger := testlog.HCLogger(t) + logUpdate := &mockUpdater{logger} allocDir := allocdir.NewAllocDir(logger, filepath.Join(conf.AllocDir, alloc.ID)) if err := allocDir.Build(); err != nil { t.Fatalf("error building alloc dir: %v", err) @@ -132,7 +132,7 @@ func TestConsul_Integration(t *testing.T) { taskDir := allocDir.NewTaskDir(task.Name) vclient := vaultclient.NewMockVaultClient() consulClient, err := consulapi.NewClient(consulConfig) - assert.Nil(err) + require.Nil(err) serviceClient := consul.NewServiceClient(consulClient.Agent(), testlog.HCLogger(t), true) defer serviceClient.Shutdown() // just-in-case cleanup @@ -141,8 +141,22 @@ func TestConsul_Integration(t *testing.T) { serviceClient.Run() close(consulRan) }() - tr := taskrunner.NewTaskRunner(logger, conf, db, logUpdate, taskDir, alloc, task, vclient, serviceClient) - tr.MarkReceived() + + // Build the config + config := &taskrunner.Config{ + Alloc: alloc, + ClientConfig: conf, + Consul: serviceClient, + Task: task, + TaskDir: taskDir, + Logger: logger, + VaultClient: vclient, + StateDB: state.NoopDB{}, + StateUpdater: logUpdate, + } + + tr, err := taskrunner.NewTaskRunner(config) + require.NoError(err) go tr.Run() defer func() { // Make sure we always shutdown task runner when the test exits @@ -150,14 +164,14 @@ func TestConsul_Integration(t *testing.T) { case <-tr.WaitCh(): // Exited cleanly, no need to kill default: - tr.Kill("", "", true) // just in case + tr.Kill(context.Background(), &structs.TaskEvent{}) // just in case } }() // Block waiting for the service to appear catalog := consulClient.Catalog() res, meta, err := catalog.Service("httpd2", "test", nil) - assert.Nil(err) + require.Nil(err) for i := 0; len(res) == 0 && i < 10; i++ { //Expected initial request to fail, do a blocking query @@ -166,7 +180,7 @@ func TestConsul_Integration(t *testing.T) { t.Fatalf("error querying for service: %v", err) } } - assert.Len(res, 1) + require.Len(res, 1) // Truncate results res = res[:] @@ -174,16 +188,16 @@ func TestConsul_Integration(t *testing.T) { // Assert the service with the checks exists for i := 0; len(res) == 0 && i < 10; i++ { res, meta, err = catalog.Service("httpd", "http", &consulapi.QueryOptions{WaitIndex: meta.LastIndex + 1, WaitTime: 3 * time.Second}) - assert.Nil(err) + require.Nil(err) } - assert.Len(res, 1) + require.Len(res, 1) // Assert the script check passes (mock_driver script checks always // pass) after having time to run once time.Sleep(2 * time.Second) checks, _, err := consulClient.Health().Checks("httpd", nil) - assert.Nil(err) - assert.Len(checks, 2) + require.Nil(err) + require.Len(checks, 2) for _, check := range checks { if expected := "httpd"; check.ServiceName != expected { @@ -220,10 +234,10 @@ func TestConsul_Integration(t *testing.T) { t.Fatalf("Unexpected number of checks registered. Got %d; want 2", cnum) } - logger.Printf("[TEST] consul.test: killing task") + logger.Debug("killing task") // Kill the task - tr.Kill("", "", false) + tr.Kill(context.Background(), &structs.TaskEvent{}) select { case <-tr.WaitCh(): @@ -238,7 +252,7 @@ func TestConsul_Integration(t *testing.T) { // Ensure Consul is clean services, _, err := catalog.Services(nil) - assert.Nil(err) - assert.Len(services, 1) - assert.Contains(services, "consul") + require.Nil(err) + require.Len(services, 1) + require.Contains(services, "consul") } diff --git a/command/agent/consul/unit_test.go b/command/agent/consul/unit_test.go index 285746213..25dbf2bfb 100644 --- a/command/agent/consul/unit_test.go +++ b/command/agent/consul/unit_test.go @@ -85,8 +85,9 @@ type restartRecorder struct { restarts int64 } -func (r *restartRecorder) Restart(source, reason string, failure bool) { +func (r *restartRecorder) Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error { atomic.AddInt64(&r.restarts, 1) + return nil } // testFakeCtx contains a fake Consul AgentAPI diff --git a/drivers/rawexec/driver_test.go b/drivers/rawexec/driver_test.go index 1c8ed6dbf..dad5810c5 100644 --- a/drivers/rawexec/driver_test.go +++ b/drivers/rawexec/driver_test.go @@ -292,6 +292,8 @@ func TestRawExecDriver_Start_Wait_AllocDir(t *testing.T) { // Task should terminate quickly waitCh, err := harness.WaitTask(context.Background(), task.ID) + require.NoError(err) + select { case res := <-waitCh: require.NoError(res.Err) diff --git a/drivers/rawexec/handle.go b/drivers/rawexec/handle.go index 09fe04e81..d4c8847bb 100644 --- a/drivers/rawexec/handle.go +++ b/drivers/rawexec/handle.go @@ -32,7 +32,7 @@ func (h *rawExecTaskHandle) IsRunning() bool { func (h *rawExecTaskHandle) run() { - // since run is called immediatly after the handle is created this + // since run is called immediately after the handle is created this // ensures the exitResult is initialized so we avoid a nil pointer // thus it does not need to be included in the lock if h.exitResult == nil { diff --git a/plugins/drivers/driver.go b/plugins/drivers/driver.go index 1f47a7165..fec8804d4 100644 --- a/plugins/drivers/driver.go +++ b/plugins/drivers/driver.go @@ -68,7 +68,7 @@ type Fingerprint struct { Health HealthState HealthDescription string - // Err is set by the plugin if an error occured during fingerprinting + // Err is set by the plugin if an error occurred during fingerprinting Err error } @@ -207,7 +207,7 @@ type TaskEvent struct { Message string Annotations map[string]string - // Err is only used if an error occured while consuming the RPC stream + // Err is only used if an error occurred while consuming the RPC stream Err error } From 4a4c7dbbfc1cc15fe26b2da9d372b2a69e80c29c Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Thu, 4 Oct 2018 15:08:20 -0400 Subject: [PATCH 119/152] client: begin driver plugin integration client: fingerprint driver plugins --- client/allocdir/task_dir.go | 4 + client/allocrunner/config.go | 8 + .../allocrunner/interfaces/task_lifecycle.go | 4 +- .../allocrunner/taskrunner/driver_handle.go | 64 ++ .../taskrunner/interfaces/handle.go | 47 + .../taskrunner/interfaces/lifecycle.go | 3 +- client/allocrunner/taskrunner/lifecycle.go | 32 +- .../taskrunner/restarts/restarts.go | 25 +- client/allocrunner/taskrunner/service_hook.go | 4 +- client/allocrunner/taskrunner/state/state.go | 5 + .../allocrunner/taskrunner/task_dir_hook.go | 2 +- .../taskrunner/task_runner_getters.go | 13 +- .../taskrunner/task_runner_hooks.go | 2 +- .../taskrunner/template/template.go | 2 +- client/allocrunner/taskrunner/vault_hook.go | 2 +- client/allocrunnerv2/alloc_runner.go | 550 +++++++++++ .../allocrunnerv2/taskrunner/task_runner.go | 903 ++++++++++++++++++ client/client.go | 110 +-- client/fingerprint_manager.go | 206 +--- command/agent/config.go | 4 +- command/agent/consul/script.go | 11 +- command/agent/consul/structs.go | 6 +- command/agent/plugins.go | 2 + drivers/rawexec/driver.go | 67 +- plugins/drivers/client.go | 38 +- plugins/drivers/driver.go | 54 +- plugins/drivers/proto/driver.pb.go | 450 ++++----- plugins/drivers/proto/driver.proto | 2 +- plugins/drivers/server.go | 44 +- plugins/drivers/testing.go | 17 +- plugins/drivers/utils.go | 138 ++- 31 files changed, 2240 insertions(+), 579 deletions(-) create mode 100644 client/allocrunner/taskrunner/driver_handle.go create mode 100644 client/allocrunner/taskrunner/interfaces/handle.go create mode 100644 client/allocrunnerv2/alloc_runner.go create mode 100644 client/allocrunnerv2/taskrunner/task_runner.go diff --git a/client/allocdir/task_dir.go b/client/allocdir/task_dir.go index 9486258d9..6595f82d9 100644 --- a/client/allocdir/task_dir.go +++ b/client/allocdir/task_dir.go @@ -13,6 +13,9 @@ import ( // TaskDir contains all of the paths relevant to a task. All paths are on the // host system so drivers should mount/link into task containers as necessary. type TaskDir struct { + // AllocDir is the path to the alloc directory on the host + AllocDir string + // Dir is the path to Task directory on the host Dir string @@ -50,6 +53,7 @@ func newTaskDir(logger hclog.Logger, allocDir, taskName string) *TaskDir { logger = logger.Named("task_dir").With("task_name", taskName) return &TaskDir{ + AllocDir: allocDir, Dir: taskDir, SharedAllocDir: filepath.Join(allocDir, SharedAllocName), LogDir: filepath.Join(allocDir, SharedAllocName, LogDirName), diff --git a/client/allocrunner/config.go b/client/allocrunner/config.go index f15ac255d..fae5cc4e4 100644 --- a/client/allocrunner/config.go +++ b/client/allocrunner/config.go @@ -9,6 +9,7 @@ import ( cstate "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/shared/loader" ) // Config holds the configuration for creating an allocation runner. @@ -37,4 +38,11 @@ type Config struct { // PrevAllocWatcher handles waiting on previous allocations and // migrating their ephemeral disk when necessary. PrevAllocWatcher allocwatcher.PrevAllocWatcher + + // PluginLoader is used to load plugins. + PluginLoader loader.PluginCatalog + + // PluginSingletonLoader is a plugin loader that will returns singleton + // instances of the plugins. + PluginSingletonLoader loader.PluginCatalog } diff --git a/client/allocrunner/interfaces/task_lifecycle.go b/client/allocrunner/interfaces/task_lifecycle.go index 8be2bfd6e..9fe16097a 100644 --- a/client/allocrunner/interfaces/task_lifecycle.go +++ b/client/allocrunner/interfaces/task_lifecycle.go @@ -3,7 +3,7 @@ package interfaces import ( "context" - "github.com/hashicorp/nomad/client/driver" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" "github.com/hashicorp/nomad/client/driver/env" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" @@ -82,7 +82,7 @@ type DriverStats interface { type TaskPoststartRequest struct { // Exec hook (may be nil) - DriverExec driver.ScriptExecutor + DriverExec interfaces.ScriptExecutor // Network info (may be nil) DriverNetwork *cstructs.DriverNetwork diff --git a/client/allocrunner/taskrunner/driver_handle.go b/client/allocrunner/taskrunner/driver_handle.go new file mode 100644 index 000000000..708209b45 --- /dev/null +++ b/client/allocrunner/taskrunner/driver_handle.go @@ -0,0 +1,64 @@ +package taskrunner + +import ( + "context" + "time" + + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/drivers" +) + +func NewDriverHandle(driver drivers.DriverPlugin, taskID string, task *structs.Task, net *cstructs.DriverNetwork) interfaces.DriverHandle { + return &driverHandleImpl{ + driver: driver, + net: net, + taskID: taskID, + task: task, + } +} + +type driverHandleImpl struct { + driver drivers.DriverPlugin + net *cstructs.DriverNetwork + task *structs.Task + taskID string +} + +func (h *driverHandleImpl) ID() string { + return h.taskID +} + +func (h *driverHandleImpl) WaitCh(ctx context.Context) (<-chan *drivers.ExitResult, error) { + return h.driver.WaitTask(ctx, h.taskID) +} + +func (h *driverHandleImpl) Update(task *structs.Task) error { + return nil +} + +func (h *driverHandleImpl) Kill() error { + return h.driver.StopTask(h.taskID, h.task.KillTimeout, h.task.KillSignal) +} + +func (h *driverHandleImpl) Stats() (*cstructs.TaskResourceUsage, error) { + return h.driver.TaskStats(h.taskID) +} + +func (h *driverHandleImpl) Signal(s string) error { + return h.driver.SignalTask(h.taskID, s) +} + +func (h *driverHandleImpl) Exec(timeout time.Duration, cmd string, args []string) ([]byte, int, error) { + command := append([]string{cmd}, args...) + res, err := h.driver.ExecTask(h.taskID, command, timeout) + if err != nil { + return nil, 0, err + } + return res.Stdout, res.ExitResult.ExitCode, res.ExitResult.Err +} + +func (h *driverHandleImpl) Network() *cstructs.DriverNetwork { + return h.net +} diff --git a/client/allocrunner/taskrunner/interfaces/handle.go b/client/allocrunner/taskrunner/interfaces/handle.go new file mode 100644 index 000000000..a0737c74f --- /dev/null +++ b/client/allocrunner/taskrunner/interfaces/handle.go @@ -0,0 +1,47 @@ +package interfaces + +import ( + "context" + "time" + + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/drivers" +) + +// DriverHandle wraps operations to a driver such that they are operated on a specific +// task +type DriverHandle interface { + // ID returns the task ID + ID() string + + // WaitCh is used to return a channel used to wait for task completion + WaitCh(context.Context) (<-chan *drivers.ExitResult, error) + + // Update is used to update the task if possible and update task related + // configurations. + Update(task *structs.Task) error + + // Kill is used to stop the task + Kill() error + + // Stats returns aggregated stats of the driver + Stats() (*cstructs.TaskResourceUsage, error) + + // Signal is used to send a signal to the task + Signal(s string) error + + // ScriptExecutor is an interface used to execute commands such as + // health check scripts in the a DriverHandle's context. + ScriptExecutor + + // Network returns the driver's network or nil if the driver did not + // create a network. + Network() *cstructs.DriverNetwork +} + +// ScriptExecutor is an interface that supports Exec()ing commands in the +// driver's context. Split out of DriverHandle to ease testing. +type ScriptExecutor interface { + Exec(timeout time.Duration, cmd string, args []string) ([]byte, int, error) +} diff --git a/client/allocrunner/taskrunner/interfaces/lifecycle.go b/client/allocrunner/taskrunner/interfaces/lifecycle.go index 10e8edf82..84bbda228 100644 --- a/client/allocrunner/taskrunner/interfaces/lifecycle.go +++ b/client/allocrunner/taskrunner/interfaces/lifecycle.go @@ -2,13 +2,12 @@ package interfaces import ( "context" - "os" "github.com/hashicorp/nomad/nomad/structs" ) type TaskLifecycle interface { Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error - Signal(event *structs.TaskEvent, s os.Signal) error + Signal(event *structs.TaskEvent, signal string) error Kill(ctx context.Context, event *structs.TaskEvent) error } diff --git a/client/allocrunner/taskrunner/lifecycle.go b/client/allocrunner/taskrunner/lifecycle.go index 4c9eabe10..551f8eb10 100644 --- a/client/allocrunner/taskrunner/lifecycle.go +++ b/client/allocrunner/taskrunner/lifecycle.go @@ -2,17 +2,16 @@ package taskrunner import ( "context" - "os" "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/drivers" ) // Restart a task. Returns immediately if no task is running. Blocks until // existing task exits or passed-in context is canceled. func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, failure bool) error { // Grab the handle - handle, result := tr.getDriverHandle() - + handle := tr.getDriverHandle() // Check it is running if handle == nil { return ErrTaskNotRunning @@ -32,13 +31,17 @@ func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, fai } // Drain the wait channel or wait for the request context to be canceled - result.Wait(ctx) + waitCh, err := handle.WaitCh(ctx) + if err != nil { + tr.logger.Error("failed to kill task. Resources may have been leaked", "error", err) + } + <-waitCh return nil } -func (tr *TaskRunner) Signal(event *structs.TaskEvent, s os.Signal) error { +func (tr *TaskRunner) Signal(event *structs.TaskEvent, s string) error { // Grab the handle - handle, _ := tr.getDriverHandle() + handle := tr.getDriverHandle() // Check it is running if handle == nil { @@ -49,6 +52,7 @@ func (tr *TaskRunner) Signal(event *structs.TaskEvent, s os.Signal) error { tr.EmitEvent(event) // Send the signal + return handle.Signal(s) } @@ -60,9 +64,9 @@ func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error tr.ctxCancel() // Grab the handle - handle, result := tr.getDriverHandle() + handle := tr.getDriverHandle() - // Check if the handle is running + // Check it is running if handle == nil { return ErrTaskNotRunning } @@ -84,7 +88,17 @@ func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error } // Block until task has exited. - result.Wait(ctx) + waitCh, err := handle.WaitCh(ctx) + + // The task may have already been cleaned up + if err != nil && err != drivers.ErrTaskNotFound { + tr.logger.Error("failed to wait on task. Resources may have been leaked", "error", err) + return err + } + + if waitCh != nil { + <-waitCh + } // Store that the task has been destroyed and any associated error. tr.UpdateState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskKilled).SetKillError(destroyErr)) diff --git a/client/allocrunner/taskrunner/restarts/restarts.go b/client/allocrunner/taskrunner/restarts/restarts.go index ec7a2594d..e46d30312 100644 --- a/client/allocrunner/taskrunner/restarts/restarts.go +++ b/client/allocrunner/taskrunner/restarts/restarts.go @@ -6,8 +6,8 @@ import ( "sync" "time" - dstructs "github.com/hashicorp/nomad/client/driver/structs" "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/drivers" ) const ( @@ -34,7 +34,7 @@ func NewRestartTracker(policy *structs.RestartPolicy, jobType string) *RestartTr } type RestartTracker struct { - waitRes *dstructs.WaitResult + exitRes *drivers.ExitResult startErr error killed bool // Whether the task has been killed restartTriggered bool // Whether the task has been signalled to be restarted @@ -72,11 +72,11 @@ func (r *RestartTracker) SetStartError(err error) *RestartTracker { return r } -// SetWaitResult is used to mark the most recent wait result. -func (r *RestartTracker) SetWaitResult(res *dstructs.WaitResult) *RestartTracker { +// SetExitResult is used to mark the most recent wait result. +func (r *RestartTracker) SetExitResult(res *drivers.ExitResult) *RestartTracker { r.lock.Lock() defer r.lock.Unlock() - r.waitRes = res + r.exitRes = res r.failure = true return r } @@ -112,6 +112,13 @@ func (r *RestartTracker) GetReason() string { return r.reason } +// GetCount returns the current restart count +func (r *RestartTracker) GetCount() int { + r.lock.Lock() + defer r.lock.Unlock() + return r.count +} + // GetState returns the tasks next state given the set exit code and start // error. One of the following states are returned: // * TaskRestarting - Task should be restarted @@ -129,7 +136,7 @@ func (r *RestartTracker) GetState() (string, time.Duration) { // Clear out the existing state defer func() { r.startErr = nil - r.waitRes = nil + r.exitRes = nil r.restartTriggered = false r.failure = false r.killed = false @@ -153,7 +160,7 @@ func (r *RestartTracker) GetState() (string, time.Duration) { // If the task does not restart on a successful exit code and // the exit code was successful: terminate. - if !r.onSuccess && r.waitRes != nil && r.waitRes.Successful() { + if !r.onSuccess && r.exitRes != nil && r.exitRes.Successful() { return structs.TaskTerminated, 0 } @@ -183,10 +190,10 @@ func (r *RestartTracker) GetState() (string, time.Duration) { r.reason = ReasonUnrecoverableErrror return structs.TaskNotRestarting, 0 } - } else if r.waitRes != nil { + } else if r.exitRes != nil { // If the task started successfully and restart on success isn't specified, // don't restart but don't mark as failed. - if r.waitRes.Successful() && !r.onSuccess { + if r.exitRes.Successful() && !r.onSuccess { r.reason = "Restart unnecessary as task terminated successfully" return structs.TaskTerminated, 0 } diff --git a/client/allocrunner/taskrunner/service_hook.go b/client/allocrunner/taskrunner/service_hook.go index 7fdbba50a..801cc2073 100644 --- a/client/allocrunner/taskrunner/service_hook.go +++ b/client/allocrunner/taskrunner/service_hook.go @@ -7,8 +7,8 @@ import ( log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/allocrunner/interfaces" + tinterfaces "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/client/consul" - "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" cstructs "github.com/hashicorp/nomad/client/structs" agentconsul "github.com/hashicorp/nomad/command/agent/consul" @@ -34,7 +34,7 @@ type serviceHook struct { logger log.Logger // The following fields may be updated - driverExec driver.ScriptExecutor + driverExec tinterfaces.ScriptExecutor driverNet *cstructs.DriverNetwork canary bool services []*structs.Service diff --git a/client/allocrunner/taskrunner/state/state.go b/client/allocrunner/taskrunner/state/state.go index 43a73d4bf..1b4f07aba 100644 --- a/client/allocrunner/taskrunner/state/state.go +++ b/client/allocrunner/taskrunner/state/state.go @@ -3,6 +3,7 @@ package state import ( "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/helper" + "github.com/hashicorp/nomad/plugins/drivers" ) // LocalState is Task state which is persisted for use when restarting Nomad @@ -13,6 +14,9 @@ type LocalState struct { // DriverNetwork is the network information returned by the task // driver's Start method DriverNetwork *structs.DriverNetwork + + // TaskHandle is the handle used to reattach to the task during recovery + TaskHandle *drivers.TaskHandle } func NewLocalState() *LocalState { @@ -27,6 +31,7 @@ func (s *LocalState) Copy() *LocalState { c := &LocalState{ Hooks: make(map[string]*HookState, len(s.Hooks)), DriverNetwork: s.DriverNetwork, + TaskHandle: s.TaskHandle, } // Copy the hooks diff --git a/client/allocrunner/taskrunner/task_dir_hook.go b/client/allocrunner/taskrunner/task_dir_hook.go index 4053f54ad..109d69101 100644 --- a/client/allocrunner/taskrunner/task_dir_hook.go +++ b/client/allocrunner/taskrunner/task_dir_hook.go @@ -38,7 +38,7 @@ func (h *taskDirHook) Prestart(ctx context.Context, req *interfaces.TaskPrestart h.runner.EmitEvent(structs.NewTaskEvent(structs.TaskSetup).SetMessage(structs.TaskBuildingTaskDir)) // Build the task directory structure - fsi := h.runner.driver.FSIsolation() + fsi := h.runner.driverCapabilities.FSIsolation err := h.runner.taskDir.Build(false, chroot, fsi) if err != nil { return err diff --git a/client/allocrunner/taskrunner/task_runner_getters.go b/client/allocrunner/taskrunner/task_runner_getters.go index abce06593..67d3a4810 100644 --- a/client/allocrunner/taskrunner/task_runner_getters.go +++ b/client/allocrunner/taskrunner/task_runner_getters.go @@ -1,7 +1,7 @@ package taskrunner import ( - "github.com/hashicorp/nomad/client/driver" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" "github.com/hashicorp/nomad/nomad/structs" ) @@ -51,23 +51,24 @@ func (tr *TaskRunner) setVaultToken(token string) { // getDriverHandle returns a driver handle and its result proxy. Use the // result proxy instead of the handle's WaitCh. -func (tr *TaskRunner) getDriverHandle() (driver.DriverHandle, *handleResult) { +func (tr *TaskRunner) getDriverHandle() interfaces.DriverHandle { tr.handleLock.Lock() defer tr.handleLock.Unlock() - return tr.handle, tr.handleResult + return tr.handle } // setDriverHanlde sets the driver handle and creates a new result proxy. -func (tr *TaskRunner) setDriverHandle(handle driver.DriverHandle) { +func (tr *TaskRunner) setDriverHandle(handle interfaces.DriverHandle) { tr.handleLock.Lock() defer tr.handleLock.Unlock() tr.handle = handle - tr.handleResult = newHandleResult(handle.WaitCh()) } func (tr *TaskRunner) clearDriverHandle() { tr.handleLock.Lock() defer tr.handleLock.Unlock() + if tr.handle != nil { + tr.driver.DestroyTask(tr.handle.ID(), true) + } tr.handle = nil - tr.handleResult = nil } diff --git a/client/allocrunner/taskrunner/task_runner_hooks.go b/client/allocrunner/taskrunner/task_runner_hooks.go index 7ad89f64b..6ccb808a4 100644 --- a/client/allocrunner/taskrunner/task_runner_hooks.go +++ b/client/allocrunner/taskrunner/task_runner_hooks.go @@ -168,7 +168,7 @@ func (tr *TaskRunner) poststart() error { }() } - handle, _ := tr.getDriverHandle() + handle := tr.getDriverHandle() net := handle.Network() var merr multierror.Error diff --git a/client/allocrunner/taskrunner/template/template.go b/client/allocrunner/taskrunner/template/template.go index 0d90306ec..54f035a9a 100644 --- a/client/allocrunner/taskrunner/template/template.go +++ b/client/allocrunner/taskrunner/template/template.go @@ -448,7 +448,7 @@ func (tm *TaskTemplateManager) handleTemplateRerenders(allRenderedTime time.Time for signal := range signals { s := tm.signals[signal] event := structs.NewTaskEvent(structs.TaskSignaling).SetTaskSignal(s).SetDisplayMessage("Template re-rendered") - if err := tm.config.Lifecycle.Signal(event, s); err != nil { + if err := tm.config.Lifecycle.Signal(event, signal); err != nil { multierror.Append(&mErr, err) } } diff --git a/client/allocrunner/taskrunner/vault_hook.go b/client/allocrunner/taskrunner/vault_hook.go index 1329e6cfe..d06ecbd81 100644 --- a/client/allocrunner/taskrunner/vault_hook.go +++ b/client/allocrunner/taskrunner/vault_hook.go @@ -240,7 +240,7 @@ OUTER: } event := structs.NewTaskEvent(structs.TaskSignaling).SetTaskSignal(s).SetDisplayMessage("Vault: new Vault token acquired") - if err := h.lifecycle.Signal(event, s); err != nil { + if err := h.lifecycle.Signal(event, h.vaultStanza.ChangeSignal); err != nil { h.logger.Error("failed to send signal", "error", err) h.lifecycle.Kill(h.ctx, structs.NewTaskEvent(structs.TaskKilling). diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go new file mode 100644 index 000000000..171d6b8c7 --- /dev/null +++ b/client/allocrunnerv2/alloc_runner.go @@ -0,0 +1,550 @@ +package allocrunnerv2 + +import ( + "context" + "fmt" + "path/filepath" + "sync" + "time" + + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocdir" + "github.com/hashicorp/nomad/client/allocrunner" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + "github.com/hashicorp/nomad/client/allocrunnerv2/state" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" + "github.com/hashicorp/nomad/client/allocwatcher" + "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/consul" + cinterfaces "github.com/hashicorp/nomad/client/interfaces" + cstate "github.com/hashicorp/nomad/client/state" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/client/vaultclient" + "github.com/hashicorp/nomad/helper" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/shared/loader" +) + +// allocRunner is used to run all the tasks in a given allocation +type allocRunner struct { + // id is the ID of the allocation. Can be accessed without a lock + id string + + // Logger is the logger for the alloc runner. + logger log.Logger + + clientConfig *config.Config + + // stateUpdater is used to emit updated task state + stateUpdater cinterfaces.AllocStateHandler + + // consulClient is the client used by the consul service hook for + // registering services and checks + consulClient consul.ConsulServiceAPI + + // vaultClient is the used to manage Vault tokens + vaultClient vaultclient.VaultClient + + // waitCh is closed when the Run() loop has exited + waitCh chan struct{} + + // destroyed is true when the Run() loop has exited, postrun hooks have + // run, and alloc runner has been destroyed + destroyed bool + destroyedLock sync.Mutex + + // Alloc captures the allocation being run. + alloc *structs.Allocation + allocLock sync.RWMutex + + // state is the alloc runner's state + state *state.State + stateLock sync.RWMutex + + stateDB cstate.StateDB + + // allocDir is used to build the allocations directory structure. + allocDir *allocdir.AllocDir + + // runnerHooks are alloc runner lifecycle hooks that should be run on state + // transistions. + runnerHooks []interfaces.RunnerHook + + // tasks are the set of task runners + tasks map[string]*taskrunner.TaskRunner + tasksLock sync.RWMutex + + // allocBroadcaster sends client allocation updates to all listeners + allocBroadcaster *cstructs.AllocBroadcaster + + // prevAllocWatcher allows waiting for a previous allocation to exit + // and if necessary migrate its alloc dir. + prevAllocWatcher allocwatcher.PrevAllocWatcher + + // pluginLoader is used to load plugins. + pluginLoader loader.PluginCatalog + + // pluginSingletonLoader is a plugin loader that will returns singleton + // instances of the plugins. + pluginSingletonLoader loader.PluginCatalog +} + +// NewAllocRunner returns a new allocation runner. +func NewAllocRunner(config *Config) (*allocRunner, error) { + alloc := config.Alloc + tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup) + if tg == nil { + return nil, fmt.Errorf("failed to lookup task group %q", alloc.TaskGroup) + } + + ar := &allocRunner{ + id: alloc.ID, + alloc: alloc, + clientConfig: config.ClientConfig, + consulClient: config.Consul, + vaultClient: config.Vault, + tasks: make(map[string]*taskrunner.TaskRunner, len(tg.Tasks)), + waitCh: make(chan struct{}), + state: &state.State{}, + stateDB: config.StateDB, + stateUpdater: config.StateUpdater, + allocBroadcaster: cstructs.NewAllocBroadcaster(), + prevAllocWatcher: config.PrevAllocWatcher, + pluginLoader: config.PluginLoader, + pluginSingletonLoader: config.PluginSingletonLoader, + } + + // Create the logger based on the allocation ID + ar.logger = config.Logger.Named("alloc_runner").With("alloc_id", alloc.ID) + + // Create alloc dir + ar.allocDir = allocdir.NewAllocDir(ar.logger, filepath.Join(config.ClientConfig.AllocDir, alloc.ID)) + + // Initialize the runners hooks. + ar.initRunnerHooks() + + // Create the TaskRunners + if err := ar.initTaskRunners(tg.Tasks); err != nil { + return nil, err + } + + return ar, nil +} + +// initTaskRunners creates task runners but does *not* run them. +func (ar *allocRunner) initTaskRunners(tasks []*structs.Task) error { + for _, task := range tasks { + config := &taskrunner.Config{ + Alloc: ar.alloc, + ClientConfig: ar.clientConfig, + Task: task, + TaskDir: ar.allocDir.NewTaskDir(task.Name), + Logger: ar.logger, + StateDB: ar.stateDB, + StateUpdater: ar, + Consul: ar.consulClient, + VaultClient: ar.vaultClient, + PluginLoader: ar.pluginLoader, + PluginSingletonLoader: ar.pluginSingletonLoader, + } + + // Create, but do not Run, the task runner + tr, err := taskrunner.NewTaskRunner(config) + if err != nil { + return fmt.Errorf("failed creating runner for task %q: %v", task.Name, err) + } + + ar.tasks[task.Name] = tr + } + return nil +} + +func (ar *allocRunner) WaitCh() <-chan struct{} { + return ar.waitCh +} + +// XXX How does alloc Restart work +// Run is the main goroutine that executes all the tasks. +func (ar *allocRunner) Run() { + // Close the wait channel + defer close(ar.waitCh) + + var taskWaitCh <-chan struct{} + + // Run the prestart hooks + // XXX Equivalent to TR.Prestart hook + if err := ar.prerun(); err != nil { + ar.logger.Error("prerun failed", "error", err) + goto POST + } + + // Run the runners + taskWaitCh = ar.runImpl() + +MAIN: + for { + select { + case <-taskWaitCh: + // TaskRunners have all exited + break MAIN + } + } + +POST: + // Run the postrun hooks + // XXX Equivalent to TR.Poststop hook + if err := ar.postrun(); err != nil { + ar.logger.Error("postrun failed", "error", err) + } +} + +// runImpl is used to run the runners. +func (ar *allocRunner) runImpl() <-chan struct{} { + for _, task := range ar.tasks { + go task.Run() + } + + // Return a combined WaitCh that is closed when all task runners have + // exited. + waitCh := make(chan struct{}) + go func() { + defer close(waitCh) + for _, task := range ar.tasks { + <-task.WaitCh() + } + }() + + return waitCh +} + +// Alloc returns the current allocation being run by this runner as sent by the +// server. This view of the allocation does not have updated task states. +func (ar *allocRunner) Alloc() *structs.Allocation { + ar.allocLock.RLock() + defer ar.allocLock.RUnlock() + return ar.alloc +} + +func (ar *allocRunner) setAlloc(updated *structs.Allocation) { + ar.allocLock.Lock() + ar.alloc = updated + ar.allocLock.Unlock() +} + +// GetAllocDir returns the alloc dir which is safe for concurrent use. +func (ar *allocRunner) GetAllocDir() *allocdir.AllocDir { + return ar.allocDir +} + +// Restore state from database. Must be called after NewAllocRunner but before +// Run. +func (ar *allocRunner) Restore() error { + // Restore task runners + for _, tr := range ar.tasks { + if err := tr.Restore(); err != nil { + return err + } + } + + return nil +} + +// TaskStateUpdated is called by TaskRunner when a task's state has been +// updated. This hook is used to compute changes to the alloc's ClientStatus +// and to update the server with the new state. +func (ar *allocRunner) TaskStateUpdated(taskName string, state *structs.TaskState) { + // If a task is dead, we potentially want to kill other tasks in the group + if state.State == structs.TaskStateDead { + // Find all tasks that are not the one that is dead and check if the one + // that is dead is a leader + var otherTaskRunners []*taskrunner.TaskRunner + var otherTaskNames []string + leader := false + for name, tr := range ar.tasks { + if name != taskName { + otherTaskRunners = append(otherTaskRunners, tr) + otherTaskNames = append(otherTaskNames, name) + } else if tr.Task().Leader { + leader = true + } + } + + // If the task failed, we should kill all the other tasks in the task group. + if state.Failed { + if len(otherTaskRunners) > 0 { + ar.logger.Debug("task failure, destroying all tasks", "failed_task", taskName, "destroying", otherTaskNames) + } + for _, tr := range otherTaskRunners { + tr.Kill(context.Background(), structs.NewTaskEvent(structs.TaskSiblingFailed).SetFailedSibling(taskName)) + } + } else if leader { + if len(otherTaskRunners) > 0 { + ar.logger.Debug("leader task dead, destroying all tasks", "leader_task", taskName, "destroying", otherTaskNames) + } + // If the task was a leader task we should kill all the other tasks. + for _, tr := range otherTaskRunners { + tr.Kill(context.Background(), structs.NewTaskEvent(structs.TaskLeaderDead)) + } + } + } + + // Gather the state of the other tasks + ar.tasksLock.RLock() + states := make(map[string]*structs.TaskState, len(ar.tasks)) + for name, tr := range ar.tasks { + if name == taskName { + states[name] = state + } else { + states[name] = tr.TaskState() + } + } + ar.tasksLock.RUnlock() + + // Get the client allocation + calloc := ar.clientAlloc(states) + + // Update the server + ar.stateUpdater.AllocStateUpdated(calloc) + + // Broadcast client alloc to listeners + ar.allocBroadcaster.Send(calloc) +} + +// clientAlloc takes in the task states and returns an Allocation populated +// with Client specific fields +func (ar *allocRunner) clientAlloc(taskStates map[string]*structs.TaskState) *structs.Allocation { + ar.stateLock.RLock() + defer ar.stateLock.RUnlock() + + // store task states for AllocState to expose + ar.state.TaskStates = taskStates + + a := &structs.Allocation{ + ID: ar.id, + TaskStates: taskStates, + } + + if d := ar.state.DeploymentStatus; d != nil { + a.DeploymentStatus = d.Copy() + } + + // Compute the ClientStatus + if ar.state.ClientStatus != "" { + // The client status is being forced + a.ClientStatus, a.ClientDescription = ar.state.ClientStatus, ar.state.ClientDescription + } else { + a.ClientStatus, a.ClientDescription = getClientStatus(taskStates) + } + + // If the allocation is terminal, make sure all required fields are properly + // set. + if a.ClientTerminalStatus() { + alloc := ar.Alloc() + + // If we are part of a deployment and the task has failed, mark the + // alloc as unhealthy. This guards against the watcher not be started. + if a.ClientStatus == structs.AllocClientStatusFailed && + alloc.DeploymentID != "" && !a.DeploymentStatus.IsUnhealthy() { + a.DeploymentStatus = &structs.AllocDeploymentStatus{ + Healthy: helper.BoolToPtr(false), + } + } + + // Make sure we have marked the finished at for every task. This is used + // to calculate the reschedule time for failed allocations. + now := time.Now() + for _, task := range alloc.Job.LookupTaskGroup(alloc.TaskGroup).Tasks { + ts, ok := a.TaskStates[task.Name] + if !ok { + ts = &structs.TaskState{} + a.TaskStates[task.Name] = ts + } + if ts.FinishedAt.IsZero() { + ts.FinishedAt = now + } + } + } + + return a +} + +// getClientStatus takes in the task states for a given allocation and computes +// the client status and description +func getClientStatus(taskStates map[string]*structs.TaskState) (status, description string) { + var pending, running, dead, failed bool + for _, state := range taskStates { + switch state.State { + case structs.TaskStateRunning: + running = true + case structs.TaskStatePending: + pending = true + case structs.TaskStateDead: + if state.Failed { + failed = true + } else { + dead = true + } + } + } + + // Determine the alloc status + if failed { + return structs.AllocClientStatusFailed, "Failed tasks" + } else if running { + return structs.AllocClientStatusRunning, "Tasks are running" + } else if pending { + return structs.AllocClientStatusPending, "No tasks have started" + } else if dead { + return structs.AllocClientStatusComplete, "All tasks have completed" + } + + return "", "" +} + +// AllocState returns a copy of allocation state including a snapshot of task +// states. +func (ar *allocRunner) AllocState() *state.State { + ar.stateLock.RLock() + state := ar.state.Copy() + ar.stateLock.RUnlock() + + // If TaskStateUpdated has not been called yet, ar.state.TaskStates + // won't be set as it is not the canonical source of TaskStates. + if len(state.TaskStates) == 0 { + ar.tasksLock.RLock() + ar.state.TaskStates = make(map[string]*structs.TaskState, len(ar.tasks)) + for k, tr := range ar.tasks { + state.TaskStates[k] = tr.TaskState() + } + ar.tasksLock.RUnlock() + } + + return state +} + +// Update the running allocation with a new version received from the server. +// +// This method sends the updated alloc to Run for serially processing updates. +// If there is already a pending update it will be discarded and replaced by +// the latest update. +func (ar *allocRunner) Update(update *structs.Allocation) { + // Update ar.alloc + ar.setAlloc(update) + + // Run hooks + if err := ar.update(update); err != nil { + ar.logger.Error("error running update hooks", "error", err) + } + + // Update task runners + for _, tr := range ar.tasks { + tr.Update(update) + } +} + +func (ar *allocRunner) Listener() *cstructs.AllocListener { + return ar.allocBroadcaster.Listen() +} + +// Destroy the alloc runner by synchronously stopping it if it is still running +// and cleaning up all of its resources. +// +// This method is safe for calling concurrently with Run() and will cause it to +// exit (thus closing WaitCh). +func (ar *allocRunner) Destroy() { + // Stop tasks + ar.tasksLock.RLock() + for name, tr := range ar.tasks { + err := tr.Kill(context.TODO(), structs.NewTaskEvent(structs.TaskKilled)) + if err != nil { + if err == taskrunner.ErrTaskNotRunning { + ar.logger.Trace("task not running", "task_name", name) + } else { + ar.logger.Warn("failed to kill task", "error", err, "task_name", name) + } + } + } + ar.tasksLock.RUnlock() + + // Wait for tasks to exit and postrun hooks to finish + <-ar.waitCh + + // Run destroy hooks + if err := ar.destroy(); err != nil { + ar.logger.Warn("error running destroy hooks", "error", err) + } + + // Cleanup state db + if err := ar.stateDB.DeleteAllocationBucket(ar.id); err != nil { + ar.logger.Warn("failed to delete allocation state", "error", err) + } + + // Mark alloc as destroyed + ar.destroyedLock.Lock() + ar.destroyed = true + ar.destroyedLock.Unlock() +} + +// IsDestroyed returns true if the alloc runner has been destroyed (stopped and +// garbage collected). +// +// This method is safe for calling concurrently with Run(). Callers must +// receive on WaitCh() to block until alloc runner has stopped and been +// destroyed. +func (ar *allocRunner) IsDestroyed() bool { + ar.destroyedLock.Lock() + defer ar.destroyedLock.Unlock() + return ar.destroyed +} + +// IsWaiting returns true if the alloc runner is waiting for its previous +// allocation to terminate. +// +// This method is safe for calling concurrently with Run(). +func (ar *allocRunner) IsWaiting() bool { + return ar.prevAllocWatcher.IsWaiting() +} + +// IsMigrating returns true if the alloc runner is migrating data from its +// previous allocation. +// +// This method is safe for calling concurrently with Run(). +func (ar *allocRunner) IsMigrating() bool { + return ar.prevAllocWatcher.IsMigrating() +} + +func (ar *allocRunner) StatsReporter() allocrunner.AllocStatsReporter { + return ar +} + +// LatestAllocStats returns the latest stats for an allocation. If taskFilter +// is set, only stats for that task -- if it exists -- are returned. +func (ar *allocRunner) LatestAllocStats(taskFilter string) (*cstructs.AllocResourceUsage, error) { + ar.tasksLock.RLock() + defer ar.tasksLock.RUnlock() + + astat := &cstructs.AllocResourceUsage{ + Tasks: make(map[string]*cstructs.TaskResourceUsage, len(ar.tasks)), + ResourceUsage: &cstructs.ResourceUsage{ + MemoryStats: &cstructs.MemoryStats{}, + CpuStats: &cstructs.CpuStats{}, + }, + } + + for name, tr := range ar.tasks { + if taskFilter != "" && taskFilter != name { + // Getting stats for a particular task and its not this one! + continue + } + + if usage := tr.LatestResourceUsage(); usage != nil { + astat.Tasks[name] = usage + astat.ResourceUsage.Add(usage.ResourceUsage) + if usage.Timestamp > astat.Timestamp { + astat.Timestamp = usage.Timestamp + } + } + } + + return astat, nil +} diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go new file mode 100644 index 000000000..14f64789e --- /dev/null +++ b/client/allocrunnerv2/taskrunner/task_runner.go @@ -0,0 +1,903 @@ +package taskrunner + +import ( + "context" + "errors" + "fmt" + "sync" + "time" + + metrics "github.com/armon/go-metrics" + log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/hcl2/hcl" + "github.com/hashicorp/hcl2/hcldec" + "github.com/hashicorp/nomad/client/allocdir" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" + "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" + tinterfaces "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" + "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/consul" + "github.com/hashicorp/nomad/client/driver/env" + cstate "github.com/hashicorp/nomad/client/state" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/client/vaultclient" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/drivers" + "github.com/hashicorp/nomad/plugins/shared" + "github.com/hashicorp/nomad/plugins/shared/hclspec" + "github.com/hashicorp/nomad/plugins/shared/loader" + "github.com/zclconf/go-cty/cty" +) + +const ( + // killBackoffBaseline is the baseline time for exponential backoff while + // killing a task. + killBackoffBaseline = 5 * time.Second + + // killBackoffLimit is the limit of the exponential backoff for killing + // the task. + killBackoffLimit = 2 * time.Minute + + // killFailureLimit is how many times we will attempt to kill a task before + // giving up and potentially leaking resources. + killFailureLimit = 5 + + // triggerUpdatechCap is the capacity for the triggerUpdateCh used for + // triggering updates. It should be exactly 1 as even if multiple + // updates have come in since the last one was handled, we only need to + // handle the last one. + triggerUpdateChCap = 1 +) + +type TaskRunner struct { + // allocID and taskName are immutable so these fields may be accessed + // without locks + allocID string + taskName string + + alloc *structs.Allocation + allocLock sync.Mutex + + clientConfig *config.Config + + // stateUpdater is used to emit updated task state + stateUpdater interfaces.TaskStateHandler + + // state captures the state of the task for updating the allocation + state *structs.TaskState + stateLock sync.Mutex + + // localState captures the node-local state of the task for when the + // Nomad agent restarts + localState *state.LocalState + localStateLock sync.RWMutex + + // stateDB is for persisting localState and taskState + stateDB cstate.StateDB + + // persistedHash is the hash of the last persisted state for skipping + // unnecessary writes + persistedHash []byte + + // ctx is the task runner's context representing the tasks's lifecycle. + // Canceling the context will cause the task to be destroyed. + ctx context.Context + + // ctxCancel is used to exit the task runner's Run loop without + // stopping the task. Shutdown hooks are run. + ctxCancel context.CancelFunc + + // Logger is the logger for the task runner. + logger log.Logger + + // triggerUpdateCh is ticked whenever update hooks need to be run and + // must be created with cap=1 to signal a pending update and prevent + // callers from deadlocking if the receiver has exited. + triggerUpdateCh chan struct{} + + // waitCh is closed when the task runner has transitioned to a terminal + // state + waitCh chan struct{} + + // driver is the driver for the task. + driver drivers.DriverPlugin + + // driverCapabilities is the set capabilities the driver supports + driverCapabilities *drivers.Capabilities + + // taskSchema is the hcl spec for the task driver configuration + taskSchema hcldec.Spec + + // handleLock guards access to handle and handleResult + handleLock sync.Mutex + + // handle to the running driver + handle tinterfaces.DriverHandle + + // network is the configuration for the driver network if one was created + network *cstructs.DriverNetwork + + // task is the task being run + task *structs.Task + taskLock sync.RWMutex + + // taskDir is the directory structure for this task. + taskDir *allocdir.TaskDir + + // envBuilder is used to build the task's environment + envBuilder *env.Builder + + // restartTracker is used to decide if the task should be restarted. + restartTracker *restarts.RestartTracker + + // runnerHooks are task runner lifecycle hooks that should be run on state + // transistions. + runnerHooks []interfaces.TaskHook + + // consulClient is the client used by the consul service hook for + // registering services and checks + consulClient consul.ConsulServiceAPI + + // vaultClient is the client to use to derive and renew Vault tokens + vaultClient vaultclient.VaultClient + + // vaultToken is the current Vault token. It should be accessed with the + // getter. + vaultToken string + vaultTokenLock sync.Mutex + + // baseLabels are used when emitting tagged metrics. All task runner metrics + // will have these tags, and optionally more. + baseLabels []metrics.Label + + // logmonHookConfig is used to get the paths to the stdout and stderr fifos + // to be passed to the driver for task logging + logmonHookConfig *logmonHookConfig + + // resourceUsage is written via UpdateStats and read via + // LatestResourceUsage. May be nil at all times. + resourceUsage *cstructs.TaskResourceUsage + resourceUsageLock sync.Mutex + + // PluginLoader is used to load plugins. + pluginLoader loader.PluginCatalog + + // PluginSingletonLoader is a plugin loader that will returns singleton + // instances of the plugins. + pluginSingletonLoader loader.PluginCatalog +} + +type Config struct { + Alloc *structs.Allocation + ClientConfig *config.Config + Consul consul.ConsulServiceAPI + Task *structs.Task + TaskDir *allocdir.TaskDir + Logger log.Logger + + // VaultClient is the client to use to derive and renew Vault tokens + VaultClient vaultclient.VaultClient + + // LocalState is optionally restored task state + LocalState *state.LocalState + + // StateDB is used to store and restore state. + StateDB cstate.StateDB + + // StateUpdater is used to emit updated task state + StateUpdater interfaces.TaskStateHandler + + // PluginLoader is used to load plugins. + PluginLoader loader.PluginCatalog + + // PluginSingletonLoader is a plugin loader that will returns singleton + // instances of the plugins. + PluginSingletonLoader loader.PluginCatalog +} + +func NewTaskRunner(config *Config) (*TaskRunner, error) { + // Create a context for the runner + trCtx, trCancel := context.WithCancel(context.Background()) + + // Initialize the environment builder + envBuilder := env.NewBuilder( + config.ClientConfig.Node, + config.Alloc, + config.Task, + config.ClientConfig.Region, + ) + + tr := &TaskRunner{ + alloc: config.Alloc, + allocID: config.Alloc.ID, + clientConfig: config.ClientConfig, + task: config.Task, + taskDir: config.TaskDir, + taskName: config.Task.Name, + envBuilder: envBuilder, + consulClient: config.Consul, + vaultClient: config.VaultClient, + state: config.Alloc.TaskStates[config.Task.Name].Copy(), + localState: config.LocalState, + stateDB: config.StateDB, + stateUpdater: config.StateUpdater, + ctx: trCtx, + ctxCancel: trCancel, + triggerUpdateCh: make(chan struct{}, triggerUpdateChCap), + waitCh: make(chan struct{}), + pluginLoader: config.PluginLoader, + pluginSingletonLoader: config.PluginSingletonLoader, + } + + // Create the logger based on the allocation ID + tr.logger = config.Logger.Named("task_runner").With("task", config.Task.Name) + + // Build the restart tracker. + tg := tr.alloc.Job.LookupTaskGroup(tr.alloc.TaskGroup) + if tg == nil { + tr.logger.Error("alloc missing task group") + return nil, fmt.Errorf("alloc missing task group") + } + tr.restartTracker = restarts.NewRestartTracker(tg.RestartPolicy, tr.alloc.Job.Type) + + // Initialize the task state + tr.initState() + + // Get the driver + if err := tr.initDriver(); err != nil { + tr.logger.Error("failed to create driver", "error", err) + return nil, err + } + + // Initialize the runners hooks. + tr.initHooks() + + // Initialize base labels + tr.initLabels() + + return tr, nil +} + +func (tr *TaskRunner) initState() { + if tr.state == nil { + tr.state = &structs.TaskState{ + State: structs.TaskStatePending, + } + } + if tr.localState == nil { + tr.localState = state.NewLocalState() + } +} + +func (tr *TaskRunner) initLabels() { + alloc := tr.Alloc() + tr.baseLabels = []metrics.Label{ + { + Name: "job", + Value: alloc.Job.Name, + }, + { + Name: "task_group", + Value: alloc.TaskGroup, + }, + { + Name: "alloc_id", + Value: tr.allocID, + }, + { + Name: "task", + Value: tr.taskName, + }, + } +} + +func (tr *TaskRunner) Run() { + defer close(tr.waitCh) + var result *drivers.ExitResult + + // Updates are handled asynchronously with the other hooks but each + // triggered update - whether due to alloc updates or a new vault token + // - should be handled serially. + go tr.handleUpdates() + +MAIN: + for tr.ctx.Err() == nil { + // Run the prestart hooks + if err := tr.prestart(); err != nil { + tr.logger.Error("prestart failed", "error", err) + tr.restartTracker.SetStartError(err) + goto RESTART + } + + if tr.ctx.Err() != nil { + break MAIN + } + + // Run the task + if err := tr.runDriver(); err != nil { + tr.logger.Error("running driver failed", "error", err) + tr.restartTracker.SetStartError(err) + goto RESTART + } + + // Run the poststart hooks + if err := tr.poststart(); err != nil { + tr.logger.Error("poststart failed", "error", err) + } + + // Grab the result proxy and wait for task to exit + { + handle := tr.getDriverHandle() + + // Do *not* use tr.ctx here as it would cause Wait() to + // unblock before the task exits when Kill() is called. + if resultCh, err := handle.WaitCh(context.Background()); err != nil { + tr.logger.Error("wait task failed", "error", err) + } else { + result = <-resultCh + } + } + + // Clear the handle + tr.clearDriverHandle() + + // Store the wait result on the restart tracker + tr.restartTracker.SetExitResult(result) + + if err := tr.exited(); err != nil { + tr.logger.Error("exited hooks failed", "error", err) + } + + RESTART: + restart, restartDelay := tr.shouldRestart() + if !restart { + break MAIN + } + + // Actually restart by sleeping and also watching for destroy events + select { + case <-time.After(restartDelay): + case <-tr.ctx.Done(): + tr.logger.Trace("task killed between restarts", "delay", restartDelay) + break MAIN + } + } + + // If task terminated, update server. All other exit conditions (eg + // killed or out of restarts) will perform their own server updates. + if result != nil { + event := structs.NewTaskEvent(structs.TaskTerminated). + SetExitCode(result.ExitCode). + SetSignal(result.Signal). + SetExitMessage(result.Err) + tr.UpdateState(structs.TaskStateDead, event) + } + + // Run the stop hooks + if err := tr.stop(); err != nil { + tr.logger.Error("stop failed", "error", err) + } + + tr.logger.Debug("task run loop exiting") +} + +// handleUpdates runs update hooks when triggerUpdateCh is ticked and exits +// when Run has returned. Should only be run in a goroutine from Run. +func (tr *TaskRunner) handleUpdates() { + for { + select { + case <-tr.triggerUpdateCh: + case <-tr.waitCh: + return + } + + if tr.Alloc().TerminalStatus() { + // Terminal update: kill TaskRunner and let Run execute postrun hooks + err := tr.Kill(context.TODO(), structs.NewTaskEvent(structs.TaskKilled)) + if err != nil { + tr.logger.Warn("error stopping task", "error", err) + } + continue + } + + // Non-terminal update; run hooks + tr.updateHooks() + } +} + +// shouldRestart determines whether the task should be restarted and updates +// the task state unless the task is killed or terminated. +func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { + // Determine if we should restart + state, when := tr.restartTracker.GetState() + reason := tr.restartTracker.GetReason() + switch state { + case structs.TaskKilled: + // Never restart an explicitly killed task. Kill method handles + // updating the server. + return false, 0 + case structs.TaskNotRestarting, structs.TaskTerminated: + tr.logger.Info("not restarting task", "reason", reason) + if state == structs.TaskNotRestarting { + tr.UpdateState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskNotRestarting).SetRestartReason(reason).SetFailsTask()) + } + return false, 0 + case structs.TaskRestarting: + tr.logger.Info("restarting task", "reason", reason, "delay", when) + tr.UpdateState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskRestarting).SetRestartDelay(when).SetRestartReason(reason)) + return true, 0 + default: + tr.logger.Error("restart tracker returned unknown state", "state", state) + return true, when + } +} + +// runDriver runs the driver and waits for it to exit +func (tr *TaskRunner) runDriver() error { + + taskConfig := drivers.NewTaskConfig(tr.task, tr.taskDir, tr.envBuilder.Build()) + taskConfig.ID = tr.buildID() + taskConfig.StdoutPath = tr.logmonHookConfig.stdoutFifo + taskConfig.StderrPath = tr.logmonHookConfig.stderrFifo + + evalCtx := &hcl.EvalContext{ + Functions: shared.GetStdlibFuncs(), + Variables: map[string]cty.Value{ + "NOMAD_ENV_bin": cty.StringVal("/bin/consul"), + }, + } + + val, diag := shared.ParseHclInterface(tr.task.Config, tr.taskSchema, evalCtx) + if diag.HasErrors() { + errStr := "failed to parse config" + for _, err := range diag.Errs() { + errStr = fmt.Sprintf("%s\n* %s", errStr, err.Error()) + } + return errors.New(errStr) + } + + if err := taskConfig.EncodeDriverConfig(val); err != nil { + tr.logger.Warn("failed to encode driver config", "error", err) + return err + } + + //TODO mounts and devices + //XXX Evaluate and encode driver config + + // Start the job + handle, net, err := tr.driver.StartTask(taskConfig) + if err != nil { + tr.logger.Warn("driver start failed", "error", err) + return err + } + tr.network = net + + tr.localStateLock.Lock() + tr.localState.TaskHandle = handle + tr.localStateLock.Unlock() + + tr.updateDriverHandle(taskConfig.ID) + // Emit an event that we started + tr.UpdateState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) + return nil +} + +func (tr *TaskRunner) updateDriverHandle(taskID string) { + tr.handleLock.Lock() + defer tr.handleLock.Unlock() + tr.handle = &driverHandleImpl{ + driver: tr.driver, + net: tr.network, + taskID: taskID, + task: tr.Task(), + } +} + +// initDriver creates the driver for the task +/*func (tr *TaskRunner) initDriver() error { + // Create a task-specific event emitter callback to expose minimal + // state to drivers + //XXX Replace with EmitEvent -- no need for a shim + eventEmitter := func(m string, args ...interface{}) { + msg := fmt.Sprintf(m, args...) + tr.logger.Debug("driver event", "event", msg) + tr.EmitEvent(structs.NewTaskEvent(structs.TaskDriverMessage).SetDriverMessage(msg)) + } + + alloc := tr.Alloc() + driverCtx := driver.NewDriverContext( + alloc.Job.Name, + alloc.TaskGroup, + tr.taskName, + tr.allocID, + tr.clientConfig, // XXX Why does it need this + tr.clientConfig.Node, // XXX THIS I NEED TO FIX + tr.logger.StandardLogger(nil), // XXX Should pass this through + eventEmitter) + + driver, err := driver.NewDriver(tr.task.Driver, driverCtx) + if err != nil { + return err + } + + tr.driver = driver + return nil +}*/ + +// initDriver retrives the DriverPlugin from the plugin loader for this task +func (tr *TaskRunner) initDriver() error { + plugin, err := tr.pluginSingletonLoader.Dispense(tr.Task().Driver, base.PluginTypeDriver, tr.logger) + if err != nil { + return err + } + + // XXX need to be able to reattach to running drivers + driver, ok := plugin.Plugin().(drivers.DriverPlugin) + if !ok { + return fmt.Errorf("plugin loaded for driver %s does not implement DriverPlugin interface", tr.task.Driver) + } + + tr.driver = driver + + schema, err := tr.driver.TaskConfigSchema() + if err != nil { + return err + } + spec, _ := hclspec.Convert(schema) + tr.taskSchema = spec + + caps, err := tr.driver.Capabilities() + if err != nil { + return err + } + tr.driverCapabilities = caps + + return nil +} + +// handleDestroy kills the task handle. In the case that killing fails, +// handleDestroy will retry with an exponential backoff and will give up at a +// given limit. It returns whether the task was destroyed and the error +// associated with the last kill attempt. +func (tr *TaskRunner) handleDestroy(handle tinterfaces.DriverHandle) (destroyed bool, err error) { + // Cap the number of times we attempt to kill the task. + for i := 0; i < killFailureLimit; i++ { + if err = handle.Kill(); err != nil { + if err == drivers.ErrTaskNotFound { + tr.logger.Warn("couldn't find task to kill", "task_id", handle.ID()) + return true, nil + } + // Calculate the new backoff + backoff := (1 << (2 * uint64(i))) * killBackoffBaseline + if backoff > killBackoffLimit { + backoff = killBackoffLimit + } + + tr.logger.Error("failed to kill task", "backoff", backoff, "error", err) + time.Sleep(backoff) + } else { + // Kill was successful + return true, nil + } + } + return +} + +// persistLocalState persists local state to disk synchronously. +func (tr *TaskRunner) persistLocalState() error { + tr.localStateLock.Lock() + defer tr.localStateLock.Unlock() + + return tr.stateDB.PutTaskRunnerLocalState(tr.allocID, tr.taskName, tr.localState) +} + +// buildID builds a consistent unique ID for the task from the alloc ID, task name and restart attempt +func (tr *TaskRunner) buildID() string { + return fmt.Sprintf("%s/%s/%d", tr.allocID, tr.taskName, tr.restartTracker.GetCount()) +} + +// XXX If the objects don't exists since the client shutdown before the task +// runner ever saved state, then we should treat it as a new task runner and not +// return an error +// +// Restore task runner state. Called by AllocRunner.Restore after NewTaskRunner +// but before Run so no locks need to be acquired. +func (tr *TaskRunner) Restore() error { + ls, ts, err := tr.stateDB.GetTaskRunnerState(tr.allocID, tr.taskName) + if err != nil { + return err + } + + tr.localState = ls + tr.state = ts + return nil +} + +// UpdateState sets the task runners allocation state and triggers a server +// update. +func (tr *TaskRunner) UpdateState(state string, event *structs.TaskEvent) { + tr.logger.Debug("setting task state", "state", state, "event", event.Type) + // Update the local state + stateCopy := tr.setStateLocal(state, event) + + // Notify the alloc runner of the transition + tr.stateUpdater.TaskStateUpdated(tr.taskName, stateCopy) +} + +// setStateLocal updates the local in-memory state, persists a copy to disk and returns a +// copy of the task's state. +func (tr *TaskRunner) setStateLocal(state string, event *structs.TaskEvent) *structs.TaskState { + tr.stateLock.Lock() + defer tr.stateLock.Unlock() + + //XXX REMOVE ME AFTER TESTING + if state == "" { + panic("UpdateState must not be called with an empty state") + } + + // Update the task state + oldState := tr.state.State + taskState := tr.state + taskState.State = state + + // Append the event + tr.appendEvent(event) + + // Handle the state transition. + switch state { + case structs.TaskStateRunning: + // Capture the start time if it is just starting + if oldState != structs.TaskStateRunning { + taskState.StartedAt = time.Now().UTC() + if !tr.clientConfig.DisableTaggedMetrics { + metrics.IncrCounterWithLabels([]string{"client", "allocs", "running"}, 1, tr.baseLabels) + } + //if r.config.BackwardsCompatibleMetrics { + //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "running"}, 1) + //} + } + case structs.TaskStateDead: + // Capture the finished time if not already set + if taskState.FinishedAt.IsZero() { + taskState.FinishedAt = time.Now().UTC() + } + + // Emitting metrics to indicate task complete and failures + if taskState.Failed { + if !tr.clientConfig.DisableTaggedMetrics { + metrics.IncrCounterWithLabels([]string{"client", "allocs", "failed"}, 1, tr.baseLabels) + } + //if r.config.BackwardsCompatibleMetrics { + //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "failed"}, 1) + //} + } else { + if !tr.clientConfig.DisableTaggedMetrics { + metrics.IncrCounterWithLabels([]string{"client", "allocs", "complete"}, 1, tr.baseLabels) + } + //if r.config.BackwardsCompatibleMetrics { + //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "complete"}, 1) + //} + } + } + + // Persist the state and event + if err := tr.stateDB.PutTaskState(tr.allocID, tr.taskName, taskState); err != nil { + // Only a warning because the next event/state-transition will + // try to persist it again. + tr.logger.Error("error persisting task state", "error", err, "event", event, "state", state) + } + + return tr.state.Copy() +} + +// EmitEvent appends a new TaskEvent to this task's TaskState. The actual +// TaskState.State (pending, running, dead) is not changed. Use UpdateState to +// transition states. +// Events are persisted locally and sent to the server, but errors are simply +// logged. Use AppendEvent to simply add a new event. +func (tr *TaskRunner) EmitEvent(event *structs.TaskEvent) { + tr.stateLock.Lock() + defer tr.stateLock.Unlock() + + tr.appendEvent(event) + + if err := tr.stateDB.PutTaskState(tr.allocID, tr.taskName, tr.state); err != nil { + // Only a warning because the next event/state-transition will + // try to persist it again. + tr.logger.Warn("error persisting event", "error", err, "event", event) + } + + // Notify the alloc runner of the event + tr.stateUpdater.TaskStateUpdated(tr.taskName, tr.state.Copy()) +} + +// AppendEvent appends a new TaskEvent to this task's TaskState. The actual +// TaskState.State (pending, running, dead) is not changed. Use UpdateState to +// transition states. +// Events are persisted locally and errors are simply logged. Use EmitEvent +// also update AllocRunner. +func (tr *TaskRunner) AppendEvent(event *structs.TaskEvent) { + tr.stateLock.Lock() + defer tr.stateLock.Unlock() + + tr.appendEvent(event) + + if err := tr.stateDB.PutTaskState(tr.allocID, tr.taskName, tr.state); err != nil { + // Only a warning because the next event/state-transition will + // try to persist it again. + tr.logger.Warn("error persisting event", "error", err, "event", event) + } +} + +// appendEvent to task's event slice. Caller must acquire stateLock. +func (tr *TaskRunner) appendEvent(event *structs.TaskEvent) error { + // Ensure the event is populated with human readable strings + event.PopulateEventDisplayMessage() + + // Propogate failure from event to task state + if event.FailsTask { + tr.state.Failed = true + } + + // XXX This seems like a super awkward spot for this? Why not shouldRestart? + // Update restart metrics + if event.Type == structs.TaskRestarting { + if !tr.clientConfig.DisableTaggedMetrics { + metrics.IncrCounterWithLabels([]string{"client", "allocs", "restart"}, 1, tr.baseLabels) + } + //if r.config.BackwardsCompatibleMetrics { + //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "restart"}, 1) + //} + tr.state.Restarts++ + tr.state.LastRestart = time.Unix(0, event.Time) + } + + // Append event to slice + appendTaskEvent(tr.state, event) + + return nil +} + +// WaitCh is closed when TaskRunner.Run exits. +func (tr *TaskRunner) WaitCh() <-chan struct{} { + return tr.waitCh +} + +// Update the running allocation with a new version received from the server. +// Calls Update hooks asynchronously with Run(). +// +// This method is safe for calling concurrently with Run() and does not modify +// the passed in allocation. +func (tr *TaskRunner) Update(update *structs.Allocation) { + // Update tr.alloc + tr.setAlloc(update) + + // Trigger update hooks + tr.triggerUpdateHooks() +} + +// triggerUpdate if there isn't already an update pending. Should be called +// instead of calling updateHooks directly to serialize runs of update hooks. +// TaskRunner state should be updated prior to triggering update hooks. +// +// Does not block. +func (tr *TaskRunner) triggerUpdateHooks() { + select { + case tr.triggerUpdateCh <- struct{}{}: + default: + // already an update hook pending + } +} + +// LatestResourceUsage returns the last resource utilization datapoint +// collected. May return nil if the task is not running or no resource +// utilization has been collected yet. +func (tr *TaskRunner) LatestResourceUsage() *cstructs.TaskResourceUsage { + tr.resourceUsageLock.Lock() + ru := tr.resourceUsage + tr.resourceUsageLock.Unlock() + return ru +} + +// UpdateStats updates and emits the latest stats from the driver. +func (tr *TaskRunner) UpdateStats(ru *cstructs.TaskResourceUsage) { + tr.resourceUsageLock.Lock() + tr.resourceUsage = ru + tr.resourceUsageLock.Unlock() + if ru != nil { + tr.emitStats(ru) + } +} + +//TODO Remove Backwardscompat or use tr.Alloc()? +func (tr *TaskRunner) setGaugeForMemory(ru *cstructs.TaskResourceUsage) { + if !tr.clientConfig.DisableTaggedMetrics { + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "rss"}, + float32(ru.ResourceUsage.MemoryStats.RSS), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "rss"}, + float32(ru.ResourceUsage.MemoryStats.RSS), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "cache"}, + float32(ru.ResourceUsage.MemoryStats.Cache), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "swap"}, + float32(ru.ResourceUsage.MemoryStats.Swap), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "max_usage"}, + float32(ru.ResourceUsage.MemoryStats.MaxUsage), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "kernel_usage"}, + float32(ru.ResourceUsage.MemoryStats.KernelUsage), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "kernel_max_usage"}, + float32(ru.ResourceUsage.MemoryStats.KernelMaxUsage), tr.baseLabels) + } + + if tr.clientConfig.BackwardsCompatibleMetrics { + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "rss"}, float32(ru.ResourceUsage.MemoryStats.RSS)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "cache"}, float32(ru.ResourceUsage.MemoryStats.Cache)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "swap"}, float32(ru.ResourceUsage.MemoryStats.Swap)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "max_usage"}, float32(ru.ResourceUsage.MemoryStats.MaxUsage)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "kernel_usage"}, float32(ru.ResourceUsage.MemoryStats.KernelUsage)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "kernel_max_usage"}, float32(ru.ResourceUsage.MemoryStats.KernelMaxUsage)) + } +} + +//TODO Remove Backwardscompat or use tr.Alloc()? +func (tr *TaskRunner) setGaugeForCPU(ru *cstructs.TaskResourceUsage) { + if !tr.clientConfig.DisableTaggedMetrics { + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "total_percent"}, + float32(ru.ResourceUsage.CpuStats.Percent), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "system"}, + float32(ru.ResourceUsage.CpuStats.SystemMode), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "user"}, + float32(ru.ResourceUsage.CpuStats.UserMode), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "throttled_time"}, + float32(ru.ResourceUsage.CpuStats.ThrottledTime), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "throttled_periods"}, + float32(ru.ResourceUsage.CpuStats.ThrottledPeriods), tr.baseLabels) + metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "total_ticks"}, + float32(ru.ResourceUsage.CpuStats.TotalTicks), tr.baseLabels) + } + + if tr.clientConfig.BackwardsCompatibleMetrics { + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "total_percent"}, float32(ru.ResourceUsage.CpuStats.Percent)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "system"}, float32(ru.ResourceUsage.CpuStats.SystemMode)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "user"}, float32(ru.ResourceUsage.CpuStats.UserMode)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "throttled_time"}, float32(ru.ResourceUsage.CpuStats.ThrottledTime)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "throttled_periods"}, float32(ru.ResourceUsage.CpuStats.ThrottledPeriods)) + metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "total_ticks"}, float32(ru.ResourceUsage.CpuStats.TotalTicks)) + } +} + +// emitStats emits resource usage stats of tasks to remote metrics collector +// sinks +func (tr *TaskRunner) emitStats(ru *cstructs.TaskResourceUsage) { + if !tr.clientConfig.PublishAllocationMetrics { + return + } + + if ru.ResourceUsage.MemoryStats != nil { + tr.setGaugeForMemory(ru) + } + + if ru.ResourceUsage.CpuStats != nil { + tr.setGaugeForCPU(ru) + } +} + +// appendTaskEvent updates the task status by appending the new event. +func appendTaskEvent(state *structs.TaskState, event *structs.TaskEvent) { + const capacity = 10 + if state.Events == nil { + state.Events = make([]*structs.TaskEvent, 1, capacity) + state.Events[0] = event + return + } + + // If we hit capacity, then shift it. + if len(state.Events) == capacity { + old := state.Events + state.Events = make([]*structs.TaskEvent, 0, capacity) + state.Events = append(state.Events, old[1:]...) + } + + state.Events = append(state.Events, event) +} diff --git a/client/client.go b/client/client.go index 4f4b6f7e6..491045dab 100644 --- a/client/client.go +++ b/client/client.go @@ -297,7 +297,7 @@ func NewClient(cfg *config.Config, consulCatalog consul.CatalogAPI, consulServic c.configCopy = c.config.Copy() c.configLock.Unlock() - fingerprintManager := NewFingerprintManager(c.GetConfig, c.configCopy.Node, + fingerprintManager := NewFingerprintManager(c.configCopy.PluginSingletonLoader, c.GetConfig, c.configCopy.Node, c.shutdownCh, c.updateNodeFromFingerprint, c.updateNodeFromDriver, c.logger) @@ -761,14 +761,16 @@ func (c *Client) restoreState() error { for _, alloc := range allocs { c.configLock.RLock() - arConf := &allocrunner.Config{ - Alloc: alloc, - Logger: c.logger, - ClientConfig: c.config, - StateDB: c.stateDB, - StateUpdater: c, - Consul: c.consulService, - Vault: c.vaultClient, + arConf := &allocrunnerv2.Config{ + Alloc: alloc, + Logger: c.logger, + ClientConfig: c.config, + StateDB: c.stateDB, + StateUpdater: c, + Consul: c.consulService, + Vault: c.vaultClient, + PluginLoader: c.config.PluginLoader, + PluginSingletonLoader: c.config.PluginSingletonLoader, } c.configLock.RUnlock() @@ -1032,31 +1034,45 @@ func (c *Client) updateNodeFromFingerprint(response *cstructs.FingerprintRespons return c.configCopy.Node } -// updateNodeFromDriver receives either a fingerprint of the driver or its -// health and merges this into a single DriverInfo object -func (c *Client) updateNodeFromDriver(name string, fingerprint, health *structs.DriverInfo) *structs.Node { +// updateNodeFromDriver receives a DriverInfo struct for the driver and updates +// the node accordingly +func (c *Client) updateNodeFromDriver(name string, info *structs.DriverInfo) *structs.Node { c.configLock.Lock() defer c.configLock.Unlock() var hasChanged bool hadDriver := c.config.Node.Drivers[name] != nil - if fingerprint != nil { + if info != nil { if !hadDriver { // If the driver info has not yet been set, do that here hasChanged = true - c.config.Node.Drivers[name] = fingerprint - for attrName, newVal := range fingerprint.Attributes { + c.config.Node.Drivers[name] = info + for attrName, newVal := range info.Attributes { c.config.Node.Attributes[attrName] = newVal } } else { + oldVal := c.config.Node.Drivers[name] // The driver info has already been set, fix it up - if c.config.Node.Drivers[name].Detected != fingerprint.Detected { + if oldVal.Detected != info.Detected { hasChanged = true - c.config.Node.Drivers[name].Detected = fingerprint.Detected + c.config.Node.Drivers[name].Detected = info.Detected } - for attrName, newVal := range fingerprint.Attributes { + if oldVal.Healthy != info.Healthy || oldVal.HealthDescription != info.HealthDescription { + hasChanged = true + if info.HealthDescription != "" { + event := &structs.NodeEvent{ + Subsystem: "Driver", + Message: info.HealthDescription, + Timestamp: time.Now(), + Details: map[string]string{"driver": name}, + } + c.triggerNodeEvent(event) + } + } + + for attrName, newVal := range info.Attributes { oldVal := c.config.Node.Drivers[name].Attributes[attrName] if oldVal == newVal { continue @@ -1075,48 +1091,12 @@ func (c *Client) updateNodeFromDriver(name string, fingerprint, health *structs. // We maintain the driver enabled attribute until all drivers expose // their attributes as DriverInfo driverName := fmt.Sprintf("driver.%s", name) - if fingerprint.Detected { + if info.Detected { c.config.Node.Attributes[driverName] = "1" } else { delete(c.config.Node.Attributes, driverName) } - } - if health != nil { - if !hadDriver { - hasChanged = true - if info, ok := c.config.Node.Drivers[name]; !ok { - c.config.Node.Drivers[name] = health - } else { - info.MergeHealthCheck(health) - } - } else { - oldVal := c.config.Node.Drivers[name] - if health.HealthCheckEquals(oldVal) { - // Make sure we accurately reflect the last time a health check has been - // performed for the driver. - oldVal.UpdateTime = health.UpdateTime - } else { - hasChanged = true - - // Only emit an event if the health status has changed after node - // initial startup (the health description will not get populated until - // a health check has run; the initial status is equal to whether the - // node is detected or not). - if health.Healthy != oldVal.Healthy && health.HealthDescription != "" { - event := &structs.NodeEvent{ - Subsystem: "Driver", - Message: health.HealthDescription, - Timestamp: time.Now(), - Details: map[string]string{"driver": name}, - } - c.triggerNodeEvent(event) - } - - // Update the node with the latest information - c.config.Node.Drivers[name].MergeHealthCheck(health) - } - } } if hasChanged { @@ -1931,15 +1911,17 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error // The long term fix is to pass in the config and node separately and then // we don't have to do a copy. c.configLock.RLock() - arConf := &allocrunner.Config{ - Alloc: alloc, - Logger: c.logger, - ClientConfig: c.config, - StateDB: c.stateDB, - Consul: c.consulService, - Vault: c.vaultClient, - StateUpdater: c, - PrevAllocWatcher: prevAllocWatcher, + arConf := &allocrunnerv2.Config{ + Alloc: alloc, + Logger: c.logger, + ClientConfig: c.config, + StateDB: c.stateDB, + Consul: c.consulService, + Vault: c.vaultClient, + StateUpdater: c, + PrevAllocWatcher: prevAllocWatcher, + PluginLoader: c.config.PluginLoader, + PluginSingletonLoader: c.config.PluginSingletonLoader, } c.configLock.RUnlock() diff --git a/client/fingerprint_manager.go b/client/fingerprint_manager.go index 8e007fb05..7254235c1 100644 --- a/client/fingerprint_manager.go +++ b/client/fingerprint_manager.go @@ -1,25 +1,29 @@ package client import ( + "context" "fmt" "sync" "time" log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/config" - "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/fingerprint" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/drivers" + "github.com/hashicorp/nomad/plugins/shared/loader" ) // FingerprintManager runs a client fingerprinters on a continuous basis, and // updates the client when the node has changed type FingerprintManager struct { - getConfig func() *config.Config - node *structs.Node - nodeLock sync.Mutex - shutdownCh chan struct{} + singletonLoader loader.PluginCatalog + getConfig func() *config.Config + node *structs.Node + nodeLock sync.Mutex + shutdownCh chan struct{} // updateNodeAttributes is a callback to the client to update the state of its // associated node @@ -27,20 +31,23 @@ type FingerprintManager struct { // updateNodeFromDriver is a callback to the client to update the state of a // specific driver for the node - updateNodeFromDriver func(string, *structs.DriverInfo, *structs.DriverInfo) *structs.Node + updateNodeFromDriver func(string, *structs.DriverInfo) *structs.Node logger log.Logger } // NewFingerprintManager is a constructor that creates and returns an instance // of FingerprintManager -func NewFingerprintManager(getConfig func() *config.Config, +func NewFingerprintManager( + singletonLoader loader.PluginCatalog, + getConfig func() *config.Config, node *structs.Node, shutdownCh chan struct{}, updateNodeAttributes func(*cstructs.FingerprintResponse) *structs.Node, - updateNodeFromDriver func(string, *structs.DriverInfo, *structs.DriverInfo) *structs.Node, + updateNodeFromDriver func(string, *structs.DriverInfo) *structs.Node, logger log.Logger) *FingerprintManager { return &FingerprintManager{ + singletonLoader: singletonLoader, getConfig: getConfig, updateNodeAttributes: updateNodeAttributes, updateNodeFromDriver: updateNodeFromDriver, @@ -111,8 +118,13 @@ func (fp *FingerprintManager) Run() error { var availDrivers []string var skippedDrivers []string + var registeredDrivers []string - for name := range driver.BuiltinDrivers { + for _, pl := range fp.singletonLoader.Catalog()[base.PluginTypeDriver] { + registeredDrivers = append(registeredDrivers, pl.Name) + } + + for _, name := range registeredDrivers { // Skip fingerprinting drivers that are not in the whitelist if it is // enabled. if _, ok := whitelistDrivers[name]; whitelistDriversEnabled && !ok { @@ -173,35 +185,41 @@ func (fm *FingerprintManager) setupFingerprinters(fingerprints []string) error { // setupDrivers is used to fingerprint the node to see if these drivers are // supported -func (fm *FingerprintManager) setupDrivers(drivers []string) error { +func (fm *FingerprintManager) setupDrivers(driverNames []string) error { //TODO(alex,hclog) Update fingerprinters to hclog var availDrivers []string - driverCtx := driver.NewDriverContext("", "", "", "", fm.getConfig(), fm.getNode(), fm.logger.StandardLogger(&log.StandardLoggerOptions{InferLevels: true}), nil) - for _, name := range drivers { - - d, err := driver.NewDriver(name, driverCtx) + for _, name := range driverNames { + // TODO: driver reattach + plug, err := fm.singletonLoader.Dispense(name, base.PluginTypeDriver, fm.logger) if err != nil { return err } + driver, ok := plug.Plugin().(drivers.DriverPlugin) + if !ok { + return fmt.Errorf("registered driver plugin %q does not implement DriverPlugin interface") + } + // Pass true for whether the health check is periodic here, so that the // fingerprinter will not set the initial health check status (this is set // below, with an empty health status so that a node event is not // triggered) // Later, the periodic health checker will update this value for drivers // where health checks are enabled. - detected, err := fm.fingerprintDriver(name, d, true) + ctx, cancel := context.WithCancel(context.Background()) + fingerCh, err := driver.Fingerprint(ctx) if err != nil { - fm.logger.Debug("error fingerprinting driver", "error", err, "fingerprinter", name) return err } + finger := <-fingerCh + // Start a periodic watcher to detect changes to a drivers health and // attributes. - go fm.watchDriver(d, name) + go fm.watchDriverFingerprint(fingerCh, name, cancel) // Log the fingerprinters which have been applied - if detected { + if finger.Health != drivers.HealthStateUndetected { availDrivers = append(availDrivers, name) } } @@ -258,154 +276,26 @@ func (fm *FingerprintManager) fingerprint(name string, f fingerprint.Fingerprint // watchDrivers facilitates the different periods between fingerprint and // health checking a driver -func (fm *FingerprintManager) watchDriver(d driver.Driver, name string) { - var fingerprintTicker, healthTicker <-chan time.Time - - // Determine whether the fingerprinter is periodic and health checking - isPeriodic, fingerprintPeriod := d.Periodic() - hc, isHealthCheck := d.(fingerprint.HealthCheck) - - // Nothing to do since the state of this driver will never change - if !isPeriodic && !isHealthCheck { - return - } - - // Setup the required tickers - if isPeriodic { - ticker := time.NewTicker(fingerprintPeriod) - fingerprintTicker = ticker.C - defer ticker.Stop() - fm.logger.Debug("periodically fingerprinting driver", "name", name, "period", fingerprintPeriod) - } - - var isHealthCheckPeriodic bool - if isHealthCheck { - // Determine the interval at which to health check - req := &cstructs.HealthCheckIntervalRequest{} - var healthCheckResp cstructs.HealthCheckIntervalResponse - - if err := hc.GetHealthCheckInterval(req, &healthCheckResp); err != nil { - fm.logger.Error("error getting health check interval for driver", "error", err, "driver", name) - } else if healthCheckResp.Eligible { - isHealthCheckPeriodic = true - ticker := time.NewTicker(healthCheckResp.Period) - healthTicker = ticker.C - defer ticker.Stop() - fm.logger.Debug("periodically health checking driver", "driver", name, "period", healthCheckResp.Period) - } - } - - driverEverDetected := false +func (fm *FingerprintManager) watchDriverFingerprint(fpChan <-chan *drivers.Fingerprint, name string, cancel context.CancelFunc) { for { select { case <-fm.shutdownCh: + cancel() return - case <-fingerprintTicker: - if _, err := fm.fingerprintDriver(name, d, isHealthCheckPeriodic); err != nil { - fm.logger.Debug("error periodic fingerprinting driver", "error", err, "driver", name) + case fp := <-fpChan: + di := &structs.DriverInfo{ + Attributes: fp.Attributes, + Detected: fp.Health != drivers.HealthStateUndetected, + Healthy: fp.Health == drivers.HealthStateHealthy, + HealthDescription: fp.HealthDescription, + UpdateTime: time.Now(), } - fm.nodeLock.Lock() - driver, detected := fm.node.Drivers[name] - - // Memoize the driver detected status, so that we know whether to run the - // health check or not. - if detected && driver != nil && driver.Detected { - if !driverEverDetected { - driverEverDetected = true - } + n := fm.updateNodeFromDriver(name, di) + if n != nil { + fm.node = n } fm.nodeLock.Unlock() - case <-healthTicker: - if driverEverDetected { - if err := fm.runDriverHealthCheck(name, hc); err != nil { - fm.logger.Debug("error health checking", "error", err, "driver", name) - } - } } } } - -// fingerprintDriver is a temporary solution to move towards DriverInfo and -// away from annotating a node's attributes to demonstrate support for a -// particular driver. Takes the FingerprintResponse and converts it to the -// proper DriverInfo update and then sets the prefix attributes as well -func (fm *FingerprintManager) fingerprintDriver(name string, f fingerprint.Fingerprint, hasPeriodicHealthCheck bool) (bool, error) { - var response cstructs.FingerprintResponse - - fm.nodeLock.Lock() - - // Determine if the driver has been detected before. - originalNode, haveDriver := fm.node.Drivers[name] - firstDetection := !haveDriver - - // Determine if the driver is healthy - var driverIsHealthy bool - if haveDriver && originalNode.Healthy { - driverIsHealthy = true - } - - // Fingerprint the driver. - request := &cstructs.FingerprintRequest{Config: fm.getConfig(), Node: fm.node} - err := f.Fingerprint(request, &response) - fm.nodeLock.Unlock() - - if err != nil { - return false, err - } - - // Remove the health check attribute indicating the status of the driver, - // as the overall driver info object should indicate this. - delete(response.Attributes, fmt.Sprintf("driver.%s", name)) - - fingerprintInfo := &structs.DriverInfo{ - Attributes: response.Attributes, - Detected: response.Detected, - } - - // We set the health status based on the detection state of the driver if: - // * It is the first time we are fingerprinting the driver. This gives all - // drivers an initial health. - // * If the driver becomes undetected. This gives us an immediate unhealthy - // state and description when it transistions from detected and healthy to - // undetected. - // * If the driver does not have its own health checks. Then we always - // couple the states. - var healthInfo *structs.DriverInfo - if firstDetection || !hasPeriodicHealthCheck || !response.Detected && driverIsHealthy { - state := " " - if !response.Detected { - state = " not " - } - - healthInfo = &structs.DriverInfo{ - Healthy: response.Detected, - HealthDescription: fmt.Sprintf("Driver %s is%sdetected", name, state), - UpdateTime: time.Now(), - } - } - - if node := fm.updateNodeFromDriver(name, fingerprintInfo, healthInfo); node != nil { - fm.setNode(node) - } - - return response.Detected, nil -} - -// runDriverHealthCheck checks the health of the specified resource. -func (fm *FingerprintManager) runDriverHealthCheck(name string, hc fingerprint.HealthCheck) error { - request := &cstructs.HealthCheckRequest{} - var response cstructs.HealthCheckResponse - if err := hc.HealthCheck(request, &response); err != nil { - return err - } - - // Update the status of the node irregardless if there was an error- in the - // case of periodic health checks, an error will occur if a health check - // fails - if node := fm.updateNodeFromDriver(name, nil, response.Drivers[name]); node != nil { - fm.setNode(node) - } - - return nil -} diff --git a/command/agent/config.go b/command/agent/config.go index 470e563af..818ff430c 100644 --- a/command/agent/config.go +++ b/command/agent/config.go @@ -572,9 +572,7 @@ func DevConfig() *Config { } conf.Client.Options = map[string]string{ "driver.raw_exec.enable": "true", - } - conf.Client.Options = map[string]string{ - "driver.docker.volumes": "true", + "driver.docker.volumes": "true", } conf.Client.GCInterval = 10 * time.Minute conf.Client.GCDiskUsageThreshold = 99 diff --git a/command/agent/consul/script.go b/command/agent/consul/script.go index be922d87f..0caa3ef1e 100644 --- a/command/agent/consul/script.go +++ b/command/agent/consul/script.go @@ -8,7 +8,7 @@ import ( log "github.com/hashicorp/go-hclog" "github.com/hashicorp/consul/api" - "github.com/hashicorp/nomad/client/driver" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" "github.com/hashicorp/nomad/nomad/structs" ) @@ -39,7 +39,7 @@ type scriptCheck struct { id string check *structs.ServiceCheck - exec driver.ScriptExecutor + exec interfaces.ScriptExecutor agent heartbeater // lastCheckOk is true if the last check was ok; otherwise false @@ -52,7 +52,7 @@ type scriptCheck struct { // newScriptCheck creates a new scriptCheck. run() should be called once the // initial check is registered with Consul. func newScriptCheck(allocID, taskName, checkID string, check *structs.ServiceCheck, - exec driver.ScriptExecutor, agent heartbeater, logger log.Logger, + exec interfaces.ScriptExecutor, agent heartbeater, logger log.Logger, shutdownCh <-chan struct{}) *scriptCheck { logger = logger.ResetNamed("consul.checks").With("task", taskName, "alloc_id", allocID, "check", check.Name) @@ -93,9 +93,8 @@ func (s *scriptCheck) run() *scriptHandle { metrics.IncrCounter([]string{"client", "consul", "script_runs"}, 1) // Execute check script with timeout - execctx, cancel := context.WithTimeout(ctx, s.check.Timeout) - output, code, err := s.exec.Exec(execctx, s.check.Command, s.check.Args) - switch execctx.Err() { + output, code, err := s.exec.Exec(s.check.Timeout, s.check.Command, s.check.Args) + switch err { case context.Canceled: // check removed during execution; exit cancel() diff --git a/command/agent/consul/structs.go b/command/agent/consul/structs.go index a69b0740c..729134523 100644 --- a/command/agent/consul/structs.go +++ b/command/agent/consul/structs.go @@ -1,7 +1,7 @@ package consul import ( - "github.com/hashicorp/nomad/client/driver" + "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" ) @@ -26,13 +26,13 @@ type TaskServices struct { Networks structs.Networks // DriverExec is the script executor for the task's driver. - DriverExec driver.ScriptExecutor + DriverExec interfaces.ScriptExecutor // DriverNetwork is the network specified by the driver and may be nil. DriverNetwork *cstructs.DriverNetwork } -func NewTaskServices(alloc *structs.Allocation, task *structs.Task, restarter TaskRestarter, exec driver.ScriptExecutor, net *cstructs.DriverNetwork) *TaskServices { +func NewTaskServices(alloc *structs.Allocation, task *structs.Task, restarter TaskRestarter, exec interfaces.ScriptExecutor, net *cstructs.DriverNetwork) *TaskServices { ts := TaskServices{ AllocID: alloc.ID, Name: task.Name, diff --git a/command/agent/plugins.go b/command/agent/plugins.go index 6637d69f9..5d53129d2 100644 --- a/command/agent/plugins.go +++ b/command/agent/plugins.go @@ -6,6 +6,8 @@ import ( "github.com/hashicorp/nomad/plugins/shared/catalog" "github.com/hashicorp/nomad/plugins/shared/loader" "github.com/hashicorp/nomad/plugins/shared/singleton" + + _ "github.com/hashicorp/nomad/drivers/rawexec" ) // setupPlugins is used to setup the plugin loaders. diff --git a/drivers/rawexec/driver.go b/drivers/rawexec/driver.go index b9e04916b..2691e321c 100644 --- a/drivers/rawexec/driver.go +++ b/drivers/rawexec/driver.go @@ -12,14 +12,40 @@ import ( plugin "github.com/hashicorp/go-plugin" "github.com/hashicorp/nomad/client/driver/executor" dstructs "github.com/hashicorp/nomad/client/driver/structs" + cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/drivers/shared/eventer" "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/drivers" "github.com/hashicorp/nomad/plugins/drivers/utils" + "github.com/hashicorp/nomad/plugins/shared/catalog" "github.com/hashicorp/nomad/plugins/shared/hclspec" + "github.com/hashicorp/nomad/plugins/shared/loader" "golang.org/x/net/context" ) +func init() { + catalog.RegisterDeferredConfig(loader.PluginID{ + Name: pluginName, + PluginType: base.PluginTypeDriver, + }, &loader.InternalPluginConfig{ + Config: map[string]interface{}{}, + Factory: func(l hclog.Logger) interface{} { return NewRawExecDriver(l) }, + }, + func(opts map[string]string) (map[string]interface{}, error) { + fmt.Println(opts) + conf := map[string]interface{}{} + if v, ok := opts["driver.raw_exec.enable"]; ok && v == "true" { + conf["enabled"] = true + } + if v, ok := opts["driver.raw_exec.no_cgroups"]; ok && v == "true" { + conf["no_cgroups"] = true + } + + return conf, nil + }, + ) +} + const ( // pluginName is the name of the plugin pluginName = "raw_exec" @@ -53,7 +79,7 @@ var ( // a task within a job. It is returned in the TaskConfigSchema RPC taskConfigSpec = hclspec.NewObject(map[string]*hclspec.Spec{ "command": hclspec.NewAttr("command", "string", true), - "args": hclspec.NewAttr("command", "list(string)", false), + "args": hclspec.NewAttr("args", "list(string)", false), }) // capabilities is returned by the Capabilities RPC and indicates what @@ -61,7 +87,7 @@ var ( capabilities = &drivers.Capabilities{ SendSignals: true, Exec: true, - FSIsolation: drivers.FSIsolationNone, + FSIsolation: cstructs.FSIsolationNone, } ) @@ -96,16 +122,16 @@ type RawExecDriver struct { type Config struct { // NoCgroups tracks whether we should use a cgroup to manage the process // tree - NoCgroups bool `codec:"no_cgroups"` + NoCgroups bool `codec:"no_cgroups" cty:"no_cgroups"` // Enabled is set to true to enable the raw_exec driver - Enabled bool `codec:"enabled"` + Enabled bool `codec:"enabled" cty:"enabled"` } // TaskConfig is the driver configuration of a task within a job type TaskConfig struct { - Command string `codec:"command"` - Args []string `codec:"args"` + Command string `codec:"command" cty:"command"` + Args []string `codec:"args" cty:"args"` } // RawExecTaskState is the state which is encoded in the handle returned in @@ -247,16 +273,17 @@ func (r *RawExecDriver) RecoverTask(handle *drivers.TaskHandle) error { return nil } -func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) { +func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cstructs.DriverNetwork, error) { if _, ok := r.tasks.Get(cfg.ID); ok { - return nil, fmt.Errorf("task with ID %q already started", cfg.ID) + return nil, nil, fmt.Errorf("task with ID %q already started", cfg.ID) } var driverConfig TaskConfig if err := cfg.DecodeDriverConfig(&driverConfig); err != nil { - return nil, fmt.Errorf("failed to decode driver config: %v", err) + return nil, nil, fmt.Errorf("failed to decode driver config: %v", err) } + r.logger.Info("starting task", "driver_cfg", hclog.Fmt("%+v", driverConfig)) handle := drivers.NewTaskHandle(pluginName) handle.Config = cfg @@ -269,7 +296,7 @@ func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, // TODO: best way to pass port ranges in from client config exec, pluginClient, err := utils.CreateExecutor(os.Stderr, hclog.Debug, 14000, 14512, executorConfig) if err != nil { - return nil, fmt.Errorf("failed to create executor: %v", err) + return nil, nil, fmt.Errorf("failed to create executor: %v", err) } execCmd := &executor.ExecCommand{ @@ -286,7 +313,7 @@ func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, ps, err := exec.Launch(execCmd) if err != nil { pluginClient.Kill() - return nil, fmt.Errorf("failed to launch command with executor: %v", err) + return nil, nil, fmt.Errorf("failed to launch command with executor: %v", err) } h := &rawExecTaskHandle{ @@ -310,12 +337,12 @@ func (r *RawExecDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, r.logger.Error("failed to start task, error setting driver state", "error", err) exec.Shutdown("", 0) pluginClient.Kill() - return nil, fmt.Errorf("failed to set driver state: %v", err) + return nil, nil, fmt.Errorf("failed to set driver state: %v", err) } r.tasks.Set(cfg.ID, h) go h.run() - return handle, nil + return handle, nil, nil } func (r *RawExecDriver) WaitTask(ctx context.Context, taskID string) (<-chan *drivers.ExitResult, error) { @@ -418,23 +445,13 @@ func (r *RawExecDriver) InspectTask(taskID string) (*drivers.TaskStatus, error) return status, nil } -func (r *RawExecDriver) TaskStats(taskID string) (*drivers.TaskStats, error) { +func (r *RawExecDriver) TaskStats(taskID string) (*cstructs.TaskResourceUsage, error) { handle, ok := r.tasks.Get(taskID) if !ok { return nil, drivers.ErrTaskNotFound } - stats, err := handle.exec.Stats() - if err != nil { - return nil, err - } - - return &drivers.TaskStats{ - ID: handle.task.ID, - Timestamp: stats.Timestamp, - AggResourceUsage: stats.ResourceUsage, - ResourceUsageByPid: stats.Pids, - }, nil + return handle.exec.Stats() } func (r *RawExecDriver) TaskEvents(ctx context.Context) (<-chan *drivers.TaskEvent, error) { diff --git a/plugins/drivers/client.go b/plugins/drivers/client.go index 756bbaf05..10de472cc 100644 --- a/plugins/drivers/client.go +++ b/plugins/drivers/client.go @@ -8,6 +8,7 @@ import ( "github.com/golang/protobuf/ptypes" hclog "github.com/hashicorp/go-hclog" + cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/drivers/proto" "github.com/hashicorp/nomad/plugins/shared/hclspec" @@ -49,13 +50,13 @@ func (d *driverPluginClient) Capabilities() (*Capabilities, error) { switch resp.Capabilities.FsIsolation { case proto.DriverCapabilities_NONE: - caps.FSIsolation = FSIsolationNone + caps.FSIsolation = cstructs.FSIsolationNone case proto.DriverCapabilities_CHROOT: - caps.FSIsolation = FSIsolationChroot + caps.FSIsolation = cstructs.FSIsolationChroot case proto.DriverCapabilities_IMAGE: - caps.FSIsolation = FSIsolationImage + caps.FSIsolation = cstructs.FSIsolationImage default: - caps.FSIsolation = FSIsolationNone + caps.FSIsolation = cstructs.FSIsolationNone } } @@ -111,17 +112,29 @@ func (d *driverPluginClient) RecoverTask(h *TaskHandle) error { // StartTask starts execution of a task with the given TaskConfig. A TaskHandle // is returned to the caller that can be used to recover state of the task, // should the driver crash or exit prematurely. -func (d *driverPluginClient) StartTask(c *TaskConfig) (*TaskHandle, error) { +func (d *driverPluginClient) StartTask(c *TaskConfig) (*TaskHandle, *cstructs.DriverNetwork, error) { req := &proto.StartTaskRequest{ Task: taskConfigToProto(c), } resp, err := d.client.StartTask(context.Background(), req) if err != nil { - return nil, err + return nil, nil, err } - return taskHandleFromProto(resp.Handle), nil + var net *cstructs.DriverNetwork + if resp.NetworkOverride != nil { + net = &cstructs.DriverNetwork{ + PortMap: map[string]int{}, + IP: resp.NetworkOverride.Addr, + AutoAdvertise: resp.NetworkOverride.AutoAdvertise, + } + for k, v := range resp.NetworkOverride.PortMap { + net.PortMap[k] = int(v) + } + } + + return taskHandleFromProto(resp.Handle), net, nil } // WaitTask returns a channel that will have an ExitResult pushed to it once when the task @@ -201,18 +214,21 @@ func (d *driverPluginClient) InspectTask(taskID string) (*TaskStatus, error) { status.DriverAttributes = resp.Driver.Attributes } if resp.NetworkOverride != nil { - status.NetworkOverride = &NetworkOverride{ - PortMap: resp.NetworkOverride.PortMap, - Addr: resp.NetworkOverride.Addr, + status.NetworkOverride = &cstructs.DriverNetwork{ + PortMap: map[string]int{}, + IP: resp.NetworkOverride.Addr, AutoAdvertise: resp.NetworkOverride.AutoAdvertise, } + for k, v := range resp.NetworkOverride.PortMap { + status.NetworkOverride.PortMap[k] = int(v) + } } return status, nil } // TaskStats returns resource usage statistics for the task -func (d *driverPluginClient) TaskStats(taskID string) (*TaskStats, error) { +func (d *driverPluginClient) TaskStats(taskID string) (*cstructs.TaskResourceUsage, error) { req := &proto.TaskStatsRequest{TaskId: taskID} resp, err := d.client.TaskStats(context.Background(), req) diff --git a/plugins/drivers/driver.go b/plugins/drivers/driver.go index fec8804d4..7d503f451 100644 --- a/plugins/drivers/driver.go +++ b/plugins/drivers/driver.go @@ -8,8 +8,12 @@ import ( "github.com/hashicorp/nomad/client/allocdir" cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/shared/hclspec" + "github.com/zclconf/go-cty/cty" + "github.com/zclconf/go-cty/cty/gocty" + "github.com/zclconf/go-cty/cty/msgpack" "golang.org/x/net/context" ) @@ -25,12 +29,12 @@ type DriverPlugin interface { Fingerprint(context.Context) (<-chan *Fingerprint, error) RecoverTask(*TaskHandle) error - StartTask(*TaskConfig) (*TaskHandle, error) + StartTask(*TaskConfig) (*TaskHandle, *cstructs.DriverNetwork, error) WaitTask(ctx context.Context, taskID string) (<-chan *ExitResult, error) StopTask(taskID string, timeout time.Duration, signal string) error DestroyTask(taskID string, force bool) error InspectTask(taskID string) (*TaskStatus, error) - TaskStats(taskID string) (*TaskStats, error) + TaskStats(taskID string) (*cstructs.TaskResourceUsage, error) TaskEvents(context.Context) (<-chan *TaskEvent, error) SignalTask(taskID string, signal string) error @@ -89,14 +93,14 @@ type Capabilities struct { Exec bool //FSIsolation indicates what kind of filesystem isolation the driver supports. - FSIsolation FSIsolation + FSIsolation cstructs.FSIsolation } type TaskConfig struct { ID string Name string Env map[string]string - Resources Resources + Resources *Resources Devices []DeviceConfig Mounts []MountConfig User string @@ -129,14 +133,35 @@ func (tc *TaskConfig) TaskDir() *allocdir.TaskDir { } func (tc *TaskConfig) DecodeDriverConfig(t interface{}) error { - return base.MsgPackDecode(tc.rawDriverConfig, t) + ty, err := gocty.ImpliedType(t) + if err != nil { + return err + } + + val, err := msgpack.Unmarshal(tc.rawDriverConfig, ty) + if err != nil { + return err + } + + return gocty.FromCtyValue(val, t) } -func (tc *TaskConfig) EncodeDriverConfig(t interface{}) error { - return base.MsgPackEncode(&tc.rawDriverConfig, t) +func (tc *TaskConfig) EncodeDriverConfig(val cty.Value) error { + data, err := msgpack.Marshal(val, val.Type()) + if err != nil { + return err + } + + tc.rawDriverConfig = data + return nil } type Resources struct { + NomadResources *structs.Resources + LinuxResources *LinuxResources +} + +type LinuxResources struct { CPUPeriod int64 CPUQuota int64 CPUShares int64 @@ -166,12 +191,6 @@ const ( type TaskState string -type NetworkOverride struct { - PortMap map[string]int32 - Addr string - AutoAdvertise bool -} - type ExitResult struct { ExitCode int Signal int @@ -191,14 +210,7 @@ type TaskStatus struct { CompletedAt time.Time ExitResult *ExitResult DriverAttributes map[string]string - NetworkOverride *NetworkOverride -} - -type TaskStats struct { - ID string - Timestamp int64 - AggResourceUsage *cstructs.ResourceUsage - ResourceUsageByPid map[string]*cstructs.ResourceUsage + NetworkOverride *cstructs.DriverNetwork } type TaskEvent struct { diff --git a/plugins/drivers/proto/driver.pb.go b/plugins/drivers/proto/driver.pb.go index ee4667d47..7f60d3e07 100644 --- a/plugins/drivers/proto/driver.pb.go +++ b/plugins/drivers/proto/driver.pb.go @@ -49,7 +49,7 @@ func (x TaskState) String() string { return proto.EnumName(TaskState_name, int32(x)) } func (TaskState) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{0} + return fileDescriptor_driver_60d7e28aa66468bf, []int{0} } type FingerprintResponse_HealthState int32 @@ -75,7 +75,7 @@ func (x FingerprintResponse_HealthState) String() string { return proto.EnumName(FingerprintResponse_HealthState_name, int32(x)) } func (FingerprintResponse_HealthState) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{5, 0} + return fileDescriptor_driver_60d7e28aa66468bf, []int{5, 0} } type StartTaskResponse_Result int32 @@ -101,7 +101,7 @@ func (x StartTaskResponse_Result) String() string { return proto.EnumName(StartTaskResponse_Result_name, int32(x)) } func (StartTaskResponse_Result) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{9, 0} + return fileDescriptor_driver_60d7e28aa66468bf, []int{9, 0} } type DriverCapabilities_FSIsolation int32 @@ -127,7 +127,7 @@ func (x DriverCapabilities_FSIsolation) String() string { return proto.EnumName(DriverCapabilities_FSIsolation_name, int32(x)) } func (DriverCapabilities_FSIsolation) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{25, 0} + return fileDescriptor_driver_60d7e28aa66468bf, []int{25, 0} } type CPUUsage_Fields int32 @@ -162,7 +162,7 @@ func (x CPUUsage_Fields) String() string { return proto.EnumName(CPUUsage_Fields_name, int32(x)) } func (CPUUsage_Fields) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{41, 0} + return fileDescriptor_driver_60d7e28aa66468bf, []int{41, 0} } type MemoryUsage_Fields int32 @@ -194,7 +194,7 @@ func (x MemoryUsage_Fields) String() string { return proto.EnumName(MemoryUsage_Fields_name, int32(x)) } func (MemoryUsage_Fields) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{42, 0} + return fileDescriptor_driver_60d7e28aa66468bf, []int{42, 0} } type TaskConfigSchemaRequest struct { @@ -207,7 +207,7 @@ func (m *TaskConfigSchemaRequest) Reset() { *m = TaskConfigSchemaRequest func (m *TaskConfigSchemaRequest) String() string { return proto.CompactTextString(m) } func (*TaskConfigSchemaRequest) ProtoMessage() {} func (*TaskConfigSchemaRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{0} + return fileDescriptor_driver_60d7e28aa66468bf, []int{0} } func (m *TaskConfigSchemaRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskConfigSchemaRequest.Unmarshal(m, b) @@ -239,7 +239,7 @@ func (m *TaskConfigSchemaResponse) Reset() { *m = TaskConfigSchemaRespon func (m *TaskConfigSchemaResponse) String() string { return proto.CompactTextString(m) } func (*TaskConfigSchemaResponse) ProtoMessage() {} func (*TaskConfigSchemaResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{1} + return fileDescriptor_driver_60d7e28aa66468bf, []int{1} } func (m *TaskConfigSchemaResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskConfigSchemaResponse.Unmarshal(m, b) @@ -276,7 +276,7 @@ func (m *CapabilitiesRequest) Reset() { *m = CapabilitiesRequest{} } func (m *CapabilitiesRequest) String() string { return proto.CompactTextString(m) } func (*CapabilitiesRequest) ProtoMessage() {} func (*CapabilitiesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{2} + return fileDescriptor_driver_60d7e28aa66468bf, []int{2} } func (m *CapabilitiesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CapabilitiesRequest.Unmarshal(m, b) @@ -311,7 +311,7 @@ func (m *CapabilitiesResponse) Reset() { *m = CapabilitiesResponse{} } func (m *CapabilitiesResponse) String() string { return proto.CompactTextString(m) } func (*CapabilitiesResponse) ProtoMessage() {} func (*CapabilitiesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{3} + return fileDescriptor_driver_60d7e28aa66468bf, []int{3} } func (m *CapabilitiesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CapabilitiesResponse.Unmarshal(m, b) @@ -348,7 +348,7 @@ func (m *FingerprintRequest) Reset() { *m = FingerprintRequest{} } func (m *FingerprintRequest) String() string { return proto.CompactTextString(m) } func (*FingerprintRequest) ProtoMessage() {} func (*FingerprintRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{4} + return fileDescriptor_driver_60d7e28aa66468bf, []int{4} } func (m *FingerprintRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_FingerprintRequest.Unmarshal(m, b) @@ -391,7 +391,7 @@ func (m *FingerprintResponse) Reset() { *m = FingerprintResponse{} } func (m *FingerprintResponse) String() string { return proto.CompactTextString(m) } func (*FingerprintResponse) ProtoMessage() {} func (*FingerprintResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{5} + return fileDescriptor_driver_60d7e28aa66468bf, []int{5} } func (m *FingerprintResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_FingerprintResponse.Unmarshal(m, b) @@ -446,7 +446,7 @@ func (m *RecoverTaskRequest) Reset() { *m = RecoverTaskRequest{} } func (m *RecoverTaskRequest) String() string { return proto.CompactTextString(m) } func (*RecoverTaskRequest) ProtoMessage() {} func (*RecoverTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{6} + return fileDescriptor_driver_60d7e28aa66468bf, []int{6} } func (m *RecoverTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecoverTaskRequest.Unmarshal(m, b) @@ -490,7 +490,7 @@ func (m *RecoverTaskResponse) Reset() { *m = RecoverTaskResponse{} } func (m *RecoverTaskResponse) String() string { return proto.CompactTextString(m) } func (*RecoverTaskResponse) ProtoMessage() {} func (*RecoverTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{7} + return fileDescriptor_driver_60d7e28aa66468bf, []int{7} } func (m *RecoverTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecoverTaskResponse.Unmarshal(m, b) @@ -522,7 +522,7 @@ func (m *StartTaskRequest) Reset() { *m = StartTaskRequest{} } func (m *StartTaskRequest) String() string { return proto.CompactTextString(m) } func (*StartTaskRequest) ProtoMessage() {} func (*StartTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{8} + return fileDescriptor_driver_60d7e28aa66468bf, []int{8} } func (m *StartTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StartTaskRequest.Unmarshal(m, b) @@ -576,7 +576,7 @@ func (m *StartTaskResponse) Reset() { *m = StartTaskResponse{} } func (m *StartTaskResponse) String() string { return proto.CompactTextString(m) } func (*StartTaskResponse) ProtoMessage() {} func (*StartTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{9} + return fileDescriptor_driver_60d7e28aa66468bf, []int{9} } func (m *StartTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StartTaskResponse.Unmarshal(m, b) @@ -636,7 +636,7 @@ func (m *WaitTaskRequest) Reset() { *m = WaitTaskRequest{} } func (m *WaitTaskRequest) String() string { return proto.CompactTextString(m) } func (*WaitTaskRequest) ProtoMessage() {} func (*WaitTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{10} + return fileDescriptor_driver_60d7e28aa66468bf, []int{10} } func (m *WaitTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WaitTaskRequest.Unmarshal(m, b) @@ -677,7 +677,7 @@ func (m *WaitTaskResponse) Reset() { *m = WaitTaskResponse{} } func (m *WaitTaskResponse) String() string { return proto.CompactTextString(m) } func (*WaitTaskResponse) ProtoMessage() {} func (*WaitTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{11} + return fileDescriptor_driver_60d7e28aa66468bf, []int{11} } func (m *WaitTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WaitTaskResponse.Unmarshal(m, b) @@ -729,7 +729,7 @@ func (m *StopTaskRequest) Reset() { *m = StopTaskRequest{} } func (m *StopTaskRequest) String() string { return proto.CompactTextString(m) } func (*StopTaskRequest) ProtoMessage() {} func (*StopTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{12} + return fileDescriptor_driver_60d7e28aa66468bf, []int{12} } func (m *StopTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StopTaskRequest.Unmarshal(m, b) @@ -780,7 +780,7 @@ func (m *StopTaskResponse) Reset() { *m = StopTaskResponse{} } func (m *StopTaskResponse) String() string { return proto.CompactTextString(m) } func (*StopTaskResponse) ProtoMessage() {} func (*StopTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{13} + return fileDescriptor_driver_60d7e28aa66468bf, []int{13} } func (m *StopTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StopTaskResponse.Unmarshal(m, b) @@ -814,7 +814,7 @@ func (m *DestroyTaskRequest) Reset() { *m = DestroyTaskRequest{} } func (m *DestroyTaskRequest) String() string { return proto.CompactTextString(m) } func (*DestroyTaskRequest) ProtoMessage() {} func (*DestroyTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{14} + return fileDescriptor_driver_60d7e28aa66468bf, []int{14} } func (m *DestroyTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DestroyTaskRequest.Unmarshal(m, b) @@ -858,7 +858,7 @@ func (m *DestroyTaskResponse) Reset() { *m = DestroyTaskResponse{} } func (m *DestroyTaskResponse) String() string { return proto.CompactTextString(m) } func (*DestroyTaskResponse) ProtoMessage() {} func (*DestroyTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{15} + return fileDescriptor_driver_60d7e28aa66468bf, []int{15} } func (m *DestroyTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DestroyTaskResponse.Unmarshal(m, b) @@ -890,7 +890,7 @@ func (m *InspectTaskRequest) Reset() { *m = InspectTaskRequest{} } func (m *InspectTaskRequest) String() string { return proto.CompactTextString(m) } func (*InspectTaskRequest) ProtoMessage() {} func (*InspectTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{16} + return fileDescriptor_driver_60d7e28aa66468bf, []int{16} } func (m *InspectTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectTaskRequest.Unmarshal(m, b) @@ -933,7 +933,7 @@ func (m *InspectTaskResponse) Reset() { *m = InspectTaskResponse{} } func (m *InspectTaskResponse) String() string { return proto.CompactTextString(m) } func (*InspectTaskResponse) ProtoMessage() {} func (*InspectTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{17} + return fileDescriptor_driver_60d7e28aa66468bf, []int{17} } func (m *InspectTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectTaskResponse.Unmarshal(m, b) @@ -986,7 +986,7 @@ func (m *TaskStatsRequest) Reset() { *m = TaskStatsRequest{} } func (m *TaskStatsRequest) String() string { return proto.CompactTextString(m) } func (*TaskStatsRequest) ProtoMessage() {} func (*TaskStatsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{18} + return fileDescriptor_driver_60d7e28aa66468bf, []int{18} } func (m *TaskStatsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStatsRequest.Unmarshal(m, b) @@ -1025,7 +1025,7 @@ func (m *TaskStatsResponse) Reset() { *m = TaskStatsResponse{} } func (m *TaskStatsResponse) String() string { return proto.CompactTextString(m) } func (*TaskStatsResponse) ProtoMessage() {} func (*TaskStatsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{19} + return fileDescriptor_driver_60d7e28aa66468bf, []int{19} } func (m *TaskStatsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStatsResponse.Unmarshal(m, b) @@ -1062,7 +1062,7 @@ func (m *TaskEventsRequest) Reset() { *m = TaskEventsRequest{} } func (m *TaskEventsRequest) String() string { return proto.CompactTextString(m) } func (*TaskEventsRequest) ProtoMessage() {} func (*TaskEventsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{20} + return fileDescriptor_driver_60d7e28aa66468bf, []int{20} } func (m *TaskEventsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskEventsRequest.Unmarshal(m, b) @@ -1096,7 +1096,7 @@ func (m *SignalTaskRequest) Reset() { *m = SignalTaskRequest{} } func (m *SignalTaskRequest) String() string { return proto.CompactTextString(m) } func (*SignalTaskRequest) ProtoMessage() {} func (*SignalTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{21} + return fileDescriptor_driver_60d7e28aa66468bf, []int{21} } func (m *SignalTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SignalTaskRequest.Unmarshal(m, b) @@ -1140,7 +1140,7 @@ func (m *SignalTaskResponse) Reset() { *m = SignalTaskResponse{} } func (m *SignalTaskResponse) String() string { return proto.CompactTextString(m) } func (*SignalTaskResponse) ProtoMessage() {} func (*SignalTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{22} + return fileDescriptor_driver_60d7e28aa66468bf, []int{22} } func (m *SignalTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SignalTaskResponse.Unmarshal(m, b) @@ -1177,7 +1177,7 @@ func (m *ExecTaskRequest) Reset() { *m = ExecTaskRequest{} } func (m *ExecTaskRequest) String() string { return proto.CompactTextString(m) } func (*ExecTaskRequest) ProtoMessage() {} func (*ExecTaskRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{23} + return fileDescriptor_driver_60d7e28aa66468bf, []int{23} } func (m *ExecTaskRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExecTaskRequest.Unmarshal(m, b) @@ -1234,7 +1234,7 @@ func (m *ExecTaskResponse) Reset() { *m = ExecTaskResponse{} } func (m *ExecTaskResponse) String() string { return proto.CompactTextString(m) } func (*ExecTaskResponse) ProtoMessage() {} func (*ExecTaskResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{24} + return fileDescriptor_driver_60d7e28aa66468bf, []int{24} } func (m *ExecTaskResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExecTaskResponse.Unmarshal(m, b) @@ -1293,7 +1293,7 @@ func (m *DriverCapabilities) Reset() { *m = DriverCapabilities{} } func (m *DriverCapabilities) String() string { return proto.CompactTextString(m) } func (*DriverCapabilities) ProtoMessage() {} func (*DriverCapabilities) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{25} + return fileDescriptor_driver_60d7e28aa66468bf, []int{25} } func (m *DriverCapabilities) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DriverCapabilities.Unmarshal(m, b) @@ -1368,7 +1368,7 @@ func (m *TaskConfig) Reset() { *m = TaskConfig{} } func (m *TaskConfig) String() string { return proto.CompactTextString(m) } func (*TaskConfig) ProtoMessage() {} func (*TaskConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{26} + return fileDescriptor_driver_60d7e28aa66468bf, []int{26} } func (m *TaskConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskConfig.Unmarshal(m, b) @@ -1479,7 +1479,7 @@ func (m *Resources) Reset() { *m = Resources{} } func (m *Resources) String() string { return proto.CompactTextString(m) } func (*Resources) ProtoMessage() {} func (*Resources) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{27} + return fileDescriptor_driver_60d7e28aa66468bf, []int{27} } func (m *Resources) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Resources.Unmarshal(m, b) @@ -1528,7 +1528,7 @@ func (m *RawResources) Reset() { *m = RawResources{} } func (m *RawResources) String() string { return proto.CompactTextString(m) } func (*RawResources) ProtoMessage() {} func (*RawResources) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{28} + return fileDescriptor_driver_60d7e28aa66468bf, []int{28} } func (m *RawResources) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RawResources.Unmarshal(m, b) @@ -1599,7 +1599,7 @@ func (m *NetworkResource) Reset() { *m = NetworkResource{} } func (m *NetworkResource) String() string { return proto.CompactTextString(m) } func (*NetworkResource) ProtoMessage() {} func (*NetworkResource) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{29} + return fileDescriptor_driver_60d7e28aa66468bf, []int{29} } func (m *NetworkResource) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NetworkResource.Unmarshal(m, b) @@ -1673,7 +1673,7 @@ func (m *NetworkPort) Reset() { *m = NetworkPort{} } func (m *NetworkPort) String() string { return proto.CompactTextString(m) } func (*NetworkPort) ProtoMessage() {} func (*NetworkPort) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{30} + return fileDescriptor_driver_60d7e28aa66468bf, []int{30} } func (m *NetworkPort) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NetworkPort.Unmarshal(m, b) @@ -1715,7 +1715,7 @@ type LinuxResources struct { // CPU shares (relative weight vs. other containers). Default: 0 (not specified) CpuShares int64 `protobuf:"varint,3,opt,name=cpu_shares,json=cpuShares,proto3" json:"cpu_shares,omitempty"` // Memory limit in bytes. Default: 0 (not specified) - MemoryLimitInBytes int64 `protobuf:"varint,4,opt,name=memory_limit_in_bytes,json=memoryLimitInBytes,proto3" json:"memory_limit_in_bytes,omitempty"` + MemoryLimitBytes int64 `protobuf:"varint,4,opt,name=memory_limit_bytes,json=memoryLimitBytes,proto3" json:"memory_limit_bytes,omitempty"` // OOMScoreAdj adjusts the oom-killer score. Default: 0 (not specified) OomScoreAdj int64 `protobuf:"varint,5,opt,name=oom_score_adj,json=oomScoreAdj,proto3" json:"oom_score_adj,omitempty"` // CpusetCpus constrains the allowed set of logical CPUs. Default: "" (not specified) @@ -1731,7 +1731,7 @@ func (m *LinuxResources) Reset() { *m = LinuxResources{} } func (m *LinuxResources) String() string { return proto.CompactTextString(m) } func (*LinuxResources) ProtoMessage() {} func (*LinuxResources) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{31} + return fileDescriptor_driver_60d7e28aa66468bf, []int{31} } func (m *LinuxResources) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LinuxResources.Unmarshal(m, b) @@ -1772,9 +1772,9 @@ func (m *LinuxResources) GetCpuShares() int64 { return 0 } -func (m *LinuxResources) GetMemoryLimitInBytes() int64 { +func (m *LinuxResources) GetMemoryLimitBytes() int64 { if m != nil { - return m.MemoryLimitInBytes + return m.MemoryLimitBytes } return 0 } @@ -1816,7 +1816,7 @@ func (m *Mount) Reset() { *m = Mount{} } func (m *Mount) String() string { return proto.CompactTextString(m) } func (*Mount) ProtoMessage() {} func (*Mount) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{32} + return fileDescriptor_driver_60d7e28aa66468bf, []int{32} } func (m *Mount) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mount.Unmarshal(m, b) @@ -1879,7 +1879,7 @@ func (m *Device) Reset() { *m = Device{} } func (m *Device) String() string { return proto.CompactTextString(m) } func (*Device) ProtoMessage() {} func (*Device) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{33} + return fileDescriptor_driver_60d7e28aa66468bf, []int{33} } func (m *Device) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Device.Unmarshal(m, b) @@ -1937,7 +1937,7 @@ func (m *TaskHandle) Reset() { *m = TaskHandle{} } func (m *TaskHandle) String() string { return proto.CompactTextString(m) } func (*TaskHandle) ProtoMessage() {} func (*TaskHandle) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{34} + return fileDescriptor_driver_60d7e28aa66468bf, []int{34} } func (m *TaskHandle) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskHandle.Unmarshal(m, b) @@ -1997,7 +1997,7 @@ func (m *NetworkOverride) Reset() { *m = NetworkOverride{} } func (m *NetworkOverride) String() string { return proto.CompactTextString(m) } func (*NetworkOverride) ProtoMessage() {} func (*NetworkOverride) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{35} + return fileDescriptor_driver_60d7e28aa66468bf, []int{35} } func (m *NetworkOverride) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NetworkOverride.Unmarshal(m, b) @@ -2055,7 +2055,7 @@ func (m *ExitResult) Reset() { *m = ExitResult{} } func (m *ExitResult) String() string { return proto.CompactTextString(m) } func (*ExitResult) ProtoMessage() {} func (*ExitResult) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{36} + return fileDescriptor_driver_60d7e28aa66468bf, []int{36} } func (m *ExitResult) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExitResult.Unmarshal(m, b) @@ -2118,7 +2118,7 @@ func (m *TaskStatus) Reset() { *m = TaskStatus{} } func (m *TaskStatus) String() string { return proto.CompactTextString(m) } func (*TaskStatus) ProtoMessage() {} func (*TaskStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{37} + return fileDescriptor_driver_60d7e28aa66468bf, []int{37} } func (m *TaskStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStatus.Unmarshal(m, b) @@ -2193,7 +2193,7 @@ func (m *TaskDriverStatus) Reset() { *m = TaskDriverStatus{} } func (m *TaskDriverStatus) String() string { return proto.CompactTextString(m) } func (*TaskDriverStatus) ProtoMessage() {} func (*TaskDriverStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{38} + return fileDescriptor_driver_60d7e28aa66468bf, []int{38} } func (m *TaskDriverStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskDriverStatus.Unmarshal(m, b) @@ -2238,7 +2238,7 @@ func (m *TaskStats) Reset() { *m = TaskStats{} } func (m *TaskStats) String() string { return proto.CompactTextString(m) } func (*TaskStats) ProtoMessage() {} func (*TaskStats) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{39} + return fileDescriptor_driver_60d7e28aa66468bf, []int{39} } func (m *TaskStats) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskStats.Unmarshal(m, b) @@ -2300,7 +2300,7 @@ func (m *TaskResourceUsage) Reset() { *m = TaskResourceUsage{} } func (m *TaskResourceUsage) String() string { return proto.CompactTextString(m) } func (*TaskResourceUsage) ProtoMessage() {} func (*TaskResourceUsage) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{40} + return fileDescriptor_driver_60d7e28aa66468bf, []int{40} } func (m *TaskResourceUsage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TaskResourceUsage.Unmarshal(m, b) @@ -2352,7 +2352,7 @@ func (m *CPUUsage) Reset() { *m = CPUUsage{} } func (m *CPUUsage) String() string { return proto.CompactTextString(m) } func (*CPUUsage) ProtoMessage() {} func (*CPUUsage) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{41} + return fileDescriptor_driver_60d7e28aa66468bf, []int{41} } func (m *CPUUsage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CPUUsage.Unmarshal(m, b) @@ -2438,7 +2438,7 @@ func (m *MemoryUsage) Reset() { *m = MemoryUsage{} } func (m *MemoryUsage) String() string { return proto.CompactTextString(m) } func (*MemoryUsage) ProtoMessage() {} func (*MemoryUsage) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{42} + return fileDescriptor_driver_60d7e28aa66468bf, []int{42} } func (m *MemoryUsage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MemoryUsage.Unmarshal(m, b) @@ -2518,7 +2518,7 @@ func (m *DriverTaskEvent) Reset() { *m = DriverTaskEvent{} } func (m *DriverTaskEvent) String() string { return proto.CompactTextString(m) } func (*DriverTaskEvent) ProtoMessage() {} func (*DriverTaskEvent) Descriptor() ([]byte, []int) { - return fileDescriptor_driver_16b159caa26c45f9, []int{43} + return fileDescriptor_driver_60d7e28aa66468bf, []int{43} } func (m *DriverTaskEvent) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DriverTaskEvent.Unmarshal(m, b) @@ -3208,181 +3208,181 @@ var _Driver_serviceDesc = grpc.ServiceDesc{ Metadata: "driver.proto", } -func init() { proto.RegisterFile("driver.proto", fileDescriptor_driver_16b159caa26c45f9) } +func init() { proto.RegisterFile("driver.proto", fileDescriptor_driver_60d7e28aa66468bf) } -var fileDescriptor_driver_16b159caa26c45f9 = []byte{ - // 2757 bytes of a gzipped FileDescriptorProto +var fileDescriptor_driver_60d7e28aa66468bf = []byte{ + // 2753 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x59, 0xcb, 0x6f, 0x23, 0xc7, 0xd1, 0x17, 0x9f, 0x22, 0x8b, 0x14, 0x35, 0xdb, 0xbb, 0xfb, 0x99, 0xa6, 0xf1, 0x7d, 0x5e, 0x0f, 0x60, 0x40, 0xb0, 0xbd, 0x94, 0x2d, 0xe3, 0xf3, 0x3e, 0x02, 0x3f, 0x68, 0x72, 0x76, 0x25, 0xaf, 0x44, 0x29, 0x4d, 0x0a, 0xeb, 0x4d, 0x62, 0x4f, 0x46, 0x33, 0x2d, 0x72, 0x2c, 0xce, 0xc3, 0xdd, - 0x43, 0xad, 0x84, 0x20, 0x48, 0x90, 0x04, 0x41, 0x12, 0x20, 0x40, 0x2e, 0x41, 0xee, 0xb9, 0x05, - 0xb9, 0xe6, 0x94, 0xc7, 0x25, 0x40, 0xfe, 0x87, 0x1c, 0x73, 0x09, 0x90, 0x6b, 0xfe, 0x83, 0xa0, - 0x1f, 0x33, 0x1c, 0x4a, 0x5a, 0x7b, 0x48, 0xe5, 0xc4, 0xa9, 0xea, 0xee, 0x5f, 0x17, 0xab, 0xaa, - 0xab, 0xaa, 0xbb, 0xa0, 0xee, 0x50, 0xf7, 0x94, 0xd0, 0x76, 0x48, 0x83, 0x28, 0x40, 0xaf, 0x8f, - 0x2d, 0x36, 0x76, 0xed, 0x80, 0x86, 0x6d, 0x3f, 0xf0, 0x2c, 0xa7, 0x1d, 0x4e, 0xa6, 0x23, 0xd7, - 0x67, 0x6d, 0x39, 0x8b, 0xc9, 0x69, 0xad, 0xff, 0x1b, 0x05, 0xc1, 0x68, 0x42, 0x36, 0x05, 0x75, - 0x34, 0x3d, 0xde, 0x74, 0xa6, 0xd4, 0x8a, 0xdc, 0xc0, 0x57, 0xe3, 0xaf, 0x5e, 0x1c, 0x8f, 0x5c, - 0x8f, 0xb0, 0xc8, 0xf2, 0x42, 0x35, 0xe1, 0xa3, 0x91, 0x1b, 0x8d, 0xa7, 0x47, 0x6d, 0x3b, 0xf0, - 0x36, 0x93, 0x2d, 0x37, 0xc5, 0x96, 0x9b, 0x6a, 0xcb, 0x4d, 0x36, 0xb6, 0x28, 0x71, 0x36, 0xc7, - 0xf6, 0x84, 0x85, 0xc4, 0xe6, 0xbf, 0x26, 0xff, 0x90, 0x08, 0xfa, 0xcb, 0xf0, 0xd2, 0xd0, 0x62, - 0x27, 0xdd, 0xc0, 0x3f, 0x76, 0x47, 0x03, 0x7b, 0x4c, 0x3c, 0x0b, 0x93, 0x2f, 0xa7, 0x84, 0x45, - 0xfa, 0x77, 0xa0, 0x79, 0x79, 0x88, 0x85, 0x81, 0xcf, 0x08, 0xfa, 0x08, 0x8a, 0x1c, 0xa4, 0x99, - 0xbb, 0x93, 0xdb, 0xa8, 0x6d, 0xbd, 0xd5, 0x7e, 0xd1, 0xff, 0x95, 0x9b, 0xb7, 0xd5, 0xe6, 0xed, - 0x41, 0x48, 0x6c, 0x2c, 0x56, 0xea, 0xb7, 0xe1, 0x66, 0xd7, 0x0a, 0xad, 0x23, 0x77, 0xe2, 0x46, - 0x2e, 0x61, 0xf1, 0xa6, 0x53, 0xb8, 0x35, 0xcf, 0x56, 0x1b, 0x7e, 0x06, 0x75, 0x3b, 0xc5, 0x57, - 0x1b, 0x3f, 0x68, 0x67, 0x52, 0x74, 0xbb, 0x27, 0xa8, 0x39, 0xe0, 0x39, 0x38, 0xfd, 0x16, 0xa0, - 0x47, 0xae, 0x3f, 0x22, 0x34, 0xa4, 0xae, 0x1f, 0xc5, 0xc2, 0xfc, 0xa2, 0x00, 0x37, 0xe7, 0xd8, - 0x4a, 0x98, 0x2f, 0x00, 0xac, 0x28, 0xa2, 0xee, 0xd1, 0x34, 0x12, 0xa2, 0x14, 0x36, 0x6a, 0x5b, - 0x9f, 0x64, 0x14, 0xe5, 0x0a, 0xbc, 0x76, 0x27, 0x01, 0x33, 0xfc, 0x88, 0x9e, 0xe3, 0x14, 0x3a, - 0xfa, 0x1c, 0xca, 0x63, 0x62, 0x4d, 0xa2, 0x71, 0x33, 0x7f, 0x27, 0xb7, 0xd1, 0xd8, 0x7a, 0x74, - 0x8d, 0x7d, 0xb6, 0x05, 0xd0, 0x20, 0xb2, 0x22, 0x82, 0x15, 0x2a, 0xba, 0x0b, 0x48, 0x7e, 0x99, - 0x0e, 0x61, 0x36, 0x75, 0x43, 0xee, 0x7f, 0xcd, 0xc2, 0x9d, 0xdc, 0x46, 0x15, 0xdf, 0x90, 0x23, - 0xbd, 0xd9, 0x40, 0xeb, 0x7d, 0x58, 0xbf, 0x20, 0x2d, 0xd2, 0xa0, 0x70, 0x42, 0xce, 0x85, 0x45, - 0xaa, 0x98, 0x7f, 0xa2, 0x5b, 0x50, 0x3a, 0xb5, 0x26, 0x53, 0x22, 0x44, 0xae, 0x62, 0x49, 0x3c, - 0xcc, 0xdf, 0xcf, 0xe9, 0x0f, 0xa0, 0x96, 0x12, 0x02, 0x35, 0x00, 0x0e, 0xfb, 0x3d, 0x63, 0x68, - 0x74, 0x87, 0x46, 0x4f, 0x5b, 0x41, 0x6b, 0x50, 0x3d, 0xec, 0x6f, 0x1b, 0x9d, 0xdd, 0xe1, 0xf6, - 0x33, 0x2d, 0x87, 0x6a, 0xb0, 0x1a, 0x13, 0x79, 0xfd, 0x0c, 0x10, 0x26, 0x76, 0x70, 0x4a, 0x28, - 0xf7, 0x4a, 0x65, 0x22, 0xf4, 0x12, 0xac, 0x46, 0x16, 0x3b, 0x31, 0x5d, 0x47, 0x09, 0x50, 0xe6, - 0xe4, 0x8e, 0x83, 0x76, 0xa0, 0x3c, 0xb6, 0x7c, 0x67, 0x22, 0x85, 0xa8, 0x6d, 0xbd, 0x93, 0x51, - 0x6f, 0x1c, 0x7c, 0x5b, 0x2c, 0xc4, 0x0a, 0x80, 0xbb, 0xea, 0xdc, 0xce, 0x52, 0x9b, 0xfa, 0x33, - 0xd0, 0x06, 0x91, 0x45, 0xa3, 0xb4, 0x38, 0x06, 0x14, 0xf9, 0xfe, 0xca, 0x3d, 0x17, 0xd9, 0x53, - 0x1e, 0x33, 0x2c, 0x96, 0xeb, 0xff, 0xce, 0xc3, 0x8d, 0x14, 0xb6, 0x72, 0xbb, 0xa7, 0x50, 0xa6, - 0x84, 0x4d, 0x27, 0x91, 0x80, 0x6f, 0x6c, 0x7d, 0x98, 0x11, 0xfe, 0x12, 0x52, 0x1b, 0x0b, 0x18, - 0xac, 0xe0, 0xd0, 0x06, 0x68, 0x72, 0x85, 0x49, 0x28, 0x0d, 0xa8, 0xe9, 0xb1, 0x91, 0x32, 0x5d, - 0x43, 0xf2, 0x0d, 0xce, 0xde, 0x63, 0xa3, 0x94, 0x56, 0x0b, 0xd7, 0xd4, 0x2a, 0xb2, 0x40, 0xf3, - 0x49, 0xf4, 0x3c, 0xa0, 0x27, 0x26, 0x57, 0x2d, 0x75, 0x1d, 0xd2, 0x2c, 0x0a, 0xd0, 0xf7, 0x32, - 0x82, 0xf6, 0xe5, 0xf2, 0x7d, 0xb5, 0x1a, 0xaf, 0xfb, 0xf3, 0x0c, 0xfd, 0x4d, 0x28, 0xcb, 0x7f, - 0xca, 0x3d, 0x69, 0x70, 0xd8, 0xed, 0x1a, 0x83, 0x81, 0xb6, 0x82, 0xaa, 0x50, 0xc2, 0xc6, 0x10, - 0x73, 0x0f, 0xab, 0x42, 0xe9, 0x51, 0x67, 0xd8, 0xd9, 0xd5, 0xf2, 0xfa, 0x1b, 0xb0, 0xfe, 0xd4, - 0x72, 0xa3, 0x2c, 0xce, 0xa5, 0x07, 0xa0, 0xcd, 0xe6, 0x2a, 0xeb, 0xec, 0xcc, 0x59, 0x27, 0xbb, - 0x6a, 0x8c, 0x33, 0x37, 0xba, 0x60, 0x0f, 0x0d, 0x0a, 0x84, 0x52, 0x65, 0x02, 0xfe, 0xa9, 0x3f, - 0x87, 0xf5, 0x41, 0x14, 0x84, 0x99, 0x3c, 0xff, 0x5d, 0x58, 0xe5, 0x79, 0x22, 0x98, 0x46, 0xca, - 0xf5, 0x5f, 0x6e, 0xcb, 0x3c, 0xd2, 0x8e, 0xf3, 0x48, 0xbb, 0xa7, 0xf2, 0x0c, 0x8e, 0x67, 0xa2, - 0xff, 0x81, 0x32, 0x73, 0x47, 0xbe, 0x35, 0x51, 0x47, 0x5f, 0x51, 0x3a, 0xe2, 0x4e, 0x1e, 0x6f, - 0xac, 0x1c, 0xbf, 0x0b, 0xa8, 0x47, 0x58, 0x44, 0x83, 0xf3, 0x4c, 0xf2, 0xdc, 0x82, 0xd2, 0x71, - 0x40, 0x6d, 0x79, 0x10, 0x2b, 0x58, 0x12, 0xfc, 0x50, 0xcd, 0x81, 0x28, 0xec, 0xbb, 0x80, 0x76, - 0x7c, 0x9e, 0x20, 0xb2, 0x19, 0xe2, 0x57, 0x79, 0xb8, 0x39, 0x37, 0x5f, 0x19, 0x63, 0xf9, 0x73, - 0xc8, 0x03, 0xd3, 0x94, 0xc9, 0x73, 0x88, 0xf6, 0xa1, 0x2c, 0x67, 0x28, 0x4d, 0xde, 0x5b, 0x00, - 0x48, 0xe6, 0x1c, 0x05, 0xa7, 0x60, 0xae, 0x74, 0xfa, 0xc2, 0x7f, 0xdb, 0xe9, 0xb5, 0xf8, 0x7f, - 0xb0, 0xaf, 0xd5, 0xdf, 0xb7, 0xe1, 0x46, 0x6a, 0xb2, 0x52, 0xde, 0x23, 0x28, 0x31, 0xce, 0x50, - 0xda, 0x7b, 0x7b, 0x41, 0xed, 0x31, 0x2c, 0x97, 0xeb, 0x37, 0x25, 0xb8, 0x71, 0x4a, 0xfc, 0x44, - 0x14, 0xbd, 0x07, 0x37, 0x06, 0xc2, 0xb5, 0x32, 0xf9, 0xce, 0xcc, 0x2d, 0xf3, 0x73, 0x6e, 0x79, - 0x0b, 0x50, 0x1a, 0x45, 0x39, 0xcf, 0x39, 0xac, 0x1b, 0x67, 0xc4, 0xce, 0x84, 0xdc, 0x84, 0x55, - 0x3b, 0xf0, 0x3c, 0xcb, 0x77, 0x9a, 0xf9, 0x3b, 0x85, 0x8d, 0x2a, 0x8e, 0xc9, 0xf4, 0xf9, 0x29, - 0x64, 0x3d, 0x3f, 0xfa, 0x2f, 0x73, 0xa0, 0xcd, 0xf6, 0x56, 0x8a, 0xe4, 0xd2, 0x47, 0x0e, 0x07, - 0xe2, 0x7b, 0xd7, 0xb1, 0xa2, 0x14, 0x3f, 0x3e, 0xe2, 0x92, 0x4f, 0x28, 0x4d, 0x85, 0x90, 0xc2, - 0x35, 0x43, 0x88, 0xfe, 0xcf, 0x1c, 0xa0, 0xcb, 0x55, 0x0f, 0x7a, 0x0d, 0xea, 0x8c, 0xf8, 0x8e, - 0x29, 0xd5, 0x28, 0x2d, 0x5c, 0xc1, 0x35, 0xce, 0x93, 0xfa, 0x64, 0x08, 0x41, 0x91, 0x9c, 0x11, - 0x5b, 0x9d, 0x56, 0xf1, 0x8d, 0xc6, 0x50, 0x3f, 0x66, 0xa6, 0xcb, 0x82, 0x89, 0x95, 0x94, 0x07, - 0x8d, 0x2d, 0x63, 0xe9, 0xea, 0xab, 0xfd, 0x68, 0xb0, 0x13, 0x83, 0xe1, 0xda, 0x31, 0x4b, 0x08, - 0xbd, 0x0d, 0xb5, 0xd4, 0x18, 0xaa, 0x40, 0xb1, 0xbf, 0xdf, 0x37, 0xb4, 0x15, 0x04, 0x50, 0xee, - 0x6e, 0xe3, 0xfd, 0xfd, 0xa1, 0x8c, 0xda, 0x3b, 0x7b, 0x9d, 0xc7, 0x86, 0x96, 0xd7, 0xff, 0x5c, - 0x04, 0x98, 0xa5, 0x4f, 0xd4, 0x80, 0x7c, 0x62, 0xe9, 0xbc, 0xeb, 0xf0, 0x3f, 0xe3, 0x5b, 0x5e, - 0x5c, 0x88, 0x88, 0x6f, 0xb4, 0x05, 0xb7, 0x3d, 0x36, 0x0a, 0x2d, 0xfb, 0xc4, 0x54, 0x59, 0xcf, - 0x16, 0x8b, 0xc5, 0xbf, 0xaa, 0xe3, 0x9b, 0x6a, 0x50, 0x49, 0x2d, 0x71, 0x77, 0xa1, 0x40, 0xfc, - 0xd3, 0x66, 0x51, 0x94, 0x7a, 0x0f, 0x17, 0x4e, 0xeb, 0x6d, 0xc3, 0x3f, 0x95, 0xa5, 0x1d, 0x87, - 0x41, 0x7d, 0xa8, 0x52, 0xc2, 0x82, 0x29, 0xb5, 0x09, 0x6b, 0x96, 0x16, 0x3a, 0x64, 0x38, 0x5e, - 0x87, 0x67, 0x10, 0xa8, 0x07, 0x65, 0x2f, 0x98, 0xfa, 0x11, 0x6b, 0x96, 0x85, 0x80, 0x6f, 0x65, - 0x04, 0xdb, 0xe3, 0x8b, 0xb0, 0x5a, 0x8b, 0x1e, 0xc3, 0xaa, 0x43, 0x4e, 0x5d, 0x2e, 0xd3, 0xaa, - 0x80, 0xb9, 0x9b, 0xd5, 0xbe, 0x62, 0x15, 0x8e, 0x57, 0x73, 0xa5, 0x4f, 0x19, 0xa1, 0xcd, 0x8a, - 0x54, 0x3a, 0xff, 0x46, 0xaf, 0x40, 0xd5, 0x9a, 0x4c, 0x02, 0xdb, 0x74, 0x5c, 0xda, 0xac, 0x8a, - 0x81, 0x8a, 0x60, 0xf4, 0x5c, 0x8a, 0x5e, 0x85, 0x9a, 0x3c, 0x19, 0x66, 0x68, 0x45, 0xe3, 0x26, - 0x88, 0x61, 0x90, 0xac, 0x03, 0x2b, 0x1a, 0xab, 0x09, 0x84, 0x52, 0x39, 0xa1, 0x96, 0x4c, 0x20, - 0x94, 0xf2, 0x09, 0xad, 0xf7, 0xa0, 0x12, 0xab, 0x78, 0xa1, 0x7a, 0xf4, 0x6f, 0x39, 0xa8, 0x26, - 0x2a, 0x45, 0x9f, 0xc2, 0x1a, 0xb5, 0x9e, 0x9b, 0x33, 0xdb, 0xc8, 0x00, 0xf8, 0x6e, 0x56, 0xdb, - 0x58, 0xcf, 0x67, 0xe6, 0xa9, 0xd3, 0x14, 0x85, 0x3e, 0x87, 0xf5, 0x89, 0xeb, 0x4f, 0xcf, 0x52, - 0xd8, 0x32, 0xa3, 0xfc, 0x7f, 0x46, 0xec, 0x5d, 0xbe, 0x7a, 0x86, 0xde, 0x98, 0xcc, 0xd1, 0xfa, - 0x1f, 0x72, 0x50, 0x4f, 0x6f, 0xcf, 0x95, 0x60, 0x87, 0x53, 0xf1, 0x07, 0x0a, 0x98, 0x7f, 0xf2, - 0xa0, 0xe3, 0x11, 0x2f, 0xa0, 0xe7, 0x62, 0xe7, 0x02, 0x56, 0x14, 0xb7, 0x96, 0xe3, 0xb2, 0x13, - 0xe1, 0xfd, 0x05, 0x2c, 0xbe, 0x39, 0xcf, 0x0d, 0x42, 0x26, 0xea, 0xb1, 0x02, 0x16, 0xdf, 0x08, - 0x43, 0x45, 0xa5, 0x1a, 0xee, 0xb3, 0x85, 0xc5, 0x53, 0x56, 0x2c, 0x1c, 0x4e, 0x70, 0xf4, 0xdf, - 0xe4, 0x61, 0xfd, 0xc2, 0x28, 0x97, 0x53, 0x3a, 0x52, 0x1c, 0xb0, 0x25, 0xc5, 0x65, 0xb2, 0x5d, - 0x27, 0xae, 0x8a, 0xc4, 0xb7, 0x38, 0xee, 0xa1, 0xaa, 0x58, 0xf2, 0x6e, 0xc8, 0x0d, 0xed, 0x1d, - 0xb9, 0x91, 0x14, 0xbc, 0x84, 0x25, 0x81, 0x9e, 0x41, 0x83, 0x12, 0x46, 0xe8, 0x29, 0x71, 0xcc, - 0x30, 0xa0, 0x51, 0x2c, 0xff, 0xd6, 0x62, 0xf2, 0x1f, 0x04, 0x34, 0xc2, 0x6b, 0x31, 0x12, 0xa7, - 0x18, 0x7a, 0x0a, 0x6b, 0xce, 0xb9, 0x6f, 0x79, 0xae, 0xad, 0x90, 0xcb, 0x4b, 0x23, 0xd7, 0x15, - 0x90, 0x00, 0xe6, 0x17, 0xa5, 0xd4, 0x20, 0xff, 0x63, 0x13, 0xeb, 0x88, 0x4c, 0x94, 0x4e, 0x24, - 0x31, 0xef, 0xd7, 0x25, 0xe5, 0xd7, 0xfa, 0x4f, 0xf2, 0xd0, 0x98, 0x77, 0x17, 0xf4, 0xbf, 0x00, - 0x76, 0x38, 0x35, 0x43, 0x42, 0xdd, 0xc0, 0x51, 0x4e, 0x51, 0xb5, 0xc3, 0xe9, 0x81, 0x60, 0xf0, - 0xc3, 0xc9, 0x87, 0xbf, 0x9c, 0x06, 0x91, 0xa5, 0xbc, 0xa3, 0x62, 0x87, 0xd3, 0x6f, 0x72, 0x3a, - 0x5e, 0x2b, 0x6e, 0xf2, 0x4c, 0x79, 0x09, 0x9f, 0x3e, 0x10, 0x0c, 0xf4, 0x0e, 0xdc, 0x96, 0x8e, - 0x64, 0x4e, 0x5c, 0xcf, 0x8d, 0x4c, 0xd7, 0x37, 0x8f, 0xce, 0xf9, 0xb5, 0x58, 0xfa, 0x0e, 0x92, - 0x83, 0xbb, 0x7c, 0x6c, 0xc7, 0xff, 0x98, 0x8f, 0x20, 0x1d, 0xd6, 0x82, 0xc0, 0x33, 0x99, 0x1d, - 0x50, 0x62, 0x5a, 0xce, 0x17, 0x22, 0x04, 0x16, 0x70, 0x2d, 0x08, 0xbc, 0x01, 0xe7, 0x75, 0x9c, - 0x2f, 0xf8, 0x89, 0xb7, 0xc3, 0x29, 0x23, 0x91, 0xc9, 0x7f, 0x9a, 0x65, 0x79, 0xe2, 0x25, 0xab, - 0x1b, 0x4e, 0x59, 0x6a, 0x82, 0x47, 0x3c, 0x1e, 0xb1, 0x52, 0x13, 0xf6, 0x88, 0xc7, 0xf4, 0xcf, - 0xa0, 0x24, 0xe2, 0x1b, 0xff, 0x77, 0xa2, 0x04, 0x10, 0xa1, 0x43, 0xea, 0xaf, 0xc2, 0x19, 0x22, - 0xb2, 0xbc, 0x02, 0xd5, 0x71, 0xc0, 0x54, 0xe0, 0x91, 0xae, 0x55, 0xe1, 0x0c, 0x31, 0xd8, 0x82, - 0x0a, 0x25, 0x96, 0x13, 0xf8, 0x93, 0x73, 0xf1, 0xc7, 0x2b, 0x38, 0xa1, 0xf5, 0x2f, 0xa1, 0x2c, - 0xe3, 0xde, 0x35, 0xf0, 0xef, 0x02, 0xb2, 0x47, 0x34, 0x98, 0x86, 0xdc, 0x32, 0x9e, 0xcb, 0x98, - 0x1b, 0xf8, 0x2c, 0xbe, 0x7b, 0xcb, 0x91, 0x83, 0xd9, 0x80, 0xfe, 0xd7, 0x9c, 0xcc, 0x75, 0xf2, - 0x22, 0xc5, 0xab, 0x05, 0x95, 0xb8, 0x96, 0xbe, 0x6d, 0x2a, 0x80, 0xb8, 0xe2, 0x23, 0xea, 0x8d, - 0x61, 0xd1, 0x8a, 0x8f, 0xc8, 0x8a, 0x8f, 0xf0, 0xf2, 0x42, 0xa5, 0x54, 0x09, 0x27, 0x33, 0x6a, - 0xcd, 0x49, 0x4a, 0x61, 0xa2, 0xff, 0x2b, 0x97, 0x1c, 0xf9, 0xb8, 0x64, 0x45, 0x9f, 0x43, 0x85, - 0x9f, 0x1e, 0xd3, 0xb3, 0x42, 0xf5, 0x9a, 0xd2, 0x5d, 0xae, 0x1a, 0x6e, 0xf3, 0xc3, 0xb2, 0x67, - 0x85, 0x32, 0xd7, 0xae, 0x86, 0x92, 0xe2, 0xa1, 0xc3, 0x72, 0x66, 0xa1, 0x83, 0x7f, 0xa3, 0xd7, - 0xa1, 0x61, 0x4d, 0xa3, 0xc0, 0xb4, 0x9c, 0x53, 0x42, 0x23, 0x97, 0x11, 0x65, 0xe1, 0x35, 0xce, - 0xed, 0xc4, 0xcc, 0xd6, 0x43, 0xa8, 0xa7, 0x31, 0xbf, 0x2e, 0xb9, 0x94, 0xd2, 0xc9, 0xe5, 0xbb, - 0x00, 0xb3, 0xca, 0x8c, 0x7b, 0x02, 0x39, 0x73, 0x23, 0xd3, 0x0e, 0x1c, 0x19, 0xda, 0x4a, 0xb8, - 0xc2, 0x19, 0xdd, 0xc0, 0x21, 0x17, 0xea, 0xdc, 0x52, 0x5c, 0xe7, 0xf2, 0xc3, 0xc7, 0x8f, 0xca, - 0x89, 0x3b, 0x99, 0x10, 0x47, 0x49, 0x58, 0x0d, 0x02, 0xef, 0x89, 0x60, 0xe8, 0x7f, 0xc9, 0x4b, - 0x8f, 0x90, 0xb7, 0x8c, 0x4c, 0xd5, 0x4f, 0x62, 0xea, 0xc2, 0xf5, 0x4c, 0xfd, 0x00, 0x80, 0x45, - 0x16, 0x8d, 0x88, 0x63, 0x5a, 0x91, 0xba, 0xb8, 0xb7, 0x2e, 0x15, 0xca, 0xc3, 0xf8, 0xc1, 0x12, - 0x57, 0xd5, 0xec, 0x4e, 0x84, 0xde, 0x87, 0xba, 0x1d, 0x78, 0xe1, 0x84, 0xa8, 0xc5, 0xa5, 0xaf, - 0x5d, 0x5c, 0x4b, 0xe6, 0x77, 0xa2, 0x54, 0x95, 0x5c, 0xbe, 0x6e, 0x95, 0xfc, 0xc7, 0x9c, 0xbc, - 0x2c, 0xa5, 0xef, 0x6a, 0x68, 0x74, 0xc5, 0xeb, 0xde, 0xe3, 0x25, 0x2f, 0x7e, 0x5f, 0xf5, 0xb4, - 0x77, 0xdd, 0xb7, 0xb4, 0x3f, 0x15, 0xa0, 0x9a, 0xdc, 0xb9, 0x2e, 0xd9, 0xfe, 0x3e, 0x54, 0x93, - 0xd7, 0x62, 0x55, 0x6b, 0x7c, 0xa5, 0x79, 0x92, 0xc9, 0xe8, 0x18, 0x90, 0x35, 0x1a, 0x25, 0x95, - 0x8a, 0x39, 0x65, 0xd6, 0x28, 0xbe, 0xa5, 0xde, 0x5f, 0x40, 0x0f, 0x71, 0xfa, 0x39, 0xe4, 0xeb, - 0xb1, 0x66, 0x8d, 0x46, 0x73, 0x1c, 0xf4, 0x3d, 0xb8, 0x3d, 0xbf, 0x87, 0x79, 0x74, 0x6e, 0x86, - 0xae, 0xa3, 0xaa, 0xec, 0xed, 0x45, 0xaf, 0x9d, 0xed, 0x39, 0xf8, 0x8f, 0xcf, 0x0f, 0x5c, 0x47, - 0xea, 0x1c, 0xd1, 0x4b, 0x03, 0xad, 0x1f, 0xc0, 0x4b, 0x2f, 0x98, 0x7e, 0x85, 0x0d, 0xfa, 0x69, - 0x1b, 0x5c, 0x47, 0x09, 0x29, 0xeb, 0xfd, 0x36, 0x27, 0x6f, 0xc7, 0xf3, 0x3a, 0xe9, 0xcc, 0xca, - 0xb6, 0xda, 0xd6, 0x66, 0xc6, 0x7d, 0xba, 0x07, 0x87, 0x12, 0x5e, 0xd4, 0x79, 0x9f, 0xcc, 0xd5, - 0x79, 0xd9, 0x6b, 0x91, 0x3d, 0xb1, 0x48, 0x02, 0x29, 0x04, 0xfd, 0xf7, 0x05, 0xa8, 0xc4, 0xe8, - 0xa2, 0x08, 0x3f, 0x67, 0x11, 0xf1, 0x4c, 0x2f, 0x0e, 0x61, 0x39, 0x0c, 0x92, 0xb5, 0xc7, 0x83, - 0xd8, 0x2b, 0x50, 0xe5, 0xb5, 0xbe, 0x1c, 0xce, 0x8b, 0xe1, 0x0a, 0x67, 0x88, 0xc1, 0x57, 0xa1, - 0x16, 0x05, 0x91, 0x35, 0x31, 0x23, 0xd7, 0x3e, 0x91, 0x49, 0x2e, 0x87, 0x41, 0xb0, 0x86, 0x9c, - 0x83, 0xde, 0x84, 0x1b, 0xd1, 0x98, 0x06, 0x51, 0x34, 0xe1, 0x65, 0x9a, 0x28, 0x4c, 0x64, 0x11, - 0x51, 0xc4, 0x5a, 0x32, 0x20, 0x0b, 0x16, 0xc6, 0xa3, 0xf7, 0x6c, 0x32, 0x77, 0x5d, 0x11, 0x44, - 0x8a, 0x78, 0x2d, 0xe1, 0x72, 0xd7, 0xe6, 0x97, 0xfc, 0x90, 0x50, 0x9b, 0xf8, 0x32, 0x56, 0xe4, - 0x70, 0x4c, 0x22, 0x13, 0xd6, 0x3d, 0x62, 0xb1, 0x29, 0x25, 0x8e, 0x79, 0xec, 0x92, 0x89, 0x23, - 0x2f, 0x3d, 0x8d, 0xcc, 0x45, 0x6d, 0xac, 0x96, 0xf6, 0x23, 0xb1, 0x1a, 0x37, 0x62, 0x38, 0x49, - 0xf3, 0xfa, 0x40, 0x7e, 0xa1, 0x75, 0xa8, 0x0d, 0x9e, 0x0d, 0x86, 0xc6, 0x9e, 0xb9, 0xb7, 0xdf, - 0x33, 0xd4, 0x2b, 0xf7, 0xc0, 0xc0, 0x92, 0xcc, 0xf1, 0xf1, 0xe1, 0xfe, 0xb0, 0xb3, 0x6b, 0x0e, - 0x77, 0xba, 0x4f, 0x06, 0x5a, 0x1e, 0xdd, 0x86, 0x1b, 0xc3, 0x6d, 0xbc, 0x3f, 0x1c, 0xee, 0x1a, - 0x3d, 0xf3, 0xc0, 0xc0, 0x3b, 0xfb, 0xbd, 0x81, 0x56, 0x40, 0x08, 0x1a, 0x33, 0xf6, 0x70, 0x67, - 0xcf, 0xd0, 0x8a, 0xa8, 0x06, 0xab, 0x07, 0x06, 0xee, 0x1a, 0xfd, 0xa1, 0x56, 0xd2, 0xff, 0x9e, - 0x87, 0x5a, 0xca, 0x8a, 0xdc, 0x91, 0x29, 0x93, 0x97, 0x98, 0x22, 0xe6, 0x9f, 0x3c, 0x98, 0xd8, - 0x96, 0x3d, 0x96, 0xd6, 0x29, 0x62, 0x49, 0x70, 0xbb, 0x79, 0xd6, 0x59, 0xea, 0x9c, 0x17, 0x71, - 0xc5, 0xb3, 0xce, 0x24, 0xc8, 0x6b, 0x50, 0x3f, 0x21, 0xd4, 0x27, 0x13, 0x35, 0x2e, 0x2d, 0x52, - 0x93, 0x3c, 0x39, 0x65, 0x03, 0x34, 0x35, 0x65, 0x06, 0x23, 0xcd, 0xd1, 0x90, 0xfc, 0xbd, 0x18, - 0xec, 0xe8, 0xb2, 0xd6, 0xcb, 0x42, 0xeb, 0x0f, 0x16, 0x77, 0xd2, 0x17, 0x29, 0x7e, 0x90, 0x28, - 0x7e, 0x15, 0x0a, 0x38, 0x7e, 0xf0, 0xed, 0x76, 0xba, 0xdb, 0x5c, 0xd9, 0x6b, 0x50, 0xdd, 0xeb, - 0x7c, 0x6a, 0x1e, 0x0e, 0xc4, 0xf3, 0x01, 0xd2, 0xa0, 0xfe, 0xc4, 0xc0, 0x7d, 0x63, 0x57, 0x71, - 0x0a, 0xe8, 0x16, 0x68, 0x8a, 0x33, 0x9b, 0x57, 0xd4, 0x7f, 0x97, 0x87, 0x75, 0x19, 0xd7, 0x93, - 0x17, 0xad, 0x17, 0x3f, 0x2d, 0x2d, 0x1f, 0x7a, 0x9b, 0xb0, 0xea, 0x11, 0x96, 0xd8, 0xa1, 0x8a, - 0x63, 0x12, 0xb9, 0x50, 0xb3, 0x7c, 0x3f, 0x88, 0xc4, 0xb3, 0x08, 0x53, 0x21, 0xf2, 0xf1, 0x42, - 0x0f, 0x30, 0x89, 0xe4, 0xed, 0xce, 0x0c, 0x49, 0x46, 0xc8, 0x34, 0x76, 0xeb, 0x03, 0xd0, 0x2e, - 0x4e, 0x58, 0x24, 0x2f, 0xbd, 0xf1, 0xce, 0x2c, 0x2d, 0x11, 0xee, 0xa0, 0x87, 0xfd, 0x27, 0xfd, - 0xfd, 0xa7, 0x7d, 0x6d, 0x85, 0x13, 0xf8, 0xb0, 0xdf, 0xdf, 0xe9, 0x3f, 0xd6, 0x72, 0x08, 0xa0, - 0x6c, 0x7c, 0xba, 0x33, 0x34, 0x7a, 0x5a, 0x7e, 0xeb, 0x1f, 0x6b, 0x50, 0x96, 0x42, 0xa2, 0x5f, - 0xab, 0x94, 0x9c, 0x6e, 0x3b, 0xa2, 0x0f, 0x16, 0x2e, 0x6d, 0xe7, 0x5a, 0x99, 0xad, 0x0f, 0x97, - 0x5e, 0xaf, 0x5e, 0x16, 0x57, 0xd0, 0xcf, 0x73, 0x50, 0x9f, 0x7b, 0x4a, 0xcb, 0xfa, 0x0a, 0x74, - 0x45, 0x97, 0xb3, 0xf5, 0x8d, 0xa5, 0xd6, 0x26, 0xb2, 0xfc, 0x2c, 0x07, 0xb5, 0x54, 0x7f, 0x0f, - 0x3d, 0x58, 0xa6, 0x27, 0x28, 0x25, 0x79, 0xb8, 0x7c, 0x3b, 0x51, 0x5f, 0x79, 0x3b, 0x87, 0x7e, - 0x9a, 0x83, 0x5a, 0xaa, 0x39, 0x96, 0x59, 0x94, 0xcb, 0xad, 0xbc, 0xcc, 0xa2, 0x5c, 0xd5, 0x8b, - 0x5b, 0x41, 0x3f, 0xcc, 0x41, 0x35, 0x69, 0x74, 0xa1, 0x7b, 0x8b, 0xb7, 0xc6, 0xa4, 0x10, 0xf7, - 0x97, 0xed, 0xa9, 0xe9, 0x2b, 0xe8, 0xfb, 0x50, 0x89, 0xbb, 0x42, 0x28, 0x6b, 0x1a, 0xb9, 0xd0, - 0x72, 0x6a, 0xdd, 0x5b, 0x78, 0x5d, 0x7a, 0xfb, 0xb8, 0x55, 0x93, 0x79, 0xfb, 0x0b, 0x4d, 0xa5, - 0xd6, 0xbd, 0x85, 0xd7, 0x25, 0xdb, 0x73, 0x4f, 0x48, 0x75, 0x74, 0x32, 0x7b, 0xc2, 0xe5, 0x56, - 0x52, 0x66, 0x4f, 0xb8, 0xaa, 0x81, 0x24, 0x05, 0x49, 0xf5, 0x84, 0x32, 0x0b, 0x72, 0xb9, 0xef, - 0x94, 0x59, 0x90, 0x2b, 0x5a, 0x50, 0xca, 0x25, 0x67, 0x05, 0xfa, 0xbd, 0x85, 0xdb, 0x28, 0x0b, - 0xba, 0xe4, 0xa5, 0x46, 0x8e, 0xbe, 0x82, 0x7e, 0xa4, 0x9e, 0x0c, 0x64, 0x0f, 0x06, 0x2d, 0x02, - 0x35, 0xd7, 0xb6, 0x69, 0xbd, 0xb7, 0x5c, 0xaa, 0x11, 0x31, 0xe2, 0xc7, 0x39, 0x80, 0x59, 0xb7, - 0x26, 0xb3, 0x10, 0x97, 0xda, 0x44, 0xad, 0x07, 0x4b, 0xac, 0x4c, 0x1f, 0x8f, 0xb8, 0x41, 0x93, - 0xf9, 0x78, 0x5c, 0xe8, 0x26, 0x65, 0x3e, 0x1e, 0x17, 0x3b, 0x41, 0xfa, 0xca, 0xc7, 0xab, 0xdf, - 0x2a, 0xc9, 0xdc, 0x5f, 0x16, 0x3f, 0xef, 0xfe, 0x27, 0x00, 0x00, 0xff, 0xff, 0x44, 0x47, 0xcf, - 0xbe, 0x32, 0x24, 0x00, 0x00, + 0x43, 0xad, 0x84, 0x20, 0x48, 0x90, 0x00, 0x41, 0x12, 0x20, 0x40, 0x2e, 0x41, 0xee, 0xb9, 0x05, + 0xb9, 0xe6, 0x94, 0xc7, 0x25, 0x40, 0xfe, 0x87, 0x1c, 0x73, 0x09, 0x90, 0x6b, 0xfe, 0x82, 0x04, + 0xfd, 0x98, 0xe1, 0x50, 0xd2, 0xda, 0x43, 0x2a, 0x27, 0x4e, 0x55, 0x77, 0xff, 0xba, 0x58, 0x55, + 0x5d, 0x55, 0xdd, 0x05, 0x75, 0x87, 0xba, 0xa7, 0x84, 0xb6, 0x43, 0x1a, 0x44, 0x01, 0x7a, 0x7d, + 0x6c, 0xb1, 0xb1, 0x6b, 0x07, 0x34, 0x6c, 0xfb, 0x81, 0x67, 0x39, 0xed, 0x70, 0x32, 0x1d, 0xb9, + 0x3e, 0x6b, 0xcb, 0x59, 0x4c, 0x4e, 0x6b, 0xfd, 0xdf, 0x28, 0x08, 0x46, 0x13, 0xb2, 0x29, 0xa8, + 0xa3, 0xe9, 0xf1, 0xa6, 0x33, 0xa5, 0x56, 0xe4, 0x06, 0xbe, 0x1a, 0x7f, 0xf5, 0xe2, 0x78, 0xe4, + 0x7a, 0x84, 0x45, 0x96, 0x17, 0xaa, 0x09, 0x1f, 0x8d, 0xdc, 0x68, 0x3c, 0x3d, 0x6a, 0xdb, 0x81, + 0xb7, 0x99, 0x6c, 0xb9, 0x29, 0xb6, 0xdc, 0x54, 0x5b, 0x6e, 0xb2, 0xb1, 0x45, 0x89, 0xb3, 0x39, + 0xb6, 0x27, 0x2c, 0x24, 0x36, 0xff, 0x35, 0xf9, 0x87, 0x44, 0xd0, 0x5f, 0x86, 0x97, 0x86, 0x16, + 0x3b, 0xe9, 0x06, 0xfe, 0xb1, 0x3b, 0x1a, 0xd8, 0x63, 0xe2, 0x59, 0x98, 0x7c, 0x39, 0x25, 0x2c, + 0xd2, 0xbf, 0x03, 0xcd, 0xcb, 0x43, 0x2c, 0x0c, 0x7c, 0x46, 0xd0, 0x47, 0x50, 0xe4, 0x20, 0xcd, + 0xdc, 0x9d, 0xdc, 0x46, 0x6d, 0xeb, 0xad, 0xf6, 0x8b, 0xfe, 0xaf, 0xdc, 0xbc, 0xad, 0x36, 0x6f, + 0x0f, 0x42, 0x62, 0x63, 0xb1, 0x52, 0xbf, 0x0d, 0x37, 0xbb, 0x56, 0x68, 0x1d, 0xb9, 0x13, 0x37, + 0x72, 0x09, 0x8b, 0x37, 0x9d, 0xc2, 0xad, 0x79, 0xb6, 0xda, 0xf0, 0x33, 0xa8, 0xdb, 0x29, 0xbe, + 0xda, 0xf8, 0x41, 0x3b, 0x93, 0xa2, 0xdb, 0x3d, 0x41, 0xcd, 0x01, 0xcf, 0xc1, 0xe9, 0xb7, 0x00, + 0x3d, 0x72, 0xfd, 0x11, 0xa1, 0x21, 0x75, 0xfd, 0x28, 0x16, 0xe6, 0xe7, 0x05, 0xb8, 0x39, 0xc7, + 0x56, 0xc2, 0x7c, 0x01, 0x60, 0x45, 0x11, 0x75, 0x8f, 0xa6, 0x91, 0x10, 0xa5, 0xb0, 0x51, 0xdb, + 0xfa, 0x24, 0xa3, 0x28, 0x57, 0xe0, 0xb5, 0x3b, 0x09, 0x98, 0xe1, 0x47, 0xf4, 0x1c, 0xa7, 0xd0, + 0xd1, 0xe7, 0x50, 0x1e, 0x13, 0x6b, 0x12, 0x8d, 0x9b, 0xf9, 0x3b, 0xb9, 0x8d, 0xc6, 0xd6, 0xa3, + 0x6b, 0xec, 0xb3, 0x2d, 0x80, 0x06, 0x91, 0x15, 0x11, 0xac, 0x50, 0xd1, 0x5d, 0x40, 0xf2, 0xcb, + 0x74, 0x08, 0xb3, 0xa9, 0x1b, 0x72, 0xff, 0x6b, 0x16, 0xee, 0xe4, 0x36, 0xaa, 0xf8, 0x86, 0x1c, + 0xe9, 0xcd, 0x06, 0x5a, 0xef, 0xc3, 0xfa, 0x05, 0x69, 0x91, 0x06, 0x85, 0x13, 0x72, 0x2e, 0x2c, + 0x52, 0xc5, 0xfc, 0x13, 0xdd, 0x82, 0xd2, 0xa9, 0x35, 0x99, 0x12, 0x21, 0x72, 0x15, 0x4b, 0xe2, + 0x61, 0xfe, 0x7e, 0x4e, 0x7f, 0x00, 0xb5, 0x94, 0x10, 0xa8, 0x01, 0x70, 0xd8, 0xef, 0x19, 0x43, + 0xa3, 0x3b, 0x34, 0x7a, 0xda, 0x0a, 0x5a, 0x83, 0xea, 0x61, 0x7f, 0xdb, 0xe8, 0xec, 0x0e, 0xb7, + 0x9f, 0x69, 0x39, 0x54, 0x83, 0xd5, 0x98, 0xc8, 0xeb, 0x67, 0x80, 0x30, 0xb1, 0x83, 0x53, 0x42, + 0xb9, 0x57, 0x2a, 0x13, 0xa1, 0x97, 0x60, 0x35, 0xb2, 0xd8, 0x89, 0xe9, 0x3a, 0x4a, 0x80, 0x32, + 0x27, 0x77, 0x1c, 0xb4, 0x03, 0xe5, 0xb1, 0xe5, 0x3b, 0x13, 0x29, 0x44, 0x6d, 0xeb, 0x9d, 0x8c, + 0x7a, 0xe3, 0xe0, 0xdb, 0x62, 0x21, 0x56, 0x00, 0xdc, 0x55, 0xe7, 0x76, 0x96, 0xda, 0xd4, 0x9f, + 0x81, 0x36, 0x88, 0x2c, 0x1a, 0xa5, 0xc5, 0x31, 0xa0, 0xc8, 0xf7, 0x57, 0xee, 0xb9, 0xc8, 0x9e, + 0xf2, 0x98, 0x61, 0xb1, 0x5c, 0xff, 0x57, 0x1e, 0x6e, 0xa4, 0xb0, 0x95, 0xdb, 0x3d, 0x85, 0x32, + 0x25, 0x6c, 0x3a, 0x89, 0x04, 0x7c, 0x63, 0xeb, 0xc3, 0x8c, 0xf0, 0x97, 0x90, 0xda, 0x58, 0xc0, + 0x60, 0x05, 0x87, 0x36, 0x40, 0x93, 0x2b, 0x4c, 0x42, 0x69, 0x40, 0x4d, 0x8f, 0x8d, 0x94, 0xe9, + 0x1a, 0x92, 0x6f, 0x70, 0xf6, 0x1e, 0x1b, 0xa5, 0xb4, 0x5a, 0xb8, 0xa6, 0x56, 0x91, 0x05, 0x9a, + 0x4f, 0xa2, 0xe7, 0x01, 0x3d, 0x31, 0xb9, 0x6a, 0xa9, 0xeb, 0x90, 0x66, 0x51, 0x80, 0xbe, 0x97, + 0x11, 0xb4, 0x2f, 0x97, 0xef, 0xab, 0xd5, 0x78, 0xdd, 0x9f, 0x67, 0xe8, 0x6f, 0x42, 0x59, 0xfe, + 0x53, 0xee, 0x49, 0x83, 0xc3, 0x6e, 0xd7, 0x18, 0x0c, 0xb4, 0x15, 0x54, 0x85, 0x12, 0x36, 0x86, + 0x98, 0x7b, 0x58, 0x15, 0x4a, 0x8f, 0x3a, 0xc3, 0xce, 0xae, 0x96, 0xd7, 0xdf, 0x80, 0xf5, 0xa7, + 0x96, 0x1b, 0x65, 0x71, 0x2e, 0x3d, 0x00, 0x6d, 0x36, 0x57, 0x59, 0x67, 0x67, 0xce, 0x3a, 0xd9, + 0x55, 0x63, 0x9c, 0xb9, 0xd1, 0x05, 0x7b, 0x68, 0x50, 0x20, 0x94, 0x2a, 0x13, 0xf0, 0x4f, 0xfd, + 0x39, 0xac, 0x0f, 0xa2, 0x20, 0xcc, 0xe4, 0xf9, 0xef, 0xc2, 0x2a, 0xcf, 0x13, 0xc1, 0x34, 0x52, + 0xae, 0xff, 0x72, 0x5b, 0xe6, 0x91, 0x76, 0x9c, 0x47, 0xda, 0x3d, 0x95, 0x67, 0x70, 0x3c, 0x13, + 0xfd, 0x0f, 0x94, 0x99, 0x3b, 0xf2, 0xad, 0x89, 0x3a, 0xfa, 0x8a, 0xd2, 0x11, 0x77, 0xf2, 0x78, + 0x63, 0xe5, 0xf8, 0x5d, 0x40, 0x3d, 0xc2, 0x22, 0x1a, 0x9c, 0x67, 0x92, 0xe7, 0x16, 0x94, 0x8e, + 0x03, 0x6a, 0xcb, 0x83, 0x58, 0xc1, 0x92, 0xe0, 0x87, 0x6a, 0x0e, 0x44, 0x61, 0xdf, 0x05, 0xb4, + 0xe3, 0xf3, 0x04, 0x91, 0xcd, 0x10, 0xbf, 0xcc, 0xc3, 0xcd, 0xb9, 0xf9, 0xca, 0x18, 0xcb, 0x9f, + 0x43, 0x1e, 0x98, 0xa6, 0x4c, 0x9e, 0x43, 0xb4, 0x0f, 0x65, 0x39, 0x43, 0x69, 0xf2, 0xde, 0x02, + 0x40, 0x32, 0xe7, 0x28, 0x38, 0x05, 0x73, 0xa5, 0xd3, 0x17, 0xfe, 0xdb, 0x4e, 0xaf, 0xc5, 0xff, + 0x83, 0x7d, 0xad, 0xfe, 0xbe, 0x0d, 0x37, 0x52, 0x93, 0x95, 0xf2, 0x1e, 0x41, 0x89, 0x71, 0x86, + 0xd2, 0xde, 0xdb, 0x0b, 0x6a, 0x8f, 0x61, 0xb9, 0x5c, 0xbf, 0x29, 0xc1, 0x8d, 0x53, 0xe2, 0x27, + 0xa2, 0xe8, 0x3d, 0xb8, 0x31, 0x10, 0xae, 0x95, 0xc9, 0x77, 0x66, 0x6e, 0x99, 0x9f, 0x73, 0xcb, + 0x5b, 0x80, 0xd2, 0x28, 0xca, 0x79, 0xce, 0x61, 0xdd, 0x38, 0x23, 0x76, 0x26, 0xe4, 0x26, 0xac, + 0xda, 0x81, 0xe7, 0x59, 0xbe, 0xd3, 0xcc, 0xdf, 0x29, 0x6c, 0x54, 0x71, 0x4c, 0xa6, 0xcf, 0x4f, + 0x21, 0xeb, 0xf9, 0xd1, 0x7f, 0x91, 0x03, 0x6d, 0xb6, 0xb7, 0x52, 0x24, 0x97, 0x3e, 0x72, 0x38, + 0x10, 0xdf, 0xbb, 0x8e, 0x15, 0xa5, 0xf8, 0xf1, 0x11, 0x97, 0x7c, 0x42, 0x69, 0x2a, 0x84, 0x14, + 0xae, 0x19, 0x42, 0xf4, 0x7f, 0xe4, 0x00, 0x5d, 0xae, 0x7a, 0xd0, 0x6b, 0x50, 0x67, 0xc4, 0x77, + 0x4c, 0xa9, 0x46, 0x69, 0xe1, 0x0a, 0xae, 0x71, 0x9e, 0xd4, 0x27, 0x43, 0x08, 0x8a, 0xe4, 0x8c, + 0xd8, 0xea, 0xb4, 0x8a, 0x6f, 0x34, 0x86, 0xfa, 0x31, 0x33, 0x5d, 0x16, 0x4c, 0xac, 0xa4, 0x3c, + 0x68, 0x6c, 0x19, 0x4b, 0x57, 0x5f, 0xed, 0x47, 0x83, 0x9d, 0x18, 0x0c, 0xd7, 0x8e, 0x59, 0x42, + 0xe8, 0x6d, 0xa8, 0xa5, 0xc6, 0x50, 0x05, 0x8a, 0xfd, 0xfd, 0xbe, 0xa1, 0xad, 0x20, 0x80, 0x72, + 0x77, 0x1b, 0xef, 0xef, 0x0f, 0x65, 0xd4, 0xde, 0xd9, 0xeb, 0x3c, 0x36, 0xb4, 0xbc, 0xfe, 0xa7, + 0x22, 0xc0, 0x2c, 0x7d, 0xa2, 0x06, 0xe4, 0x13, 0x4b, 0xe7, 0x5d, 0x87, 0xff, 0x19, 0xdf, 0xf2, + 0xe2, 0x42, 0x44, 0x7c, 0xa3, 0x2d, 0xb8, 0xed, 0xb1, 0x51, 0x68, 0xd9, 0x27, 0xa6, 0xca, 0x7a, + 0xb6, 0x58, 0x2c, 0xfe, 0x55, 0x1d, 0xdf, 0x54, 0x83, 0x4a, 0x6a, 0x89, 0xbb, 0x0b, 0x05, 0xe2, + 0x9f, 0x36, 0x8b, 0xa2, 0xd4, 0x7b, 0xb8, 0x70, 0x5a, 0x6f, 0x1b, 0xfe, 0xa9, 0x2c, 0xed, 0x38, + 0x0c, 0xea, 0x43, 0x95, 0x12, 0x16, 0x4c, 0xa9, 0x4d, 0x58, 0xb3, 0xb4, 0xd0, 0x21, 0xc3, 0xf1, + 0x3a, 0x3c, 0x83, 0x40, 0x3d, 0x28, 0x7b, 0xc1, 0xd4, 0x8f, 0x58, 0xb3, 0x2c, 0x04, 0x7c, 0x2b, + 0x23, 0xd8, 0x1e, 0x5f, 0x84, 0xd5, 0x5a, 0xf4, 0x18, 0x56, 0x1d, 0x72, 0xea, 0x72, 0x99, 0x56, + 0x05, 0xcc, 0xdd, 0xac, 0xf6, 0x15, 0xab, 0x70, 0xbc, 0x9a, 0x2b, 0x7d, 0xca, 0x08, 0x6d, 0x56, + 0xa4, 0xd2, 0xf9, 0x37, 0x7a, 0x05, 0xaa, 0xd6, 0x64, 0x12, 0xd8, 0xa6, 0xe3, 0xd2, 0x66, 0x55, + 0x0c, 0x54, 0x04, 0xa3, 0xe7, 0x52, 0xf4, 0x2a, 0xd4, 0xe4, 0xc9, 0x30, 0x43, 0x2b, 0x1a, 0x37, + 0x41, 0x0c, 0x83, 0x64, 0x1d, 0x58, 0xd1, 0x58, 0x4d, 0x20, 0x94, 0xca, 0x09, 0xb5, 0x64, 0x02, + 0xa1, 0x94, 0x4f, 0x68, 0xbd, 0x07, 0x95, 0x58, 0xc5, 0x0b, 0xd5, 0xa3, 0x7f, 0xcd, 0x41, 0x35, + 0x51, 0x29, 0xfa, 0x14, 0xd6, 0xa8, 0xf5, 0xdc, 0x9c, 0xd9, 0x46, 0x06, 0xc0, 0x77, 0xb3, 0xda, + 0xc6, 0x7a, 0x3e, 0x33, 0x4f, 0x9d, 0xa6, 0x28, 0xf4, 0x39, 0xac, 0x4f, 0x5c, 0x7f, 0x7a, 0x96, + 0xc2, 0x96, 0x19, 0xe5, 0xff, 0x33, 0x62, 0xef, 0xf2, 0xd5, 0x33, 0xf4, 0xc6, 0x64, 0x8e, 0xd6, + 0x7f, 0x9f, 0x83, 0x7a, 0x7a, 0x7b, 0xae, 0x04, 0x3b, 0x9c, 0x8a, 0x3f, 0x50, 0xc0, 0xfc, 0x93, + 0x07, 0x1d, 0x8f, 0x78, 0x01, 0x3d, 0x17, 0x3b, 0x17, 0xb0, 0xa2, 0xb8, 0xb5, 0x1c, 0x97, 0x9d, + 0x08, 0xef, 0x2f, 0x60, 0xf1, 0xcd, 0x79, 0x6e, 0x10, 0x32, 0x51, 0x8f, 0x15, 0xb0, 0xf8, 0x46, + 0x18, 0x2a, 0x2a, 0xd5, 0x70, 0x9f, 0x2d, 0x2c, 0x9e, 0xb2, 0x62, 0xe1, 0x70, 0x82, 0xa3, 0xff, + 0x3a, 0x0f, 0xeb, 0x17, 0x46, 0xb9, 0x9c, 0xd2, 0x91, 0xe2, 0x80, 0x2d, 0x29, 0x2e, 0x93, 0xed, + 0x3a, 0x71, 0x55, 0x24, 0xbe, 0xc5, 0x71, 0x0f, 0x55, 0xc5, 0x92, 0x77, 0x43, 0x6e, 0x68, 0xef, + 0xc8, 0x8d, 0xa4, 0xe0, 0x25, 0x2c, 0x09, 0xf4, 0x0c, 0x1a, 0x94, 0x30, 0x42, 0x4f, 0x89, 0x63, + 0x86, 0x01, 0x8d, 0x62, 0xf9, 0xb7, 0x16, 0x93, 0xff, 0x20, 0xa0, 0x11, 0x5e, 0x8b, 0x91, 0x38, + 0xc5, 0xd0, 0x53, 0x58, 0x73, 0xce, 0x7d, 0xcb, 0x73, 0x6d, 0x85, 0x5c, 0x5e, 0x1a, 0xb9, 0xae, + 0x80, 0x04, 0x30, 0xbf, 0x28, 0xa5, 0x06, 0xf9, 0x1f, 0x9b, 0x58, 0x47, 0x64, 0xa2, 0x74, 0x22, + 0x89, 0x79, 0xbf, 0x2e, 0x29, 0xbf, 0xd6, 0xff, 0x9d, 0x83, 0xc6, 0xbc, 0xbb, 0xa0, 0xff, 0x05, + 0xb0, 0xc3, 0xa9, 0x19, 0x12, 0xea, 0x06, 0x8e, 0x72, 0x8a, 0xaa, 0x1d, 0x4e, 0x0f, 0x04, 0x83, + 0x1f, 0x4e, 0x3e, 0xfc, 0xe5, 0x34, 0x88, 0x2c, 0xe5, 0x1d, 0x15, 0x3b, 0x9c, 0x7e, 0x93, 0xd3, + 0xf1, 0x5a, 0x71, 0x93, 0x67, 0xca, 0x4b, 0xf8, 0xf4, 0x81, 0x60, 0xa0, 0xb7, 0x00, 0x49, 0x47, + 0x32, 0x27, 0xae, 0xe7, 0x46, 0xe6, 0xd1, 0x39, 0xbf, 0x13, 0x4b, 0xc7, 0xd1, 0xe4, 0xc8, 0x2e, + 0x1f, 0xf8, 0x98, 0xf3, 0x91, 0x0e, 0x6b, 0x41, 0xe0, 0x99, 0xcc, 0x0e, 0x28, 0x31, 0x2d, 0xe7, + 0x0b, 0x11, 0xfd, 0x0a, 0xb8, 0x16, 0x04, 0xde, 0x80, 0xf3, 0x3a, 0xce, 0x17, 0xfc, 0xb0, 0xdb, + 0xe1, 0x94, 0x91, 0xc8, 0xe4, 0x3f, 0xcd, 0xb2, 0x3c, 0xec, 0x92, 0xd5, 0x0d, 0xa7, 0x2c, 0x35, + 0xc1, 0x23, 0x1e, 0x0f, 0x56, 0xa9, 0x09, 0x7b, 0xc4, 0x63, 0xfa, 0x67, 0x50, 0x12, 0xa1, 0x8d, + 0xff, 0x31, 0x91, 0xfd, 0x45, 0xd4, 0x90, 0xaa, 0xab, 0x70, 0x86, 0x08, 0x2a, 0xaf, 0x40, 0x75, + 0x1c, 0x30, 0x15, 0x73, 0xa4, 0x57, 0x55, 0x38, 0x43, 0x0c, 0xb6, 0xa0, 0x42, 0x89, 0xe5, 0x04, + 0xfe, 0xe4, 0x5c, 0xfc, 0xe7, 0x0a, 0x4e, 0x68, 0xfd, 0x4b, 0x28, 0xcb, 0x90, 0x77, 0x0d, 0xfc, + 0xbb, 0x80, 0xec, 0x11, 0x0d, 0xa6, 0x21, 0x37, 0x8a, 0xe7, 0x32, 0xe6, 0x06, 0x3e, 0x8b, 0xaf, + 0xdd, 0x72, 0xe4, 0x60, 0x36, 0xa0, 0xff, 0x25, 0x27, 0xd3, 0x9c, 0xbc, 0x43, 0xf1, 0x42, 0x41, + 0xe5, 0xac, 0xa5, 0x2f, 0x9a, 0x0a, 0x20, 0x2e, 0xf6, 0x88, 0x7a, 0x5e, 0x58, 0xb4, 0xd8, 0x23, + 0xb2, 0xd8, 0x23, 0xbc, 0xb2, 0x50, 0xd9, 0x54, 0xc2, 0xc9, 0x64, 0x5a, 0x73, 0x92, 0x2a, 0x98, + 0xe8, 0xff, 0xcc, 0x25, 0xa7, 0x3d, 0xae, 0x56, 0xd1, 0xe7, 0x50, 0xe1, 0x07, 0xc7, 0xf4, 0xac, + 0x50, 0x3d, 0xa4, 0x74, 0x97, 0x2b, 0x84, 0xdb, 0xfc, 0x9c, 0xec, 0x59, 0xa1, 0x4c, 0xb3, 0xab, + 0xa1, 0xa4, 0x78, 0xd4, 0xb0, 0x9c, 0x59, 0xd4, 0xe0, 0xdf, 0xe8, 0x75, 0x68, 0x58, 0xd3, 0x28, + 0x30, 0x2d, 0xe7, 0x94, 0xd0, 0xc8, 0x65, 0x44, 0x59, 0x78, 0x8d, 0x73, 0x3b, 0x31, 0xb3, 0xf5, + 0x10, 0xea, 0x69, 0xcc, 0xaf, 0xcb, 0x2b, 0xa5, 0x74, 0x5e, 0xf9, 0x2e, 0xc0, 0xac, 0x28, 0xe3, + 0x9e, 0x40, 0xce, 0xdc, 0xc8, 0xb4, 0x03, 0x47, 0x46, 0xb5, 0x12, 0xae, 0x70, 0x46, 0x37, 0x70, + 0xc8, 0x85, 0x12, 0xb7, 0x14, 0x97, 0xb8, 0xfc, 0xdc, 0xf1, 0xa3, 0x72, 0xe2, 0x4e, 0x26, 0xc4, + 0x51, 0x12, 0x56, 0x83, 0xc0, 0x7b, 0x22, 0x18, 0xfa, 0x9f, 0xf3, 0xd2, 0x23, 0xe4, 0x05, 0x23, + 0x53, 0xe1, 0x93, 0x98, 0xba, 0x70, 0x3d, 0x53, 0x3f, 0x00, 0x60, 0x91, 0x45, 0x23, 0xe2, 0x98, + 0x56, 0xa4, 0xee, 0xec, 0xad, 0x4b, 0x35, 0xf2, 0x30, 0x7e, 0xab, 0xc4, 0x55, 0x35, 0xbb, 0x13, + 0xa1, 0xf7, 0xa1, 0x6e, 0x07, 0x5e, 0x38, 0x21, 0x6a, 0x71, 0xe9, 0x6b, 0x17, 0xd7, 0x92, 0xf9, + 0x9d, 0x28, 0x55, 0x20, 0x97, 0xaf, 0x5b, 0x20, 0xff, 0x21, 0x27, 0xef, 0x49, 0xe9, 0x6b, 0x1a, + 0x1a, 0x5d, 0xf1, 0xb0, 0xf7, 0x78, 0xc9, 0x3b, 0xdf, 0x57, 0xbd, 0xea, 0x5d, 0xf7, 0x19, 0xed, + 0x8f, 0x05, 0xa8, 0x26, 0xd7, 0xad, 0x4b, 0xb6, 0xbf, 0x0f, 0xd5, 0xe4, 0xa1, 0x58, 0x95, 0x19, + 0x5f, 0x69, 0x9e, 0x64, 0x32, 0x3a, 0x06, 0x64, 0x8d, 0x46, 0x49, 0x91, 0x62, 0x4e, 0x99, 0x35, + 0x8a, 0x2f, 0xa8, 0xf7, 0x17, 0xd0, 0x43, 0x9c, 0x79, 0x0e, 0xf9, 0x7a, 0xac, 0x59, 0xa3, 0xd1, + 0x1c, 0x07, 0x7d, 0x0f, 0x6e, 0xcf, 0xef, 0x61, 0x1e, 0x9d, 0x9b, 0xa1, 0xeb, 0xa8, 0x02, 0x7b, + 0x7b, 0xd1, 0x1b, 0x67, 0x7b, 0x0e, 0xfe, 0xe3, 0xf3, 0x03, 0xd7, 0x91, 0x3a, 0x47, 0xf4, 0xd2, + 0x40, 0xeb, 0x07, 0xf0, 0xd2, 0x0b, 0xa6, 0x5f, 0x61, 0x83, 0x7e, 0xda, 0x06, 0xd7, 0x51, 0x42, + 0xca, 0x7a, 0xbf, 0xc9, 0xc9, 0x8b, 0xf1, 0xbc, 0x4e, 0x3a, 0xb3, 0x8a, 0xad, 0xb6, 0xb5, 0x99, + 0x71, 0x9f, 0xee, 0xc1, 0xa1, 0x84, 0x17, 0x25, 0xde, 0x27, 0x73, 0x25, 0x5e, 0xf6, 0x32, 0x64, + 0x4f, 0x2c, 0x92, 0x40, 0x0a, 0x41, 0xff, 0x5d, 0x01, 0x2a, 0x31, 0xba, 0xa8, 0xbf, 0xcf, 0x59, + 0x44, 0x3c, 0xd3, 0x8b, 0x43, 0x58, 0x0e, 0x83, 0x64, 0xed, 0xf1, 0x20, 0xf6, 0x0a, 0x54, 0x79, + 0x99, 0x2f, 0x87, 0xf3, 0x62, 0xb8, 0xc2, 0x19, 0x62, 0xf0, 0x55, 0xa8, 0x45, 0x41, 0x64, 0x4d, + 0xcc, 0xc8, 0xb5, 0x4f, 0x64, 0x92, 0xcb, 0x61, 0x10, 0xac, 0x21, 0xe7, 0xa0, 0x37, 0xe1, 0x46, + 0x34, 0xa6, 0x41, 0x14, 0x4d, 0x78, 0x85, 0x26, 0x6a, 0x12, 0x59, 0x42, 0x14, 0xb1, 0x96, 0x0c, + 0xc8, 0x5a, 0x85, 0xf1, 0xe8, 0x3d, 0x9b, 0xcc, 0x5d, 0x57, 0x04, 0x91, 0x22, 0x5e, 0x4b, 0xb8, + 0xdc, 0xb5, 0xf9, 0xfd, 0x3e, 0x24, 0xd4, 0x26, 0xbe, 0x8c, 0x15, 0x39, 0x1c, 0x93, 0xc8, 0x84, + 0x75, 0x8f, 0x58, 0x6c, 0x4a, 0x89, 0x63, 0x1e, 0xbb, 0x64, 0xe2, 0xc8, 0xfb, 0x4e, 0x23, 0x73, + 0x3d, 0x1b, 0xab, 0xa5, 0xfd, 0x48, 0xac, 0xc6, 0x8d, 0x18, 0x4e, 0xd2, 0xbc, 0x3e, 0x90, 0x5f, + 0x68, 0x1d, 0x6a, 0x83, 0x67, 0x83, 0xa1, 0xb1, 0x67, 0xee, 0xed, 0xf7, 0x0c, 0xf5, 0xc0, 0x3d, + 0x30, 0xb0, 0x24, 0x73, 0x7c, 0x7c, 0xb8, 0x3f, 0xec, 0xec, 0x9a, 0xc3, 0x9d, 0xee, 0x93, 0x81, + 0x96, 0x47, 0xb7, 0xe1, 0xc6, 0x70, 0x1b, 0xef, 0x0f, 0x87, 0xbb, 0x46, 0xcf, 0x3c, 0x30, 0xf0, + 0xce, 0x7e, 0x6f, 0xa0, 0x15, 0x10, 0x82, 0xc6, 0x8c, 0x3d, 0xdc, 0xd9, 0x33, 0xb4, 0x22, 0xaa, + 0xc1, 0xea, 0x81, 0x81, 0xbb, 0x46, 0x7f, 0xa8, 0x95, 0xf4, 0xbf, 0xe5, 0xa1, 0x96, 0xb2, 0x22, + 0x77, 0x64, 0xca, 0xe4, 0xfd, 0xa5, 0x88, 0xf9, 0x27, 0x0f, 0x26, 0xb6, 0x65, 0x8f, 0xa5, 0x75, + 0x8a, 0x58, 0x12, 0xdc, 0x6e, 0x9e, 0x75, 0x96, 0x3a, 0xe7, 0x45, 0x5c, 0xf1, 0xac, 0x33, 0x09, + 0xf2, 0x1a, 0xd4, 0x4f, 0x08, 0xf5, 0xc9, 0x44, 0x8d, 0x4b, 0x8b, 0xd4, 0x24, 0x4f, 0x4e, 0xd9, + 0x00, 0x4d, 0x4d, 0x99, 0xc1, 0x48, 0x73, 0x34, 0x24, 0x7f, 0x2f, 0x06, 0x3b, 0xba, 0xac, 0xf5, + 0xb2, 0xd0, 0xfa, 0x83, 0xc5, 0x9d, 0xf4, 0x45, 0x8a, 0x1f, 0x24, 0x8a, 0x5f, 0x85, 0x02, 0x8e, + 0xdf, 0x7a, 0xbb, 0x9d, 0xee, 0x36, 0x57, 0xf6, 0x1a, 0x54, 0xf7, 0x3a, 0x9f, 0x9a, 0x87, 0x03, + 0xf1, 0x72, 0x80, 0x34, 0xa8, 0x3f, 0x31, 0x70, 0xdf, 0xd8, 0x55, 0x9c, 0x02, 0xba, 0x05, 0x9a, + 0xe2, 0xcc, 0xe6, 0x15, 0xf5, 0xdf, 0xe6, 0x61, 0x5d, 0xc6, 0xf5, 0xe4, 0x31, 0xeb, 0xc5, 0xaf, + 0x4a, 0xcb, 0x87, 0xde, 0x26, 0xac, 0x7a, 0x84, 0x25, 0x76, 0xa8, 0xe2, 0x98, 0x44, 0x2e, 0xd4, + 0x2c, 0xdf, 0x0f, 0x22, 0xf1, 0x22, 0xc2, 0x54, 0x88, 0x7c, 0xbc, 0xd0, 0xdb, 0x4b, 0x22, 0x79, + 0xbb, 0x33, 0x43, 0x92, 0x11, 0x32, 0x8d, 0xdd, 0xfa, 0x00, 0xb4, 0x8b, 0x13, 0x16, 0xc9, 0x4b, + 0x6f, 0xbc, 0x33, 0x4b, 0x4b, 0x84, 0x3b, 0xe8, 0x61, 0xff, 0x49, 0x7f, 0xff, 0x69, 0x5f, 0x5b, + 0xe1, 0x04, 0x3e, 0xec, 0xf7, 0x77, 0xfa, 0x8f, 0xb5, 0x1c, 0x02, 0x28, 0x1b, 0x9f, 0xee, 0x0c, + 0x8d, 0x9e, 0x96, 0xdf, 0xfa, 0xfb, 0x1a, 0x94, 0xa5, 0x90, 0xe8, 0x57, 0x2a, 0x25, 0xa7, 0x3b, + 0x8e, 0xe8, 0x83, 0x85, 0x4b, 0xdb, 0xb9, 0x2e, 0x66, 0xeb, 0xc3, 0xa5, 0xd7, 0xab, 0x47, 0xc5, + 0x15, 0xf4, 0xb3, 0x1c, 0xd4, 0xe7, 0x5e, 0xd1, 0xb2, 0x3e, 0x00, 0x5d, 0xd1, 0xe0, 0x6c, 0x7d, + 0x63, 0xa9, 0xb5, 0x89, 0x2c, 0x3f, 0xcd, 0x41, 0x2d, 0xd5, 0xda, 0x43, 0x0f, 0x96, 0x69, 0x07, + 0x4a, 0x49, 0x1e, 0x2e, 0xdf, 0x49, 0xd4, 0x57, 0xde, 0xce, 0xa1, 0x9f, 0xe4, 0xa0, 0x96, 0xea, + 0x8b, 0x65, 0x16, 0xe5, 0x72, 0x17, 0x2f, 0xb3, 0x28, 0x57, 0xb5, 0xe1, 0x56, 0xd0, 0x0f, 0x73, + 0x50, 0x4d, 0x7a, 0x5c, 0xe8, 0xde, 0xe2, 0x5d, 0x31, 0x29, 0xc4, 0xfd, 0x65, 0xdb, 0x69, 0xfa, + 0x0a, 0xfa, 0x3e, 0x54, 0xe2, 0x86, 0x10, 0xca, 0x9a, 0x46, 0x2e, 0x74, 0x9b, 0x5a, 0xf7, 0x16, + 0x5e, 0x97, 0xde, 0x3e, 0xee, 0xd2, 0x64, 0xde, 0xfe, 0x42, 0x3f, 0xa9, 0x75, 0x6f, 0xe1, 0x75, + 0xc9, 0xf6, 0xdc, 0x13, 0x52, 0xcd, 0x9c, 0xcc, 0x9e, 0x70, 0xb9, 0x8b, 0x94, 0xd9, 0x13, 0xae, + 0xea, 0x1d, 0x49, 0x41, 0x52, 0xed, 0xa0, 0xcc, 0x82, 0x5c, 0x6e, 0x39, 0x65, 0x16, 0xe4, 0x8a, + 0xee, 0x93, 0x72, 0xc9, 0x59, 0x81, 0x7e, 0x6f, 0xe1, 0x0e, 0xca, 0x82, 0x2e, 0x79, 0xa9, 0x87, + 0xa3, 0xaf, 0xa0, 0x1f, 0xa9, 0x27, 0x03, 0xd9, 0x7e, 0x41, 0x8b, 0x40, 0xcd, 0x75, 0x6c, 0x5a, + 0xef, 0x2d, 0x97, 0x6a, 0x44, 0x8c, 0xf8, 0x71, 0x0e, 0x60, 0xd6, 0xa8, 0xc9, 0x2c, 0xc4, 0xa5, + 0x0e, 0x51, 0xeb, 0xc1, 0x12, 0x2b, 0xd3, 0xc7, 0x23, 0xee, 0xcd, 0x64, 0x3e, 0x1e, 0x17, 0x1a, + 0x49, 0x99, 0x8f, 0xc7, 0xc5, 0x26, 0x90, 0xbe, 0xf2, 0xf1, 0xea, 0xb7, 0x4a, 0x32, 0xf7, 0x97, + 0xc5, 0xcf, 0xbb, 0xff, 0x09, 0x00, 0x00, 0xff, 0xff, 0x34, 0x96, 0x6b, 0x66, 0x2d, 0x24, 0x00, + 0x00, } diff --git a/plugins/drivers/proto/driver.proto b/plugins/drivers/proto/driver.proto index a5e44e36a..527419274 100644 --- a/plugins/drivers/proto/driver.proto +++ b/plugins/drivers/proto/driver.proto @@ -373,7 +373,7 @@ message LinuxResources { // CPU shares (relative weight vs. other containers). Default: 0 (not specified) int64 cpu_shares = 3; // Memory limit in bytes. Default: 0 (not specified) - int64 memory_limit_in_bytes = 4; + int64 memory_limit_bytes = 4; // OOMScoreAdj adjusts the oom-killer score. Default: 0 (not specified) int64 oom_score_adj = 5; // CpusetCpus constrains the allowed set of logical CPUs. Default: "" (not specified) diff --git a/plugins/drivers/server.go b/plugins/drivers/server.go index 4392626b6..80e232d06 100644 --- a/plugins/drivers/server.go +++ b/plugins/drivers/server.go @@ -8,6 +8,7 @@ import ( "github.com/golang/protobuf/ptypes" hclog "github.com/hashicorp/go-hclog" plugin "github.com/hashicorp/go-plugin" + cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/plugins/drivers/proto" ) @@ -42,12 +43,14 @@ func (b *driverPluginServer) Capabilities(ctx context.Context, req *proto.Capabi } switch caps.FSIsolation { - case FSIsolationNone: + case cstructs.FSIsolationNone: resp.Capabilities.FsIsolation = proto.DriverCapabilities_NONE - case FSIsolationChroot: + case cstructs.FSIsolationChroot: resp.Capabilities.FsIsolation = proto.DriverCapabilities_CHROOT - case FSIsolationImage: + case cstructs.FSIsolationImage: resp.Capabilities.FsIsolation = proto.DriverCapabilities_IMAGE + default: + resp.Capabilities.FsIsolation = proto.DriverCapabilities_NONE } return resp, nil } @@ -91,13 +94,26 @@ func (b *driverPluginServer) RecoverTask(ctx context.Context, req *proto.Recover } func (b *driverPluginServer) StartTask(ctx context.Context, req *proto.StartTaskRequest) (*proto.StartTaskResponse, error) { - handle, err := b.impl.StartTask(taskConfigFromProto(req.Task)) + handle, net, err := b.impl.StartTask(taskConfigFromProto(req.Task)) if err != nil { return nil, err } + var pbNet *proto.NetworkOverride + if net != nil { + pbNet = &proto.NetworkOverride{ + PortMap: map[string]int32{}, + Addr: net.IP, + AutoAdvertise: net.AutoAdvertise, + } + for k, v := range net.PortMap { + pbNet.PortMap[k] = int32(v) + } + } + resp := &proto.StartTaskResponse{ - Handle: taskHandleToProto(handle), + Handle: taskHandleToProto(handle), + NetworkOverride: pbNet, } return resp, nil @@ -159,16 +175,24 @@ func (b *driverPluginServer) InspectTask(ctx context.Context, req *proto.Inspect return nil, err } + var pbNet *proto.NetworkOverride + if status.NetworkOverride != nil { + pbNet = &proto.NetworkOverride{ + PortMap: map[string]int32{}, + Addr: status.NetworkOverride.IP, + AutoAdvertise: status.NetworkOverride.AutoAdvertise, + } + for k, v := range status.NetworkOverride.PortMap { + pbNet.PortMap[k] = int32(v) + } + } + resp := &proto.InspectTaskResponse{ Task: protoStatus, Driver: &proto.TaskDriverStatus{ Attributes: status.DriverAttributes, }, - NetworkOverride: &proto.NetworkOverride{ - PortMap: status.NetworkOverride.PortMap, - Addr: status.NetworkOverride.Addr, - AutoAdvertise: status.NetworkOverride.AutoAdvertise, - }, + NetworkOverride: pbNet, } return resp, nil diff --git a/plugins/drivers/testing.go b/plugins/drivers/testing.go index a3ba7b3c0..53eecb642 100644 --- a/plugins/drivers/testing.go +++ b/plugins/drivers/testing.go @@ -15,6 +15,7 @@ import ( plugin "github.com/hashicorp/go-plugin" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/logmon" + cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/helper/uuid" "github.com/hashicorp/nomad/plugins/base" @@ -150,12 +151,12 @@ type MockDriver struct { FingerprintF func(context.Context) (<-chan *Fingerprint, error) CapabilitiesF func() (*Capabilities, error) RecoverTaskF func(*TaskHandle) error - StartTaskF func(*TaskConfig) (*TaskHandle, error) + StartTaskF func(*TaskConfig) (*TaskHandle, *cstructs.DriverNetwork, error) WaitTaskF func(context.Context, string) (<-chan *ExitResult, error) StopTaskF func(string, time.Duration, string) error DestroyTaskF func(string, bool) error InspectTaskF func(string) (*TaskStatus, error) - TaskStatsF func(string) (*TaskStats, error) + TaskStatsF func(string) (*cstructs.TaskResourceUsage, error) TaskEventsF func(context.Context) (<-chan *TaskEvent, error) SignalTaskF func(string, string) error ExecTaskF func(string, []string, time.Duration) (*ExecTaskResult, error) @@ -165,9 +166,11 @@ func (d *MockDriver) TaskConfigSchema() (*hclspec.Spec, error) { return d.TaskCo func (d *MockDriver) Fingerprint(ctx context.Context) (<-chan *Fingerprint, error) { return d.FingerprintF(ctx) } -func (d *MockDriver) Capabilities() (*Capabilities, error) { return d.CapabilitiesF() } -func (d *MockDriver) RecoverTask(h *TaskHandle) error { return d.RecoverTaskF(h) } -func (d *MockDriver) StartTask(c *TaskConfig) (*TaskHandle, error) { return d.StartTaskF(c) } +func (d *MockDriver) Capabilities() (*Capabilities, error) { return d.CapabilitiesF() } +func (d *MockDriver) RecoverTask(h *TaskHandle) error { return d.RecoverTaskF(h) } +func (d *MockDriver) StartTask(c *TaskConfig) (*TaskHandle, *cstructs.DriverNetwork, error) { + return d.StartTaskF(c) +} func (d *MockDriver) WaitTask(ctx context.Context, id string) (<-chan *ExitResult, error) { return d.WaitTaskF(ctx, id) } @@ -178,7 +181,9 @@ func (d *MockDriver) DestroyTask(taskID string, force bool) error { return d.DestroyTaskF(taskID, force) } func (d *MockDriver) InspectTask(taskID string) (*TaskStatus, error) { return d.InspectTaskF(taskID) } -func (d *MockDriver) TaskStats(taskID string) (*TaskStats, error) { return d.TaskStats(taskID) } +func (d *MockDriver) TaskStats(taskID string) (*cstructs.TaskResourceUsage, error) { + return d.TaskStats(taskID) +} func (d *MockDriver) TaskEvents(ctx context.Context) (<-chan *TaskEvent, error) { return d.TaskEventsF(ctx) } diff --git a/plugins/drivers/utils.go b/plugins/drivers/utils.go index b74e13d4a..ba629abd5 100644 --- a/plugins/drivers/utils.go +++ b/plugins/drivers/utils.go @@ -4,7 +4,10 @@ import ( "time" "github.com/golang/protobuf/ptypes" + "github.com/hashicorp/nomad/client/allocdir" + "github.com/hashicorp/nomad/client/driver/env" cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/plugins/drivers/proto" ) @@ -44,6 +47,19 @@ func healthStateFromProto(pb proto.FingerprintResponse_HealthState) HealthState return HealthStateUndetected } +// NewTaskConfig builds a TaskConfig from a Task struct +func NewTaskConfig(task *structs.Task, taskDir *allocdir.TaskDir, env *env.TaskEnv) *TaskConfig { + return &TaskConfig{ + Name: task.Name, + Resources: &Resources{ + NomadResources: task.Resources, + }, + Env: env.Map(), + User: task.User, + AllocDir: taskDir.AllocDir, + } +} + func taskConfigFromProto(pb *proto.TaskConfig) *TaskConfig { if pb == nil { return &TaskConfig{} @@ -53,7 +69,7 @@ func taskConfigFromProto(pb *proto.TaskConfig) *TaskConfig { Name: pb.Name, Env: pb.Env, rawDriverConfig: pb.MsgpackDriverConfig, - Resources: Resources{}, //TODO + Resources: resourcesFromProto(pb.Resources), Devices: []DeviceConfig{}, //TODO Mounts: []MountConfig{}, //TODO User: pb.User, @@ -71,7 +87,7 @@ func taskConfigToProto(cfg *TaskConfig) *proto.TaskConfig { Id: cfg.ID, Name: cfg.Name, Env: cfg.Env, - Resources: &proto.Resources{}, + Resources: resourcesToProto(cfg.Resources), Mounts: []*proto.Mount{}, Devices: []*proto.Device{}, User: cfg.User, @@ -83,6 +99,106 @@ func taskConfigToProto(cfg *TaskConfig) *proto.TaskConfig { return pb } +func resourcesFromProto(pb *proto.Resources) *Resources { + var r Resources + if pb == nil { + return &r + } + + if pb.RawResources != nil { + r.NomadResources = &structs.Resources{ + CPU: int(pb.RawResources.Cpu), + MemoryMB: int(pb.RawResources.Memory), + IOPS: int(pb.RawResources.Iops), + DiskMB: int(pb.RawResources.Disk), + } + + for _, network := range pb.RawResources.Networks { + var n structs.NetworkResource + n.Device = network.Device + n.IP = network.Ip + n.CIDR = network.Cidr + n.MBits = int(network.Mbits) + for _, port := range network.ReservedPorts { + n.ReservedPorts = append(n.ReservedPorts, structs.Port{ + Label: port.Label, + Value: int(port.Value), + }) + } + for _, port := range network.DynamicPorts { + n.DynamicPorts = append(n.DynamicPorts, structs.Port{ + Label: port.Label, + Value: int(port.Value), + }) + } + r.NomadResources.Networks = append(r.NomadResources.Networks, &n) + } + } + + if pb.LinuxResources != nil { + r.LinuxResources = &LinuxResources{ + CPUPeriod: pb.LinuxResources.CpuPeriod, + CPUQuota: pb.LinuxResources.CpuQuota, + CPUShares: pb.LinuxResources.CpuShares, + MemoryLimitBytes: pb.LinuxResources.MemoryLimitBytes, + OOMScoreAdj: pb.LinuxResources.OomScoreAdj, + CpusetCPUs: pb.LinuxResources.CpusetCpus, + CpusetMems: pb.LinuxResources.CpusetMems, + } + } + + return &r +} + +func resourcesToProto(r *Resources) *proto.Resources { + var pb proto.Resources + + if r.NomadResources != nil { + pb.RawResources = &proto.RawResources{ + Cpu: int64(r.NomadResources.CPU), + Memory: int64(r.NomadResources.MemoryMB), + Iops: int64(r.NomadResources.IOPS), + Disk: int64(r.NomadResources.DiskMB), + Networks: []*proto.NetworkResource{}, + } + + for _, network := range r.NomadResources.Networks { + var n proto.NetworkResource + n.Device = network.Device + n.Ip = network.IP + n.Cidr = network.CIDR + n.Mbits = int32(network.MBits) + n.ReservedPorts = []*proto.NetworkPort{} + for _, port := range network.ReservedPorts { + n.ReservedPorts = append(n.ReservedPorts, &proto.NetworkPort{ + Label: port.Label, + Value: int32(port.Value), + }) + } + for _, port := range network.DynamicPorts { + n.DynamicPorts = append(n.DynamicPorts, &proto.NetworkPort{ + Label: port.Label, + Value: int32(port.Value), + }) + } + } + } + + if r.LinuxResources != nil { + pb.LinuxResources = &proto.LinuxResources{ + CpuPeriod: r.LinuxResources.CPUPeriod, + CpuQuota: r.LinuxResources.CPUQuota, + CpuShares: r.LinuxResources.CPUShares, + MemoryLimitBytes: r.LinuxResources.MemoryLimitBytes, + OomScoreAdj: r.LinuxResources.OOMScoreAdj, + CpusetCpus: r.LinuxResources.CpusetCPUs, + CpusetMems: r.LinuxResources.CpusetMems, + } + } + + return &pb +} + func taskHandleFromProto(pb *proto.TaskHandle) *TaskHandle { if pb == nil { return &TaskHandle{} @@ -158,26 +274,25 @@ func taskStatusFromProto(pb *proto.TaskStatus) (*TaskStatus, error) { }, nil } -func taskStatsToProto(stats *TaskStats) (*proto.TaskStats, error) { +func taskStatsToProto(stats *cstructs.TaskResourceUsage) (*proto.TaskStats, error) { timestamp, err := ptypes.TimestampProto(time.Unix(stats.Timestamp, 0)) if err != nil { return nil, err } pids := map[string]*proto.TaskResourceUsage{} - for pid, ru := range stats.ResourceUsageByPid { + for pid, ru := range stats.Pids { pids[pid] = resourceUsageToProto(ru) } return &proto.TaskStats{ - Id: stats.ID, Timestamp: timestamp, - AggResourceUsage: resourceUsageToProto(stats.AggResourceUsage), + AggResourceUsage: resourceUsageToProto(stats.ResourceUsage), ResourceUsageByPid: pids, }, nil } -func taskStatsFromProto(pb *proto.TaskStats) (*TaskStats, error) { +func taskStatsFromProto(pb *proto.TaskStats) (*cstructs.TaskResourceUsage, error) { timestamp, err := ptypes.Timestamp(pb.Timestamp) if err != nil { return nil, err @@ -188,11 +303,10 @@ func taskStatsFromProto(pb *proto.TaskStats) (*TaskStats, error) { pids[pid] = resourceUsageFromProto(ru) } - stats := &TaskStats{ - ID: pb.Id, - Timestamp: timestamp.Unix(), - AggResourceUsage: resourceUsageFromProto(pb.AggResourceUsage), - ResourceUsageByPid: pids, + stats := &cstructs.TaskResourceUsage{ + Timestamp: timestamp.Unix(), + ResourceUsage: resourceUsageFromProto(pb.AggResourceUsage), + Pids: pids, } return stats, nil From f192c3752ab301b7429ffdfaf4bfddb0fee2370e Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Fri, 5 Oct 2018 21:42:15 -0400 Subject: [PATCH 120/152] client: refactor post allocrunnerv2 finalization --- client/allocrunner/alloc_runner.go | 54 +- .../allocrunner/interfaces/task_lifecycle.go | 2 +- .../allocrunner/taskrunner/driver_handle.go | 2 +- client/allocrunner/taskrunner/task_runner.go | 202 +++- .../taskrunner/task_runner_getters.go | 2 +- client/allocrunnerv2/alloc_runner.go | 550 ----------- .../allocrunnerv2/taskrunner/task_runner.go | 903 ------------------ client/client.go | 4 +- command/agent/consul/script.go | 2 +- command/agent/consul/structs.go | 2 +- 10 files changed, 195 insertions(+), 1528 deletions(-) delete mode 100644 client/allocrunnerv2/alloc_runner.go delete mode 100644 client/allocrunnerv2/taskrunner/task_runner.go diff --git a/client/allocrunner/alloc_runner.go b/client/allocrunner/alloc_runner.go index f0e09b5cd..6f67e0f1f 100644 --- a/client/allocrunner/alloc_runner.go +++ b/client/allocrunner/alloc_runner.go @@ -21,6 +21,7 @@ import ( "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/helper" "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/shared/loader" ) // allocRunner is used to run all the tasks in a given allocation @@ -78,6 +79,13 @@ type allocRunner struct { // prevAllocWatcher allows waiting for a previous allocation to exit // and if necessary migrate its alloc dir. prevAllocWatcher allocwatcher.PrevAllocWatcher + + // pluginLoader is used to load plugins. + pluginLoader loader.PluginCatalog + + // pluginSingletonLoader is a plugin loader that will returns singleton + // instances of the plugins. + pluginSingletonLoader loader.PluginCatalog } // NewAllocRunner returns a new allocation runner. @@ -89,18 +97,20 @@ func NewAllocRunner(config *Config) (*allocRunner, error) { } ar := &allocRunner{ - id: alloc.ID, - alloc: alloc, - clientConfig: config.ClientConfig, - consulClient: config.Consul, - vaultClient: config.Vault, - tasks: make(map[string]*taskrunner.TaskRunner, len(tg.Tasks)), - waitCh: make(chan struct{}), - state: &state.State{}, - stateDB: config.StateDB, - stateUpdater: config.StateUpdater, - allocBroadcaster: cstructs.NewAllocBroadcaster(), - prevAllocWatcher: config.PrevAllocWatcher, + id: alloc.ID, + alloc: alloc, + clientConfig: config.ClientConfig, + consulClient: config.Consul, + vaultClient: config.Vault, + tasks: make(map[string]*taskrunner.TaskRunner, len(tg.Tasks)), + waitCh: make(chan struct{}), + state: &state.State{}, + stateDB: config.StateDB, + stateUpdater: config.StateUpdater, + allocBroadcaster: cstructs.NewAllocBroadcaster(), + prevAllocWatcher: config.PrevAllocWatcher, + pluginLoader: config.PluginLoader, + pluginSingletonLoader: config.PluginSingletonLoader, } // Create the logger based on the allocation ID @@ -124,15 +134,17 @@ func NewAllocRunner(config *Config) (*allocRunner, error) { func (ar *allocRunner) initTaskRunners(tasks []*structs.Task) error { for _, task := range tasks { config := &taskrunner.Config{ - Alloc: ar.alloc, - ClientConfig: ar.clientConfig, - Task: task, - TaskDir: ar.allocDir.NewTaskDir(task.Name), - Logger: ar.logger, - StateDB: ar.stateDB, - StateUpdater: ar, - Consul: ar.consulClient, - VaultClient: ar.vaultClient, + Alloc: ar.alloc, + ClientConfig: ar.clientConfig, + Task: task, + TaskDir: ar.allocDir.NewTaskDir(task.Name), + Logger: ar.logger, + StateDB: ar.stateDB, + StateUpdater: ar, + Consul: ar.consulClient, + VaultClient: ar.vaultClient, + PluginLoader: ar.pluginLoader, + PluginSingletonLoader: ar.pluginSingletonLoader, } // Create, but do not Run, the task runner diff --git a/client/allocrunner/interfaces/task_lifecycle.go b/client/allocrunner/interfaces/task_lifecycle.go index 9fe16097a..6b6314a69 100644 --- a/client/allocrunner/interfaces/task_lifecycle.go +++ b/client/allocrunner/interfaces/task_lifecycle.go @@ -3,7 +3,7 @@ package interfaces import ( "context" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/client/driver/env" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" diff --git a/client/allocrunner/taskrunner/driver_handle.go b/client/allocrunner/taskrunner/driver_handle.go index 708209b45..9aea784d9 100644 --- a/client/allocrunner/taskrunner/driver_handle.go +++ b/client/allocrunner/taskrunner/driver_handle.go @@ -4,7 +4,7 @@ import ( "context" "time" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/plugins/drivers" diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index 3f93313ca..87f060e22 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -2,25 +2,33 @@ package taskrunner import ( "context" + "errors" "fmt" "sync" "time" metrics "github.com/armon/go-metrics" log "github.com/hashicorp/go-hclog" + "github.com/hashicorp/hcl2/hcl" + "github.com/hashicorp/hcl2/hcldec" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner/interfaces" + tinterfaces "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" "github.com/hashicorp/nomad/client/allocrunner/taskrunner/state" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/consul" - "github.com/hashicorp/nomad/client/driver" "github.com/hashicorp/nomad/client/driver/env" - dstructs "github.com/hashicorp/nomad/client/driver/structs" cstate "github.com/hashicorp/nomad/client/state" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/client/vaultclient" "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/drivers" + "github.com/hashicorp/nomad/plugins/shared" + "github.com/hashicorp/nomad/plugins/shared/hclspec" + "github.com/hashicorp/nomad/plugins/shared/loader" + "github.com/zclconf/go-cty/cty" ) const ( @@ -69,6 +77,10 @@ type TaskRunner struct { // stateDB is for persisting localState and taskState stateDB cstate.StateDB + // persistedHash is the hash of the last persisted state for skipping + // unnecessary writes + persistedHash []byte + // ctx is the task runner's context representing the tasks's lifecycle. // Canceling the context will cause the task to be destroyed. ctx context.Context @@ -90,16 +102,22 @@ type TaskRunner struct { waitCh chan struct{} // driver is the driver for the task. - driver driver.Driver + driver drivers.DriverPlugin + + // driverCapabilities is the set capabilities the driver supports + driverCapabilities *drivers.Capabilities + + // taskSchema is the hcl spec for the task driver configuration + taskSchema hcldec.Spec // handleLock guards access to handle and handleResult handleLock sync.Mutex // handle to the running driver - handle driver.DriverHandle + handle tinterfaces.DriverHandle - // handleResult proxies wait results from drivers - handleResult *handleResult + // network is the configuration for the driver network if one was created + network *cstructs.DriverNetwork // task is the task being run task *structs.Task @@ -142,6 +160,13 @@ type TaskRunner struct { // LatestResourceUsage. May be nil at all times. resourceUsage *cstructs.TaskResourceUsage resourceUsageLock sync.Mutex + + // PluginLoader is used to load plugins. + pluginLoader loader.PluginCatalog + + // PluginSingletonLoader is a plugin loader that will returns singleton + // instances of the plugins. + pluginSingletonLoader loader.PluginCatalog } type Config struct { @@ -163,6 +188,13 @@ type Config struct { // StateUpdater is used to emit updated task state StateUpdater interfaces.TaskStateHandler + + // PluginLoader is used to load plugins. + PluginLoader loader.PluginCatalog + + // PluginSingletonLoader is a plugin loader that will returns singleton + // instances of the plugins. + PluginSingletonLoader loader.PluginCatalog } func NewTaskRunner(config *Config) (*TaskRunner, error) { @@ -178,23 +210,25 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { ) tr := &TaskRunner{ - alloc: config.Alloc, - allocID: config.Alloc.ID, - clientConfig: config.ClientConfig, - task: config.Task, - taskDir: config.TaskDir, - taskName: config.Task.Name, - envBuilder: envBuilder, - consulClient: config.Consul, - vaultClient: config.VaultClient, - state: config.Alloc.TaskStates[config.Task.Name].Copy(), - localState: config.LocalState, - stateDB: config.StateDB, - stateUpdater: config.StateUpdater, - ctx: trCtx, - ctxCancel: trCancel, - triggerUpdateCh: make(chan struct{}, triggerUpdateChCap), - waitCh: make(chan struct{}), + alloc: config.Alloc, + allocID: config.Alloc.ID, + clientConfig: config.ClientConfig, + task: config.Task, + taskDir: config.TaskDir, + taskName: config.Task.Name, + envBuilder: envBuilder, + consulClient: config.Consul, + vaultClient: config.VaultClient, + state: config.Alloc.TaskStates[config.Task.Name].Copy(), + localState: config.LocalState, + stateDB: config.StateDB, + stateUpdater: config.StateUpdater, + ctx: trCtx, + ctxCancel: trCancel, + triggerUpdateCh: make(chan struct{}, triggerUpdateChCap), + waitCh: make(chan struct{}), + pluginLoader: config.PluginLoader, + pluginSingletonLoader: config.PluginSingletonLoader, } // Create the logger based on the allocation ID @@ -261,7 +295,7 @@ func (tr *TaskRunner) initLabels() { func (tr *TaskRunner) Run() { defer close(tr.waitCh) - var waitRes *dstructs.WaitResult + var result *drivers.ExitResult // Updates are handled asynchronously with the other hooks but each // triggered update - whether due to alloc updates or a new vault token @@ -295,18 +329,22 @@ MAIN: // Grab the result proxy and wait for task to exit { - _, result := tr.getDriverHandle() + handle := tr.getDriverHandle() // Do *not* use tr.ctx here as it would cause Wait() to // unblock before the task exits when Kill() is called. - waitRes = result.Wait(context.Background()) + if resultCh, err := handle.WaitCh(context.Background()); err != nil { + tr.logger.Error("wait task failed", "error", err) + } else { + result = <-resultCh + } } // Clear the handle tr.clearDriverHandle() // Store the wait result on the restart tracker - tr.restartTracker.SetWaitResult(waitRes) + tr.restartTracker.SetExitResult(result) if err := tr.exited(); err != nil { tr.logger.Error("exited hooks failed", "error", err) @@ -329,11 +367,11 @@ MAIN: // If task terminated, update server. All other exit conditions (eg // killed or out of restarts) will perform their own server updates. - if waitRes != nil { + if result != nil { event := structs.NewTaskEvent(structs.TaskTerminated). - SetExitCode(waitRes.ExitCode). - SetSignal(waitRes.Signal). - SetExitMessage(waitRes.Err) + SetExitCode(result.ExitCode). + SetSignal(result.Signal). + SetExitMessage(result.Err) tr.UpdateState(structs.TaskStateDead, event) } @@ -398,37 +436,67 @@ func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { // runDriver runs the driver and waits for it to exit func (tr *TaskRunner) runDriver() error { - // Run prestart - ctx := driver.NewExecContext(tr.taskDir, tr.envBuilder.Build()) - _, err := tr.driver.Prestart(ctx, tr.task) - if err != nil { - tr.logger.Error("driver pre-start failed", "error", err) + + taskConfig := drivers.NewTaskConfig(tr.task, tr.taskDir, tr.envBuilder.Build()) + taskConfig.ID = tr.buildID() + taskConfig.StdoutPath = tr.logmonHookConfig.stdoutFifo + taskConfig.StderrPath = tr.logmonHookConfig.stderrFifo + + evalCtx := &hcl.EvalContext{ + Functions: shared.GetStdlibFuncs(), + Variables: map[string]cty.Value{ + "NOMAD_ENV_bin": cty.StringVal("/bin/consul"), + }, + } + + val, diag := shared.ParseHclInterface(tr.task.Config, tr.taskSchema, evalCtx) + if diag.HasErrors() { + errStr := "failed to parse config" + for _, err := range diag.Errs() { + errStr = fmt.Sprintf("%s\n* %s", errStr, err.Error()) + } + return errors.New(errStr) + } + + if err := taskConfig.EncodeDriverConfig(val); err != nil { + tr.logger.Warn("failed to encode driver config", "error", err) return err } - // Create a new context for Start since the environment may have been updated. - ctx = driver.NewExecContext(tr.taskDir, tr.envBuilder.Build()) - - ctx.StdoutFifo = tr.logmonHookConfig.stdoutFifo - ctx.StderrFifo = tr.logmonHookConfig.stderrFifo + //TODO mounts and devices + //XXX Evaluate and encode driver config // Start the job - sresp, err := tr.driver.Start(ctx, tr.task) + handle, net, err := tr.driver.StartTask(taskConfig) if err != nil { tr.logger.Warn("driver start failed", "error", err) return err } + tr.network = net - // Store the driver handle and associated metadata - tr.setDriverHandle(sresp.Handle) + tr.localStateLock.Lock() + tr.localState.TaskHandle = handle + tr.localStateLock.Unlock() + tr.updateDriverHandle(taskConfig.ID) // Emit an event that we started tr.UpdateState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) return nil } +func (tr *TaskRunner) updateDriverHandle(taskID string) { + tr.handleLock.Lock() + defer tr.handleLock.Unlock() + tr.handle = &driverHandleImpl{ + driver: tr.driver, + net: tr.network, + taskID: taskID, + task: tr.Task(), + } +} + // initDriver creates the driver for the task -func (tr *TaskRunner) initDriver() error { +/*func (tr *TaskRunner) initDriver() error { // Create a task-specific event emitter callback to expose minimal // state to drivers //XXX Replace with EmitEvent -- no need for a shim @@ -455,6 +523,37 @@ func (tr *TaskRunner) initDriver() error { } tr.driver = driver + return nil +}*/ + +// initDriver retrives the DriverPlugin from the plugin loader for this task +func (tr *TaskRunner) initDriver() error { + plugin, err := tr.pluginSingletonLoader.Dispense(tr.Task().Driver, base.PluginTypeDriver, tr.logger) + if err != nil { + return err + } + + // XXX need to be able to reattach to running drivers + driver, ok := plugin.Plugin().(drivers.DriverPlugin) + if !ok { + return fmt.Errorf("plugin loaded for driver %s does not implement DriverPlugin interface", tr.task.Driver) + } + + tr.driver = driver + + schema, err := tr.driver.TaskConfigSchema() + if err != nil { + return err + } + spec, _ := hclspec.Convert(schema) + tr.taskSchema = spec + + caps, err := tr.driver.Capabilities() + if err != nil { + return err + } + tr.driverCapabilities = caps + return nil } @@ -462,10 +561,14 @@ func (tr *TaskRunner) initDriver() error { // handleDestroy will retry with an exponential backoff and will give up at a // given limit. It returns whether the task was destroyed and the error // associated with the last kill attempt. -func (tr *TaskRunner) handleDestroy(handle driver.DriverHandle) (destroyed bool, err error) { +func (tr *TaskRunner) handleDestroy(handle tinterfaces.DriverHandle) (destroyed bool, err error) { // Cap the number of times we attempt to kill the task. for i := 0; i < killFailureLimit; i++ { if err = handle.Kill(); err != nil { + if err == drivers.ErrTaskNotFound { + tr.logger.Warn("couldn't find task to kill", "task_id", handle.ID()) + return true, nil + } // Calculate the new backoff backoff := (1 << (2 * uint64(i))) * killBackoffBaseline if backoff > killBackoffLimit { @@ -490,6 +593,11 @@ func (tr *TaskRunner) persistLocalState() error { return tr.stateDB.PutTaskRunnerLocalState(tr.allocID, tr.taskName, tr.localState) } +// buildID builds a consistent unique ID for the task from the alloc ID, task name and restart attempt +func (tr *TaskRunner) buildID() string { + return fmt.Sprintf("%s/%s/%d", tr.allocID, tr.taskName, tr.restartTracker.GetCount()) +} + // XXX If the objects don't exists since the client shutdown before the task // runner ever saved state, then we should treat it as a new task runner and not // return an error @@ -628,7 +736,7 @@ func (tr *TaskRunner) appendEvent(event *structs.TaskEvent) error { // Ensure the event is populated with human readable strings event.PopulateEventDisplayMessage() - // Propagate failure from event to task state + // Propogate failure from event to task state if event.FailsTask { tr.state.Failed = true } diff --git a/client/allocrunner/taskrunner/task_runner_getters.go b/client/allocrunner/taskrunner/task_runner_getters.go index 67d3a4810..2affbcc3e 100644 --- a/client/allocrunner/taskrunner/task_runner_getters.go +++ b/client/allocrunner/taskrunner/task_runner_getters.go @@ -1,7 +1,7 @@ package taskrunner import ( - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/nomad/structs" ) diff --git a/client/allocrunnerv2/alloc_runner.go b/client/allocrunnerv2/alloc_runner.go deleted file mode 100644 index 171d6b8c7..000000000 --- a/client/allocrunnerv2/alloc_runner.go +++ /dev/null @@ -1,550 +0,0 @@ -package allocrunnerv2 - -import ( - "context" - "fmt" - "path/filepath" - "sync" - "time" - - log "github.com/hashicorp/go-hclog" - "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunner" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - "github.com/hashicorp/nomad/client/allocrunnerv2/state" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner" - "github.com/hashicorp/nomad/client/allocwatcher" - "github.com/hashicorp/nomad/client/config" - "github.com/hashicorp/nomad/client/consul" - cinterfaces "github.com/hashicorp/nomad/client/interfaces" - cstate "github.com/hashicorp/nomad/client/state" - cstructs "github.com/hashicorp/nomad/client/structs" - "github.com/hashicorp/nomad/client/vaultclient" - "github.com/hashicorp/nomad/helper" - "github.com/hashicorp/nomad/nomad/structs" - "github.com/hashicorp/nomad/plugins/shared/loader" -) - -// allocRunner is used to run all the tasks in a given allocation -type allocRunner struct { - // id is the ID of the allocation. Can be accessed without a lock - id string - - // Logger is the logger for the alloc runner. - logger log.Logger - - clientConfig *config.Config - - // stateUpdater is used to emit updated task state - stateUpdater cinterfaces.AllocStateHandler - - // consulClient is the client used by the consul service hook for - // registering services and checks - consulClient consul.ConsulServiceAPI - - // vaultClient is the used to manage Vault tokens - vaultClient vaultclient.VaultClient - - // waitCh is closed when the Run() loop has exited - waitCh chan struct{} - - // destroyed is true when the Run() loop has exited, postrun hooks have - // run, and alloc runner has been destroyed - destroyed bool - destroyedLock sync.Mutex - - // Alloc captures the allocation being run. - alloc *structs.Allocation - allocLock sync.RWMutex - - // state is the alloc runner's state - state *state.State - stateLock sync.RWMutex - - stateDB cstate.StateDB - - // allocDir is used to build the allocations directory structure. - allocDir *allocdir.AllocDir - - // runnerHooks are alloc runner lifecycle hooks that should be run on state - // transistions. - runnerHooks []interfaces.RunnerHook - - // tasks are the set of task runners - tasks map[string]*taskrunner.TaskRunner - tasksLock sync.RWMutex - - // allocBroadcaster sends client allocation updates to all listeners - allocBroadcaster *cstructs.AllocBroadcaster - - // prevAllocWatcher allows waiting for a previous allocation to exit - // and if necessary migrate its alloc dir. - prevAllocWatcher allocwatcher.PrevAllocWatcher - - // pluginLoader is used to load plugins. - pluginLoader loader.PluginCatalog - - // pluginSingletonLoader is a plugin loader that will returns singleton - // instances of the plugins. - pluginSingletonLoader loader.PluginCatalog -} - -// NewAllocRunner returns a new allocation runner. -func NewAllocRunner(config *Config) (*allocRunner, error) { - alloc := config.Alloc - tg := alloc.Job.LookupTaskGroup(alloc.TaskGroup) - if tg == nil { - return nil, fmt.Errorf("failed to lookup task group %q", alloc.TaskGroup) - } - - ar := &allocRunner{ - id: alloc.ID, - alloc: alloc, - clientConfig: config.ClientConfig, - consulClient: config.Consul, - vaultClient: config.Vault, - tasks: make(map[string]*taskrunner.TaskRunner, len(tg.Tasks)), - waitCh: make(chan struct{}), - state: &state.State{}, - stateDB: config.StateDB, - stateUpdater: config.StateUpdater, - allocBroadcaster: cstructs.NewAllocBroadcaster(), - prevAllocWatcher: config.PrevAllocWatcher, - pluginLoader: config.PluginLoader, - pluginSingletonLoader: config.PluginSingletonLoader, - } - - // Create the logger based on the allocation ID - ar.logger = config.Logger.Named("alloc_runner").With("alloc_id", alloc.ID) - - // Create alloc dir - ar.allocDir = allocdir.NewAllocDir(ar.logger, filepath.Join(config.ClientConfig.AllocDir, alloc.ID)) - - // Initialize the runners hooks. - ar.initRunnerHooks() - - // Create the TaskRunners - if err := ar.initTaskRunners(tg.Tasks); err != nil { - return nil, err - } - - return ar, nil -} - -// initTaskRunners creates task runners but does *not* run them. -func (ar *allocRunner) initTaskRunners(tasks []*structs.Task) error { - for _, task := range tasks { - config := &taskrunner.Config{ - Alloc: ar.alloc, - ClientConfig: ar.clientConfig, - Task: task, - TaskDir: ar.allocDir.NewTaskDir(task.Name), - Logger: ar.logger, - StateDB: ar.stateDB, - StateUpdater: ar, - Consul: ar.consulClient, - VaultClient: ar.vaultClient, - PluginLoader: ar.pluginLoader, - PluginSingletonLoader: ar.pluginSingletonLoader, - } - - // Create, but do not Run, the task runner - tr, err := taskrunner.NewTaskRunner(config) - if err != nil { - return fmt.Errorf("failed creating runner for task %q: %v", task.Name, err) - } - - ar.tasks[task.Name] = tr - } - return nil -} - -func (ar *allocRunner) WaitCh() <-chan struct{} { - return ar.waitCh -} - -// XXX How does alloc Restart work -// Run is the main goroutine that executes all the tasks. -func (ar *allocRunner) Run() { - // Close the wait channel - defer close(ar.waitCh) - - var taskWaitCh <-chan struct{} - - // Run the prestart hooks - // XXX Equivalent to TR.Prestart hook - if err := ar.prerun(); err != nil { - ar.logger.Error("prerun failed", "error", err) - goto POST - } - - // Run the runners - taskWaitCh = ar.runImpl() - -MAIN: - for { - select { - case <-taskWaitCh: - // TaskRunners have all exited - break MAIN - } - } - -POST: - // Run the postrun hooks - // XXX Equivalent to TR.Poststop hook - if err := ar.postrun(); err != nil { - ar.logger.Error("postrun failed", "error", err) - } -} - -// runImpl is used to run the runners. -func (ar *allocRunner) runImpl() <-chan struct{} { - for _, task := range ar.tasks { - go task.Run() - } - - // Return a combined WaitCh that is closed when all task runners have - // exited. - waitCh := make(chan struct{}) - go func() { - defer close(waitCh) - for _, task := range ar.tasks { - <-task.WaitCh() - } - }() - - return waitCh -} - -// Alloc returns the current allocation being run by this runner as sent by the -// server. This view of the allocation does not have updated task states. -func (ar *allocRunner) Alloc() *structs.Allocation { - ar.allocLock.RLock() - defer ar.allocLock.RUnlock() - return ar.alloc -} - -func (ar *allocRunner) setAlloc(updated *structs.Allocation) { - ar.allocLock.Lock() - ar.alloc = updated - ar.allocLock.Unlock() -} - -// GetAllocDir returns the alloc dir which is safe for concurrent use. -func (ar *allocRunner) GetAllocDir() *allocdir.AllocDir { - return ar.allocDir -} - -// Restore state from database. Must be called after NewAllocRunner but before -// Run. -func (ar *allocRunner) Restore() error { - // Restore task runners - for _, tr := range ar.tasks { - if err := tr.Restore(); err != nil { - return err - } - } - - return nil -} - -// TaskStateUpdated is called by TaskRunner when a task's state has been -// updated. This hook is used to compute changes to the alloc's ClientStatus -// and to update the server with the new state. -func (ar *allocRunner) TaskStateUpdated(taskName string, state *structs.TaskState) { - // If a task is dead, we potentially want to kill other tasks in the group - if state.State == structs.TaskStateDead { - // Find all tasks that are not the one that is dead and check if the one - // that is dead is a leader - var otherTaskRunners []*taskrunner.TaskRunner - var otherTaskNames []string - leader := false - for name, tr := range ar.tasks { - if name != taskName { - otherTaskRunners = append(otherTaskRunners, tr) - otherTaskNames = append(otherTaskNames, name) - } else if tr.Task().Leader { - leader = true - } - } - - // If the task failed, we should kill all the other tasks in the task group. - if state.Failed { - if len(otherTaskRunners) > 0 { - ar.logger.Debug("task failure, destroying all tasks", "failed_task", taskName, "destroying", otherTaskNames) - } - for _, tr := range otherTaskRunners { - tr.Kill(context.Background(), structs.NewTaskEvent(structs.TaskSiblingFailed).SetFailedSibling(taskName)) - } - } else if leader { - if len(otherTaskRunners) > 0 { - ar.logger.Debug("leader task dead, destroying all tasks", "leader_task", taskName, "destroying", otherTaskNames) - } - // If the task was a leader task we should kill all the other tasks. - for _, tr := range otherTaskRunners { - tr.Kill(context.Background(), structs.NewTaskEvent(structs.TaskLeaderDead)) - } - } - } - - // Gather the state of the other tasks - ar.tasksLock.RLock() - states := make(map[string]*structs.TaskState, len(ar.tasks)) - for name, tr := range ar.tasks { - if name == taskName { - states[name] = state - } else { - states[name] = tr.TaskState() - } - } - ar.tasksLock.RUnlock() - - // Get the client allocation - calloc := ar.clientAlloc(states) - - // Update the server - ar.stateUpdater.AllocStateUpdated(calloc) - - // Broadcast client alloc to listeners - ar.allocBroadcaster.Send(calloc) -} - -// clientAlloc takes in the task states and returns an Allocation populated -// with Client specific fields -func (ar *allocRunner) clientAlloc(taskStates map[string]*structs.TaskState) *structs.Allocation { - ar.stateLock.RLock() - defer ar.stateLock.RUnlock() - - // store task states for AllocState to expose - ar.state.TaskStates = taskStates - - a := &structs.Allocation{ - ID: ar.id, - TaskStates: taskStates, - } - - if d := ar.state.DeploymentStatus; d != nil { - a.DeploymentStatus = d.Copy() - } - - // Compute the ClientStatus - if ar.state.ClientStatus != "" { - // The client status is being forced - a.ClientStatus, a.ClientDescription = ar.state.ClientStatus, ar.state.ClientDescription - } else { - a.ClientStatus, a.ClientDescription = getClientStatus(taskStates) - } - - // If the allocation is terminal, make sure all required fields are properly - // set. - if a.ClientTerminalStatus() { - alloc := ar.Alloc() - - // If we are part of a deployment and the task has failed, mark the - // alloc as unhealthy. This guards against the watcher not be started. - if a.ClientStatus == structs.AllocClientStatusFailed && - alloc.DeploymentID != "" && !a.DeploymentStatus.IsUnhealthy() { - a.DeploymentStatus = &structs.AllocDeploymentStatus{ - Healthy: helper.BoolToPtr(false), - } - } - - // Make sure we have marked the finished at for every task. This is used - // to calculate the reschedule time for failed allocations. - now := time.Now() - for _, task := range alloc.Job.LookupTaskGroup(alloc.TaskGroup).Tasks { - ts, ok := a.TaskStates[task.Name] - if !ok { - ts = &structs.TaskState{} - a.TaskStates[task.Name] = ts - } - if ts.FinishedAt.IsZero() { - ts.FinishedAt = now - } - } - } - - return a -} - -// getClientStatus takes in the task states for a given allocation and computes -// the client status and description -func getClientStatus(taskStates map[string]*structs.TaskState) (status, description string) { - var pending, running, dead, failed bool - for _, state := range taskStates { - switch state.State { - case structs.TaskStateRunning: - running = true - case structs.TaskStatePending: - pending = true - case structs.TaskStateDead: - if state.Failed { - failed = true - } else { - dead = true - } - } - } - - // Determine the alloc status - if failed { - return structs.AllocClientStatusFailed, "Failed tasks" - } else if running { - return structs.AllocClientStatusRunning, "Tasks are running" - } else if pending { - return structs.AllocClientStatusPending, "No tasks have started" - } else if dead { - return structs.AllocClientStatusComplete, "All tasks have completed" - } - - return "", "" -} - -// AllocState returns a copy of allocation state including a snapshot of task -// states. -func (ar *allocRunner) AllocState() *state.State { - ar.stateLock.RLock() - state := ar.state.Copy() - ar.stateLock.RUnlock() - - // If TaskStateUpdated has not been called yet, ar.state.TaskStates - // won't be set as it is not the canonical source of TaskStates. - if len(state.TaskStates) == 0 { - ar.tasksLock.RLock() - ar.state.TaskStates = make(map[string]*structs.TaskState, len(ar.tasks)) - for k, tr := range ar.tasks { - state.TaskStates[k] = tr.TaskState() - } - ar.tasksLock.RUnlock() - } - - return state -} - -// Update the running allocation with a new version received from the server. -// -// This method sends the updated alloc to Run for serially processing updates. -// If there is already a pending update it will be discarded and replaced by -// the latest update. -func (ar *allocRunner) Update(update *structs.Allocation) { - // Update ar.alloc - ar.setAlloc(update) - - // Run hooks - if err := ar.update(update); err != nil { - ar.logger.Error("error running update hooks", "error", err) - } - - // Update task runners - for _, tr := range ar.tasks { - tr.Update(update) - } -} - -func (ar *allocRunner) Listener() *cstructs.AllocListener { - return ar.allocBroadcaster.Listen() -} - -// Destroy the alloc runner by synchronously stopping it if it is still running -// and cleaning up all of its resources. -// -// This method is safe for calling concurrently with Run() and will cause it to -// exit (thus closing WaitCh). -func (ar *allocRunner) Destroy() { - // Stop tasks - ar.tasksLock.RLock() - for name, tr := range ar.tasks { - err := tr.Kill(context.TODO(), structs.NewTaskEvent(structs.TaskKilled)) - if err != nil { - if err == taskrunner.ErrTaskNotRunning { - ar.logger.Trace("task not running", "task_name", name) - } else { - ar.logger.Warn("failed to kill task", "error", err, "task_name", name) - } - } - } - ar.tasksLock.RUnlock() - - // Wait for tasks to exit and postrun hooks to finish - <-ar.waitCh - - // Run destroy hooks - if err := ar.destroy(); err != nil { - ar.logger.Warn("error running destroy hooks", "error", err) - } - - // Cleanup state db - if err := ar.stateDB.DeleteAllocationBucket(ar.id); err != nil { - ar.logger.Warn("failed to delete allocation state", "error", err) - } - - // Mark alloc as destroyed - ar.destroyedLock.Lock() - ar.destroyed = true - ar.destroyedLock.Unlock() -} - -// IsDestroyed returns true if the alloc runner has been destroyed (stopped and -// garbage collected). -// -// This method is safe for calling concurrently with Run(). Callers must -// receive on WaitCh() to block until alloc runner has stopped and been -// destroyed. -func (ar *allocRunner) IsDestroyed() bool { - ar.destroyedLock.Lock() - defer ar.destroyedLock.Unlock() - return ar.destroyed -} - -// IsWaiting returns true if the alloc runner is waiting for its previous -// allocation to terminate. -// -// This method is safe for calling concurrently with Run(). -func (ar *allocRunner) IsWaiting() bool { - return ar.prevAllocWatcher.IsWaiting() -} - -// IsMigrating returns true if the alloc runner is migrating data from its -// previous allocation. -// -// This method is safe for calling concurrently with Run(). -func (ar *allocRunner) IsMigrating() bool { - return ar.prevAllocWatcher.IsMigrating() -} - -func (ar *allocRunner) StatsReporter() allocrunner.AllocStatsReporter { - return ar -} - -// LatestAllocStats returns the latest stats for an allocation. If taskFilter -// is set, only stats for that task -- if it exists -- are returned. -func (ar *allocRunner) LatestAllocStats(taskFilter string) (*cstructs.AllocResourceUsage, error) { - ar.tasksLock.RLock() - defer ar.tasksLock.RUnlock() - - astat := &cstructs.AllocResourceUsage{ - Tasks: make(map[string]*cstructs.TaskResourceUsage, len(ar.tasks)), - ResourceUsage: &cstructs.ResourceUsage{ - MemoryStats: &cstructs.MemoryStats{}, - CpuStats: &cstructs.CpuStats{}, - }, - } - - for name, tr := range ar.tasks { - if taskFilter != "" && taskFilter != name { - // Getting stats for a particular task and its not this one! - continue - } - - if usage := tr.LatestResourceUsage(); usage != nil { - astat.Tasks[name] = usage - astat.ResourceUsage.Add(usage.ResourceUsage) - if usage.Timestamp > astat.Timestamp { - astat.Timestamp = usage.Timestamp - } - } - } - - return astat, nil -} diff --git a/client/allocrunnerv2/taskrunner/task_runner.go b/client/allocrunnerv2/taskrunner/task_runner.go deleted file mode 100644 index 14f64789e..000000000 --- a/client/allocrunnerv2/taskrunner/task_runner.go +++ /dev/null @@ -1,903 +0,0 @@ -package taskrunner - -import ( - "context" - "errors" - "fmt" - "sync" - "time" - - metrics "github.com/armon/go-metrics" - log "github.com/hashicorp/go-hclog" - "github.com/hashicorp/hcl2/hcl" - "github.com/hashicorp/hcl2/hcldec" - "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" - "github.com/hashicorp/nomad/client/allocrunnerv2/interfaces" - tinterfaces "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/state" - "github.com/hashicorp/nomad/client/config" - "github.com/hashicorp/nomad/client/consul" - "github.com/hashicorp/nomad/client/driver/env" - cstate "github.com/hashicorp/nomad/client/state" - cstructs "github.com/hashicorp/nomad/client/structs" - "github.com/hashicorp/nomad/client/vaultclient" - "github.com/hashicorp/nomad/nomad/structs" - "github.com/hashicorp/nomad/plugins/base" - "github.com/hashicorp/nomad/plugins/drivers" - "github.com/hashicorp/nomad/plugins/shared" - "github.com/hashicorp/nomad/plugins/shared/hclspec" - "github.com/hashicorp/nomad/plugins/shared/loader" - "github.com/zclconf/go-cty/cty" -) - -const ( - // killBackoffBaseline is the baseline time for exponential backoff while - // killing a task. - killBackoffBaseline = 5 * time.Second - - // killBackoffLimit is the limit of the exponential backoff for killing - // the task. - killBackoffLimit = 2 * time.Minute - - // killFailureLimit is how many times we will attempt to kill a task before - // giving up and potentially leaking resources. - killFailureLimit = 5 - - // triggerUpdatechCap is the capacity for the triggerUpdateCh used for - // triggering updates. It should be exactly 1 as even if multiple - // updates have come in since the last one was handled, we only need to - // handle the last one. - triggerUpdateChCap = 1 -) - -type TaskRunner struct { - // allocID and taskName are immutable so these fields may be accessed - // without locks - allocID string - taskName string - - alloc *structs.Allocation - allocLock sync.Mutex - - clientConfig *config.Config - - // stateUpdater is used to emit updated task state - stateUpdater interfaces.TaskStateHandler - - // state captures the state of the task for updating the allocation - state *structs.TaskState - stateLock sync.Mutex - - // localState captures the node-local state of the task for when the - // Nomad agent restarts - localState *state.LocalState - localStateLock sync.RWMutex - - // stateDB is for persisting localState and taskState - stateDB cstate.StateDB - - // persistedHash is the hash of the last persisted state for skipping - // unnecessary writes - persistedHash []byte - - // ctx is the task runner's context representing the tasks's lifecycle. - // Canceling the context will cause the task to be destroyed. - ctx context.Context - - // ctxCancel is used to exit the task runner's Run loop without - // stopping the task. Shutdown hooks are run. - ctxCancel context.CancelFunc - - // Logger is the logger for the task runner. - logger log.Logger - - // triggerUpdateCh is ticked whenever update hooks need to be run and - // must be created with cap=1 to signal a pending update and prevent - // callers from deadlocking if the receiver has exited. - triggerUpdateCh chan struct{} - - // waitCh is closed when the task runner has transitioned to a terminal - // state - waitCh chan struct{} - - // driver is the driver for the task. - driver drivers.DriverPlugin - - // driverCapabilities is the set capabilities the driver supports - driverCapabilities *drivers.Capabilities - - // taskSchema is the hcl spec for the task driver configuration - taskSchema hcldec.Spec - - // handleLock guards access to handle and handleResult - handleLock sync.Mutex - - // handle to the running driver - handle tinterfaces.DriverHandle - - // network is the configuration for the driver network if one was created - network *cstructs.DriverNetwork - - // task is the task being run - task *structs.Task - taskLock sync.RWMutex - - // taskDir is the directory structure for this task. - taskDir *allocdir.TaskDir - - // envBuilder is used to build the task's environment - envBuilder *env.Builder - - // restartTracker is used to decide if the task should be restarted. - restartTracker *restarts.RestartTracker - - // runnerHooks are task runner lifecycle hooks that should be run on state - // transistions. - runnerHooks []interfaces.TaskHook - - // consulClient is the client used by the consul service hook for - // registering services and checks - consulClient consul.ConsulServiceAPI - - // vaultClient is the client to use to derive and renew Vault tokens - vaultClient vaultclient.VaultClient - - // vaultToken is the current Vault token. It should be accessed with the - // getter. - vaultToken string - vaultTokenLock sync.Mutex - - // baseLabels are used when emitting tagged metrics. All task runner metrics - // will have these tags, and optionally more. - baseLabels []metrics.Label - - // logmonHookConfig is used to get the paths to the stdout and stderr fifos - // to be passed to the driver for task logging - logmonHookConfig *logmonHookConfig - - // resourceUsage is written via UpdateStats and read via - // LatestResourceUsage. May be nil at all times. - resourceUsage *cstructs.TaskResourceUsage - resourceUsageLock sync.Mutex - - // PluginLoader is used to load plugins. - pluginLoader loader.PluginCatalog - - // PluginSingletonLoader is a plugin loader that will returns singleton - // instances of the plugins. - pluginSingletonLoader loader.PluginCatalog -} - -type Config struct { - Alloc *structs.Allocation - ClientConfig *config.Config - Consul consul.ConsulServiceAPI - Task *structs.Task - TaskDir *allocdir.TaskDir - Logger log.Logger - - // VaultClient is the client to use to derive and renew Vault tokens - VaultClient vaultclient.VaultClient - - // LocalState is optionally restored task state - LocalState *state.LocalState - - // StateDB is used to store and restore state. - StateDB cstate.StateDB - - // StateUpdater is used to emit updated task state - StateUpdater interfaces.TaskStateHandler - - // PluginLoader is used to load plugins. - PluginLoader loader.PluginCatalog - - // PluginSingletonLoader is a plugin loader that will returns singleton - // instances of the plugins. - PluginSingletonLoader loader.PluginCatalog -} - -func NewTaskRunner(config *Config) (*TaskRunner, error) { - // Create a context for the runner - trCtx, trCancel := context.WithCancel(context.Background()) - - // Initialize the environment builder - envBuilder := env.NewBuilder( - config.ClientConfig.Node, - config.Alloc, - config.Task, - config.ClientConfig.Region, - ) - - tr := &TaskRunner{ - alloc: config.Alloc, - allocID: config.Alloc.ID, - clientConfig: config.ClientConfig, - task: config.Task, - taskDir: config.TaskDir, - taskName: config.Task.Name, - envBuilder: envBuilder, - consulClient: config.Consul, - vaultClient: config.VaultClient, - state: config.Alloc.TaskStates[config.Task.Name].Copy(), - localState: config.LocalState, - stateDB: config.StateDB, - stateUpdater: config.StateUpdater, - ctx: trCtx, - ctxCancel: trCancel, - triggerUpdateCh: make(chan struct{}, triggerUpdateChCap), - waitCh: make(chan struct{}), - pluginLoader: config.PluginLoader, - pluginSingletonLoader: config.PluginSingletonLoader, - } - - // Create the logger based on the allocation ID - tr.logger = config.Logger.Named("task_runner").With("task", config.Task.Name) - - // Build the restart tracker. - tg := tr.alloc.Job.LookupTaskGroup(tr.alloc.TaskGroup) - if tg == nil { - tr.logger.Error("alloc missing task group") - return nil, fmt.Errorf("alloc missing task group") - } - tr.restartTracker = restarts.NewRestartTracker(tg.RestartPolicy, tr.alloc.Job.Type) - - // Initialize the task state - tr.initState() - - // Get the driver - if err := tr.initDriver(); err != nil { - tr.logger.Error("failed to create driver", "error", err) - return nil, err - } - - // Initialize the runners hooks. - tr.initHooks() - - // Initialize base labels - tr.initLabels() - - return tr, nil -} - -func (tr *TaskRunner) initState() { - if tr.state == nil { - tr.state = &structs.TaskState{ - State: structs.TaskStatePending, - } - } - if tr.localState == nil { - tr.localState = state.NewLocalState() - } -} - -func (tr *TaskRunner) initLabels() { - alloc := tr.Alloc() - tr.baseLabels = []metrics.Label{ - { - Name: "job", - Value: alloc.Job.Name, - }, - { - Name: "task_group", - Value: alloc.TaskGroup, - }, - { - Name: "alloc_id", - Value: tr.allocID, - }, - { - Name: "task", - Value: tr.taskName, - }, - } -} - -func (tr *TaskRunner) Run() { - defer close(tr.waitCh) - var result *drivers.ExitResult - - // Updates are handled asynchronously with the other hooks but each - // triggered update - whether due to alloc updates or a new vault token - // - should be handled serially. - go tr.handleUpdates() - -MAIN: - for tr.ctx.Err() == nil { - // Run the prestart hooks - if err := tr.prestart(); err != nil { - tr.logger.Error("prestart failed", "error", err) - tr.restartTracker.SetStartError(err) - goto RESTART - } - - if tr.ctx.Err() != nil { - break MAIN - } - - // Run the task - if err := tr.runDriver(); err != nil { - tr.logger.Error("running driver failed", "error", err) - tr.restartTracker.SetStartError(err) - goto RESTART - } - - // Run the poststart hooks - if err := tr.poststart(); err != nil { - tr.logger.Error("poststart failed", "error", err) - } - - // Grab the result proxy and wait for task to exit - { - handle := tr.getDriverHandle() - - // Do *not* use tr.ctx here as it would cause Wait() to - // unblock before the task exits when Kill() is called. - if resultCh, err := handle.WaitCh(context.Background()); err != nil { - tr.logger.Error("wait task failed", "error", err) - } else { - result = <-resultCh - } - } - - // Clear the handle - tr.clearDriverHandle() - - // Store the wait result on the restart tracker - tr.restartTracker.SetExitResult(result) - - if err := tr.exited(); err != nil { - tr.logger.Error("exited hooks failed", "error", err) - } - - RESTART: - restart, restartDelay := tr.shouldRestart() - if !restart { - break MAIN - } - - // Actually restart by sleeping and also watching for destroy events - select { - case <-time.After(restartDelay): - case <-tr.ctx.Done(): - tr.logger.Trace("task killed between restarts", "delay", restartDelay) - break MAIN - } - } - - // If task terminated, update server. All other exit conditions (eg - // killed or out of restarts) will perform their own server updates. - if result != nil { - event := structs.NewTaskEvent(structs.TaskTerminated). - SetExitCode(result.ExitCode). - SetSignal(result.Signal). - SetExitMessage(result.Err) - tr.UpdateState(structs.TaskStateDead, event) - } - - // Run the stop hooks - if err := tr.stop(); err != nil { - tr.logger.Error("stop failed", "error", err) - } - - tr.logger.Debug("task run loop exiting") -} - -// handleUpdates runs update hooks when triggerUpdateCh is ticked and exits -// when Run has returned. Should only be run in a goroutine from Run. -func (tr *TaskRunner) handleUpdates() { - for { - select { - case <-tr.triggerUpdateCh: - case <-tr.waitCh: - return - } - - if tr.Alloc().TerminalStatus() { - // Terminal update: kill TaskRunner and let Run execute postrun hooks - err := tr.Kill(context.TODO(), structs.NewTaskEvent(structs.TaskKilled)) - if err != nil { - tr.logger.Warn("error stopping task", "error", err) - } - continue - } - - // Non-terminal update; run hooks - tr.updateHooks() - } -} - -// shouldRestart determines whether the task should be restarted and updates -// the task state unless the task is killed or terminated. -func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { - // Determine if we should restart - state, when := tr.restartTracker.GetState() - reason := tr.restartTracker.GetReason() - switch state { - case structs.TaskKilled: - // Never restart an explicitly killed task. Kill method handles - // updating the server. - return false, 0 - case structs.TaskNotRestarting, structs.TaskTerminated: - tr.logger.Info("not restarting task", "reason", reason) - if state == structs.TaskNotRestarting { - tr.UpdateState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskNotRestarting).SetRestartReason(reason).SetFailsTask()) - } - return false, 0 - case structs.TaskRestarting: - tr.logger.Info("restarting task", "reason", reason, "delay", when) - tr.UpdateState(structs.TaskStatePending, structs.NewTaskEvent(structs.TaskRestarting).SetRestartDelay(when).SetRestartReason(reason)) - return true, 0 - default: - tr.logger.Error("restart tracker returned unknown state", "state", state) - return true, when - } -} - -// runDriver runs the driver and waits for it to exit -func (tr *TaskRunner) runDriver() error { - - taskConfig := drivers.NewTaskConfig(tr.task, tr.taskDir, tr.envBuilder.Build()) - taskConfig.ID = tr.buildID() - taskConfig.StdoutPath = tr.logmonHookConfig.stdoutFifo - taskConfig.StderrPath = tr.logmonHookConfig.stderrFifo - - evalCtx := &hcl.EvalContext{ - Functions: shared.GetStdlibFuncs(), - Variables: map[string]cty.Value{ - "NOMAD_ENV_bin": cty.StringVal("/bin/consul"), - }, - } - - val, diag := shared.ParseHclInterface(tr.task.Config, tr.taskSchema, evalCtx) - if diag.HasErrors() { - errStr := "failed to parse config" - for _, err := range diag.Errs() { - errStr = fmt.Sprintf("%s\n* %s", errStr, err.Error()) - } - return errors.New(errStr) - } - - if err := taskConfig.EncodeDriverConfig(val); err != nil { - tr.logger.Warn("failed to encode driver config", "error", err) - return err - } - - //TODO mounts and devices - //XXX Evaluate and encode driver config - - // Start the job - handle, net, err := tr.driver.StartTask(taskConfig) - if err != nil { - tr.logger.Warn("driver start failed", "error", err) - return err - } - tr.network = net - - tr.localStateLock.Lock() - tr.localState.TaskHandle = handle - tr.localStateLock.Unlock() - - tr.updateDriverHandle(taskConfig.ID) - // Emit an event that we started - tr.UpdateState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) - return nil -} - -func (tr *TaskRunner) updateDriverHandle(taskID string) { - tr.handleLock.Lock() - defer tr.handleLock.Unlock() - tr.handle = &driverHandleImpl{ - driver: tr.driver, - net: tr.network, - taskID: taskID, - task: tr.Task(), - } -} - -// initDriver creates the driver for the task -/*func (tr *TaskRunner) initDriver() error { - // Create a task-specific event emitter callback to expose minimal - // state to drivers - //XXX Replace with EmitEvent -- no need for a shim - eventEmitter := func(m string, args ...interface{}) { - msg := fmt.Sprintf(m, args...) - tr.logger.Debug("driver event", "event", msg) - tr.EmitEvent(structs.NewTaskEvent(structs.TaskDriverMessage).SetDriverMessage(msg)) - } - - alloc := tr.Alloc() - driverCtx := driver.NewDriverContext( - alloc.Job.Name, - alloc.TaskGroup, - tr.taskName, - tr.allocID, - tr.clientConfig, // XXX Why does it need this - tr.clientConfig.Node, // XXX THIS I NEED TO FIX - tr.logger.StandardLogger(nil), // XXX Should pass this through - eventEmitter) - - driver, err := driver.NewDriver(tr.task.Driver, driverCtx) - if err != nil { - return err - } - - tr.driver = driver - return nil -}*/ - -// initDriver retrives the DriverPlugin from the plugin loader for this task -func (tr *TaskRunner) initDriver() error { - plugin, err := tr.pluginSingletonLoader.Dispense(tr.Task().Driver, base.PluginTypeDriver, tr.logger) - if err != nil { - return err - } - - // XXX need to be able to reattach to running drivers - driver, ok := plugin.Plugin().(drivers.DriverPlugin) - if !ok { - return fmt.Errorf("plugin loaded for driver %s does not implement DriverPlugin interface", tr.task.Driver) - } - - tr.driver = driver - - schema, err := tr.driver.TaskConfigSchema() - if err != nil { - return err - } - spec, _ := hclspec.Convert(schema) - tr.taskSchema = spec - - caps, err := tr.driver.Capabilities() - if err != nil { - return err - } - tr.driverCapabilities = caps - - return nil -} - -// handleDestroy kills the task handle. In the case that killing fails, -// handleDestroy will retry with an exponential backoff and will give up at a -// given limit. It returns whether the task was destroyed and the error -// associated with the last kill attempt. -func (tr *TaskRunner) handleDestroy(handle tinterfaces.DriverHandle) (destroyed bool, err error) { - // Cap the number of times we attempt to kill the task. - for i := 0; i < killFailureLimit; i++ { - if err = handle.Kill(); err != nil { - if err == drivers.ErrTaskNotFound { - tr.logger.Warn("couldn't find task to kill", "task_id", handle.ID()) - return true, nil - } - // Calculate the new backoff - backoff := (1 << (2 * uint64(i))) * killBackoffBaseline - if backoff > killBackoffLimit { - backoff = killBackoffLimit - } - - tr.logger.Error("failed to kill task", "backoff", backoff, "error", err) - time.Sleep(backoff) - } else { - // Kill was successful - return true, nil - } - } - return -} - -// persistLocalState persists local state to disk synchronously. -func (tr *TaskRunner) persistLocalState() error { - tr.localStateLock.Lock() - defer tr.localStateLock.Unlock() - - return tr.stateDB.PutTaskRunnerLocalState(tr.allocID, tr.taskName, tr.localState) -} - -// buildID builds a consistent unique ID for the task from the alloc ID, task name and restart attempt -func (tr *TaskRunner) buildID() string { - return fmt.Sprintf("%s/%s/%d", tr.allocID, tr.taskName, tr.restartTracker.GetCount()) -} - -// XXX If the objects don't exists since the client shutdown before the task -// runner ever saved state, then we should treat it as a new task runner and not -// return an error -// -// Restore task runner state. Called by AllocRunner.Restore after NewTaskRunner -// but before Run so no locks need to be acquired. -func (tr *TaskRunner) Restore() error { - ls, ts, err := tr.stateDB.GetTaskRunnerState(tr.allocID, tr.taskName) - if err != nil { - return err - } - - tr.localState = ls - tr.state = ts - return nil -} - -// UpdateState sets the task runners allocation state and triggers a server -// update. -func (tr *TaskRunner) UpdateState(state string, event *structs.TaskEvent) { - tr.logger.Debug("setting task state", "state", state, "event", event.Type) - // Update the local state - stateCopy := tr.setStateLocal(state, event) - - // Notify the alloc runner of the transition - tr.stateUpdater.TaskStateUpdated(tr.taskName, stateCopy) -} - -// setStateLocal updates the local in-memory state, persists a copy to disk and returns a -// copy of the task's state. -func (tr *TaskRunner) setStateLocal(state string, event *structs.TaskEvent) *structs.TaskState { - tr.stateLock.Lock() - defer tr.stateLock.Unlock() - - //XXX REMOVE ME AFTER TESTING - if state == "" { - panic("UpdateState must not be called with an empty state") - } - - // Update the task state - oldState := tr.state.State - taskState := tr.state - taskState.State = state - - // Append the event - tr.appendEvent(event) - - // Handle the state transition. - switch state { - case structs.TaskStateRunning: - // Capture the start time if it is just starting - if oldState != structs.TaskStateRunning { - taskState.StartedAt = time.Now().UTC() - if !tr.clientConfig.DisableTaggedMetrics { - metrics.IncrCounterWithLabels([]string{"client", "allocs", "running"}, 1, tr.baseLabels) - } - //if r.config.BackwardsCompatibleMetrics { - //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "running"}, 1) - //} - } - case structs.TaskStateDead: - // Capture the finished time if not already set - if taskState.FinishedAt.IsZero() { - taskState.FinishedAt = time.Now().UTC() - } - - // Emitting metrics to indicate task complete and failures - if taskState.Failed { - if !tr.clientConfig.DisableTaggedMetrics { - metrics.IncrCounterWithLabels([]string{"client", "allocs", "failed"}, 1, tr.baseLabels) - } - //if r.config.BackwardsCompatibleMetrics { - //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "failed"}, 1) - //} - } else { - if !tr.clientConfig.DisableTaggedMetrics { - metrics.IncrCounterWithLabels([]string{"client", "allocs", "complete"}, 1, tr.baseLabels) - } - //if r.config.BackwardsCompatibleMetrics { - //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "complete"}, 1) - //} - } - } - - // Persist the state and event - if err := tr.stateDB.PutTaskState(tr.allocID, tr.taskName, taskState); err != nil { - // Only a warning because the next event/state-transition will - // try to persist it again. - tr.logger.Error("error persisting task state", "error", err, "event", event, "state", state) - } - - return tr.state.Copy() -} - -// EmitEvent appends a new TaskEvent to this task's TaskState. The actual -// TaskState.State (pending, running, dead) is not changed. Use UpdateState to -// transition states. -// Events are persisted locally and sent to the server, but errors are simply -// logged. Use AppendEvent to simply add a new event. -func (tr *TaskRunner) EmitEvent(event *structs.TaskEvent) { - tr.stateLock.Lock() - defer tr.stateLock.Unlock() - - tr.appendEvent(event) - - if err := tr.stateDB.PutTaskState(tr.allocID, tr.taskName, tr.state); err != nil { - // Only a warning because the next event/state-transition will - // try to persist it again. - tr.logger.Warn("error persisting event", "error", err, "event", event) - } - - // Notify the alloc runner of the event - tr.stateUpdater.TaskStateUpdated(tr.taskName, tr.state.Copy()) -} - -// AppendEvent appends a new TaskEvent to this task's TaskState. The actual -// TaskState.State (pending, running, dead) is not changed. Use UpdateState to -// transition states. -// Events are persisted locally and errors are simply logged. Use EmitEvent -// also update AllocRunner. -func (tr *TaskRunner) AppendEvent(event *structs.TaskEvent) { - tr.stateLock.Lock() - defer tr.stateLock.Unlock() - - tr.appendEvent(event) - - if err := tr.stateDB.PutTaskState(tr.allocID, tr.taskName, tr.state); err != nil { - // Only a warning because the next event/state-transition will - // try to persist it again. - tr.logger.Warn("error persisting event", "error", err, "event", event) - } -} - -// appendEvent to task's event slice. Caller must acquire stateLock. -func (tr *TaskRunner) appendEvent(event *structs.TaskEvent) error { - // Ensure the event is populated with human readable strings - event.PopulateEventDisplayMessage() - - // Propogate failure from event to task state - if event.FailsTask { - tr.state.Failed = true - } - - // XXX This seems like a super awkward spot for this? Why not shouldRestart? - // Update restart metrics - if event.Type == structs.TaskRestarting { - if !tr.clientConfig.DisableTaggedMetrics { - metrics.IncrCounterWithLabels([]string{"client", "allocs", "restart"}, 1, tr.baseLabels) - } - //if r.config.BackwardsCompatibleMetrics { - //metrics.IncrCounter([]string{"client", "allocs", r.alloc.Job.Name, r.alloc.TaskGroup, taskName, "restart"}, 1) - //} - tr.state.Restarts++ - tr.state.LastRestart = time.Unix(0, event.Time) - } - - // Append event to slice - appendTaskEvent(tr.state, event) - - return nil -} - -// WaitCh is closed when TaskRunner.Run exits. -func (tr *TaskRunner) WaitCh() <-chan struct{} { - return tr.waitCh -} - -// Update the running allocation with a new version received from the server. -// Calls Update hooks asynchronously with Run(). -// -// This method is safe for calling concurrently with Run() and does not modify -// the passed in allocation. -func (tr *TaskRunner) Update(update *structs.Allocation) { - // Update tr.alloc - tr.setAlloc(update) - - // Trigger update hooks - tr.triggerUpdateHooks() -} - -// triggerUpdate if there isn't already an update pending. Should be called -// instead of calling updateHooks directly to serialize runs of update hooks. -// TaskRunner state should be updated prior to triggering update hooks. -// -// Does not block. -func (tr *TaskRunner) triggerUpdateHooks() { - select { - case tr.triggerUpdateCh <- struct{}{}: - default: - // already an update hook pending - } -} - -// LatestResourceUsage returns the last resource utilization datapoint -// collected. May return nil if the task is not running or no resource -// utilization has been collected yet. -func (tr *TaskRunner) LatestResourceUsage() *cstructs.TaskResourceUsage { - tr.resourceUsageLock.Lock() - ru := tr.resourceUsage - tr.resourceUsageLock.Unlock() - return ru -} - -// UpdateStats updates and emits the latest stats from the driver. -func (tr *TaskRunner) UpdateStats(ru *cstructs.TaskResourceUsage) { - tr.resourceUsageLock.Lock() - tr.resourceUsage = ru - tr.resourceUsageLock.Unlock() - if ru != nil { - tr.emitStats(ru) - } -} - -//TODO Remove Backwardscompat or use tr.Alloc()? -func (tr *TaskRunner) setGaugeForMemory(ru *cstructs.TaskResourceUsage) { - if !tr.clientConfig.DisableTaggedMetrics { - metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "rss"}, - float32(ru.ResourceUsage.MemoryStats.RSS), tr.baseLabels) - metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "rss"}, - float32(ru.ResourceUsage.MemoryStats.RSS), tr.baseLabels) - metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "cache"}, - float32(ru.ResourceUsage.MemoryStats.Cache), tr.baseLabels) - metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "swap"}, - float32(ru.ResourceUsage.MemoryStats.Swap), tr.baseLabels) - metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "max_usage"}, - float32(ru.ResourceUsage.MemoryStats.MaxUsage), tr.baseLabels) - metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "kernel_usage"}, - float32(ru.ResourceUsage.MemoryStats.KernelUsage), tr.baseLabels) - metrics.SetGaugeWithLabels([]string{"client", "allocs", "memory", "kernel_max_usage"}, - float32(ru.ResourceUsage.MemoryStats.KernelMaxUsage), tr.baseLabels) - } - - if tr.clientConfig.BackwardsCompatibleMetrics { - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "rss"}, float32(ru.ResourceUsage.MemoryStats.RSS)) - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "cache"}, float32(ru.ResourceUsage.MemoryStats.Cache)) - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "swap"}, float32(ru.ResourceUsage.MemoryStats.Swap)) - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "max_usage"}, float32(ru.ResourceUsage.MemoryStats.MaxUsage)) - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "kernel_usage"}, float32(ru.ResourceUsage.MemoryStats.KernelUsage)) - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "memory", "kernel_max_usage"}, float32(ru.ResourceUsage.MemoryStats.KernelMaxUsage)) - } -} - -//TODO Remove Backwardscompat or use tr.Alloc()? -func (tr *TaskRunner) setGaugeForCPU(ru *cstructs.TaskResourceUsage) { - if !tr.clientConfig.DisableTaggedMetrics { - metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "total_percent"}, - float32(ru.ResourceUsage.CpuStats.Percent), tr.baseLabels) - metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "system"}, - float32(ru.ResourceUsage.CpuStats.SystemMode), tr.baseLabels) - metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "user"}, - float32(ru.ResourceUsage.CpuStats.UserMode), tr.baseLabels) - metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "throttled_time"}, - float32(ru.ResourceUsage.CpuStats.ThrottledTime), tr.baseLabels) - metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "throttled_periods"}, - float32(ru.ResourceUsage.CpuStats.ThrottledPeriods), tr.baseLabels) - metrics.SetGaugeWithLabels([]string{"client", "allocs", "cpu", "total_ticks"}, - float32(ru.ResourceUsage.CpuStats.TotalTicks), tr.baseLabels) - } - - if tr.clientConfig.BackwardsCompatibleMetrics { - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "total_percent"}, float32(ru.ResourceUsage.CpuStats.Percent)) - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "system"}, float32(ru.ResourceUsage.CpuStats.SystemMode)) - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "user"}, float32(ru.ResourceUsage.CpuStats.UserMode)) - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "throttled_time"}, float32(ru.ResourceUsage.CpuStats.ThrottledTime)) - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "throttled_periods"}, float32(ru.ResourceUsage.CpuStats.ThrottledPeriods)) - metrics.SetGauge([]string{"client", "allocs", tr.alloc.Job.Name, tr.alloc.TaskGroup, tr.allocID, tr.taskName, "cpu", "total_ticks"}, float32(ru.ResourceUsage.CpuStats.TotalTicks)) - } -} - -// emitStats emits resource usage stats of tasks to remote metrics collector -// sinks -func (tr *TaskRunner) emitStats(ru *cstructs.TaskResourceUsage) { - if !tr.clientConfig.PublishAllocationMetrics { - return - } - - if ru.ResourceUsage.MemoryStats != nil { - tr.setGaugeForMemory(ru) - } - - if ru.ResourceUsage.CpuStats != nil { - tr.setGaugeForCPU(ru) - } -} - -// appendTaskEvent updates the task status by appending the new event. -func appendTaskEvent(state *structs.TaskState, event *structs.TaskEvent) { - const capacity = 10 - if state.Events == nil { - state.Events = make([]*structs.TaskEvent, 1, capacity) - state.Events[0] = event - return - } - - // If we hit capacity, then shift it. - if len(state.Events) == capacity { - old := state.Events - state.Events = make([]*structs.TaskEvent, 0, capacity) - state.Events = append(state.Events, old[1:]...) - } - - state.Events = append(state.Events, event) -} diff --git a/client/client.go b/client/client.go index 491045dab..fd9935de9 100644 --- a/client/client.go +++ b/client/client.go @@ -761,7 +761,7 @@ func (c *Client) restoreState() error { for _, alloc := range allocs { c.configLock.RLock() - arConf := &allocrunnerv2.Config{ + arConf := &allocrunner.Config{ Alloc: alloc, Logger: c.logger, ClientConfig: c.config, @@ -1911,7 +1911,7 @@ func (c *Client) addAlloc(alloc *structs.Allocation, migrateToken string) error // The long term fix is to pass in the config and node separately and then // we don't have to do a copy. c.configLock.RLock() - arConf := &allocrunnerv2.Config{ + arConf := &allocrunner.Config{ Alloc: alloc, Logger: c.logger, ClientConfig: c.config, diff --git a/command/agent/consul/script.go b/command/agent/consul/script.go index 0caa3ef1e..4a24d8f99 100644 --- a/command/agent/consul/script.go +++ b/command/agent/consul/script.go @@ -8,7 +8,7 @@ import ( log "github.com/hashicorp/go-hclog" "github.com/hashicorp/consul/api" - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/nomad/structs" ) diff --git a/command/agent/consul/structs.go b/command/agent/consul/structs.go index 729134523..a5d070d8f 100644 --- a/command/agent/consul/structs.go +++ b/command/agent/consul/structs.go @@ -1,7 +1,7 @@ package consul import ( - "github.com/hashicorp/nomad/client/allocrunnerv2/taskrunner/interfaces" + "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" ) From 64e9fc379834e67b9d11cb567a5c46bc17efd22a Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Fri, 5 Oct 2018 23:32:43 -0400 Subject: [PATCH 121/152] docklog: add go-plugin for forwarding of docker logs --- drivers/docker/docklog/client.go | 35 +++ drivers/docker/docklog/docklog.go | 141 +++++++++ drivers/docker/docklog/docklog_test.go | 102 +++++++ drivers/docker/docklog/plugin.go | 68 +++++ drivers/docker/docklog/proto/docklog.pb.go | 338 +++++++++++++++++++++ drivers/docker/docklog/proto/docklog.proto | 25 ++ drivers/docker/docklog/server.go | 39 +++ 7 files changed, 748 insertions(+) create mode 100644 drivers/docker/docklog/client.go create mode 100644 drivers/docker/docklog/docklog.go create mode 100644 drivers/docker/docklog/docklog_test.go create mode 100644 drivers/docker/docklog/plugin.go create mode 100644 drivers/docker/docklog/proto/docklog.pb.go create mode 100644 drivers/docker/docklog/proto/docklog.proto create mode 100644 drivers/docker/docklog/server.go diff --git a/drivers/docker/docklog/client.go b/drivers/docker/docklog/client.go new file mode 100644 index 000000000..d07cf3339 --- /dev/null +++ b/drivers/docker/docklog/client.go @@ -0,0 +1,35 @@ +package docklog + +import ( + "context" + + "github.com/hashicorp/nomad/drivers/docker/docklog/proto" +) + +// docklogClient implements the Docklog interface for client side requests +type docklogClient struct { + client proto.DocklogClient +} + +// Start proxies the Start client side func to the protobuf interface +func (c *docklogClient) Start(opts *StartOpts) error { + req := &proto.StartRequest{ + Endpoint: opts.Endpoint, + ContainerId: opts.ContainerID, + StdoutFifo: opts.Stdout, + StderrFifo: opts.Stderr, + + TlsCert: opts.TLSCert, + TlsKey: opts.TLSKey, + TlsCa: opts.TLSCA, + } + _, err := c.client.Start(context.Background(), req) + return err +} + +// Stop proxies the Stop client side func to the protobuf interface +func (c *docklogClient) Stop() error { + req := &proto.StopRequest{} + _, err := c.client.Stop(context.Background(), req) + return err +} diff --git a/drivers/docker/docklog/docklog.go b/drivers/docker/docklog/docklog.go new file mode 100644 index 000000000..5557dcb8f --- /dev/null +++ b/drivers/docker/docklog/docklog.go @@ -0,0 +1,141 @@ +package docklog + +import ( + "fmt" + "io" + + docker "github.com/fsouza/go-dockerclient" + hclog "github.com/hashicorp/go-hclog" + multierror "github.com/hashicorp/go-multierror" + "github.com/hashicorp/nomad/client/lib/fifo" + "golang.org/x/net/context" +) + +// Docklog is a small utility to forward logs from a docker container to a target +// destination +type Docklog interface { + Start(*StartOpts) error + Stop() error +} + +// StartOpts are the options needed to start docker log monitoring +type StartOpts struct { + // Endpoint sets the docker client endpoint, defaults to environment if not set + Endpoint string + + // ContainerID of the container to monitor logs for + ContainerID string + + // Stdout path to fifo + Stdout string + //Stderr path to fifo + Stderr string + + // TLS settings for docker client + TLSCert string + TLSKey string + TLSCA string +} + +// NewDocklog returns an implementation of the Docklog interface +func NewDocklog(logger hclog.Logger) Docklog { + return &dockerLogger{logger: logger} +} + +// dockerLogger implements the Docklog interface +type dockerLogger struct { + logger hclog.Logger + + stdout io.WriteCloser + stderr io.WriteCloser + cancelCtx context.CancelFunc +} + +// Start log monitoring +func (d *dockerLogger) Start(opts *StartOpts) error { + client, err := d.getDockerClient(opts) + if err != nil { + return fmt.Errorf("failed to open docker client: %v", err) + } + + if d.stdout == nil { + stdout, err := fifo.Open(opts.Stdout) + if err != nil { + return fmt.Errorf("failed to open fifo for path %s: %v", opts.Stdout, err) + } + d.stdout = stdout + } + if d.stderr == nil { + stderr, err := fifo.Open(opts.Stderr) + if err != nil { + return fmt.Errorf("failed to open fifo for path %s: %v", opts.Stdout, err) + } + d.stderr = stderr + } + ctx, cancel := context.WithCancel(context.Background()) + d.cancelCtx = cancel + + logOpts := docker.LogsOptions{ + Context: ctx, + Container: opts.ContainerID, + OutputStream: d.stdout, + ErrorStream: d.stderr, + Since: 0, + Follow: true, + Stdout: true, + Stderr: true, + } + + go func() { client.Logs(logOpts) }() + return nil + +} + +// Stop log monitoring +func (d *dockerLogger) Stop() error { + if d.cancelCtx != nil { + d.cancelCtx() + } + if d.stdout != nil { + d.stdout.Close() + } + if d.stderr != nil { + d.stderr.Close() + } + return nil +} + +func (d *dockerLogger) getDockerClient(opts *StartOpts) (*docker.Client, error) { + var err error + var merr multierror.Error + var newClient *docker.Client + + // Default to using whatever is configured in docker.endpoint. If this is + // not specified we'll fall back on NewClientFromEnv which reads config from + // the DOCKER_* environment variables DOCKER_HOST, DOCKER_TLS_VERIFY, and + // DOCKER_CERT_PATH. This allows us to lock down the config in production + // but also accept the standard ENV configs for dev and test. + if opts.Endpoint != "" { + if opts.TLSCert+opts.TLSKey+opts.TLSCA != "" { + d.logger.Debug("using TLS client connection to docker", "endpoint", opts.Endpoint) + newClient, err = docker.NewTLSClient(opts.Endpoint, opts.TLSCert, opts.TLSKey, opts.TLSCA) + if err != nil { + merr.Errors = append(merr.Errors, err) + } + } else { + d.logger.Debug("using plaintext client connection to docker", "endpoint", opts.Endpoint) + newClient, err = docker.NewClient(opts.Endpoint) + if err != nil { + merr.Errors = append(merr.Errors, err) + } + } + } else { + d.logger.Debug("using client connection initialized from environment") + newClient, err = docker.NewClientFromEnv() + if err != nil { + merr.Errors = append(merr.Errors, err) + } + } + + return newClient, merr.ErrorOrNil() +} diff --git a/drivers/docker/docklog/docklog_test.go b/drivers/docker/docklog/docklog_test.go new file mode 100644 index 000000000..7c4848537 --- /dev/null +++ b/drivers/docker/docklog/docklog_test.go @@ -0,0 +1,102 @@ +package docklog + +import ( + "bytes" + "fmt" + "testing" + "time" + + docker "github.com/fsouza/go-dockerclient" + "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/testutil" + "github.com/stretchr/testify/require" + "golang.org/x/net/context" +) + +func TestDocklog(t *testing.T) { + t.Parallel() + require := require.New(t) + + client, err := docker.NewClientFromEnv() + if err != nil { + t.Skip("docker unavailable:", err) + } + + containerConf := docker.CreateContainerOptions{ + Config: &docker.Config{ + Cmd: []string{ + "/bin/ash", "-c", "touch /tmp/docklog; tail -f /tmp/docklog", + }, + Image: "alpine", + }, + Context: context.Background(), + } + + container, err := client.CreateContainer(containerConf) + require.NoError(err) + + defer client.RemoveContainer(docker.RemoveContainerOptions{ + ID: container.ID, + Force: true, + }) + + err = client.StartContainer(container.ID, nil) + require.NoError(err) + + var count int + for !container.State.Running { + if count > 10 { + t.Fatal("timeout waiting for container to start") + } + time.Sleep(100 * time.Millisecond) + container, err = client.InspectContainer(container.ID) + count++ + } + + stdout := &noopCloser{bytes.NewBufferString("")} + stderr := &noopCloser{bytes.NewBufferString("")} + + dl := NewDocklog(testlog.HCLogger(t)).(*dockerLogger) + dl.stdout = stdout + dl.stderr = stderr + require.NoError(dl.Start(&StartOpts{ + ContainerID: container.ID, + })) + + echoToContainer(t, client, container.ID, "abc") + echoToContainer(t, client, container.ID, "123") + + time.Sleep(2000 * time.Millisecond) + testutil.WaitForResult(func() (bool, error) { + act := stdout.String() + if "abc\n123\n" != act { + return false, fmt.Errorf("expected abc\\n123\\n for stdout but got %s", act) + } + + return true, nil + }, func(err error) { + require.NoError(err) + }) +} + +func echoToContainer(t *testing.T, client *docker.Client, id string, line string) { + op := docker.CreateExecOptions{ + Container: id, + Cmd: []string{ + "/bin/ash", "-c", + fmt.Sprintf("echo %s >>/tmp/docklog", line), + }, + } + + exec, err := client.CreateExec(op) + require.NoError(t, err) + require.NoError(t, client.StartExec(exec.ID, docker.StartExecOptions{Detach: true})) +} + +type noopCloser struct { + *bytes.Buffer +} + +func (*noopCloser) Close() error { + return nil +} diff --git a/drivers/docker/docklog/plugin.go b/drivers/docker/docklog/plugin.go new file mode 100644 index 000000000..97498525a --- /dev/null +++ b/drivers/docker/docklog/plugin.go @@ -0,0 +1,68 @@ +package docklog + +import ( + "context" + "os/exec" + + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/drivers/docker/docklog/proto" + "github.com/hashicorp/nomad/helper/discover" + "github.com/hashicorp/nomad/plugins/base" + "google.golang.org/grpc" +) + +// LaunchDocklog launches an instance of docklog +// TODO: Integrate with base plugin loader +func LaunchDocklog(logger hclog.Logger) (Docklog, *plugin.Client, error) { + logger = logger.Named("docklog-launcher") + bin, err := discover.NomadExecutable() + if err != nil { + return nil, nil, err + } + + client := plugin.NewClient(&plugin.ClientConfig{ + HandshakeConfig: base.Handshake, + Plugins: map[string]plugin.Plugin{ + "docklog": &Plugin{impl: NewDocklog(hclog.L().Named("docklog"))}, + }, + Cmd: exec.Command(bin, "docklog"), + AllowedProtocols: []plugin.Protocol{ + plugin.ProtocolGRPC, + }, + }) + + rpcClient, err := client.Client() + if err != nil { + return nil, nil, err + } + + raw, err := rpcClient.Dispense("docklog") + if err != nil { + return nil, nil, err + } + + l := raw.(Docklog) + return l, client, nil + +} + +// Plugin is the go-plugin implementation +type Plugin struct { + plugin.NetRPCUnsupportedPlugin + impl Docklog +} + +// GRPCServer registered the server side implementation with the grpc server +func (p *Plugin) GRPCServer(broker *plugin.GRPCBroker, s *grpc.Server) error { + proto.RegisterDocklogServer(s, &docklogServer{ + impl: p.impl, + broker: broker, + }) + return nil +} + +// GRPCClient returns a client side implementation of the plugin +func (p *Plugin) GRPCClient(ctx context.Context, broker *plugin.GRPCBroker, c *grpc.ClientConn) (interface{}, error) { + return &docklogClient{client: proto.NewDocklogClient(c)}, nil +} diff --git a/drivers/docker/docklog/proto/docklog.pb.go b/drivers/docker/docklog/proto/docklog.pb.go new file mode 100644 index 000000000..506455821 --- /dev/null +++ b/drivers/docker/docklog/proto/docklog.pb.go @@ -0,0 +1,338 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: docklog.proto + +package proto + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" + +import ( + context "golang.org/x/net/context" + grpc "google.golang.org/grpc" +) + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +type StartRequest struct { + Endpoint string `protobuf:"bytes,1,opt,name=endpoint,proto3" json:"endpoint,omitempty"` + ContainerId string `protobuf:"bytes,2,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"` + StdoutFifo string `protobuf:"bytes,3,opt,name=stdout_fifo,json=stdoutFifo,proto3" json:"stdout_fifo,omitempty"` + StderrFifo string `protobuf:"bytes,4,opt,name=stderr_fifo,json=stderrFifo,proto3" json:"stderr_fifo,omitempty"` + TlsCert string `protobuf:"bytes,5,opt,name=tls_cert,json=tlsCert,proto3" json:"tls_cert,omitempty"` + TlsKey string `protobuf:"bytes,6,opt,name=tls_key,json=tlsKey,proto3" json:"tls_key,omitempty"` + TlsCa string `protobuf:"bytes,7,opt,name=tls_ca,json=tlsCa,proto3" json:"tls_ca,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *StartRequest) Reset() { *m = StartRequest{} } +func (m *StartRequest) String() string { return proto.CompactTextString(m) } +func (*StartRequest) ProtoMessage() {} +func (*StartRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_docklog_6cb3edde8d295224, []int{0} +} +func (m *StartRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StartRequest.Unmarshal(m, b) +} +func (m *StartRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StartRequest.Marshal(b, m, deterministic) +} +func (dst *StartRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_StartRequest.Merge(dst, src) +} +func (m *StartRequest) XXX_Size() int { + return xxx_messageInfo_StartRequest.Size(m) +} +func (m *StartRequest) XXX_DiscardUnknown() { + xxx_messageInfo_StartRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_StartRequest proto.InternalMessageInfo + +func (m *StartRequest) GetEndpoint() string { + if m != nil { + return m.Endpoint + } + return "" +} + +func (m *StartRequest) GetContainerId() string { + if m != nil { + return m.ContainerId + } + return "" +} + +func (m *StartRequest) GetStdoutFifo() string { + if m != nil { + return m.StdoutFifo + } + return "" +} + +func (m *StartRequest) GetStderrFifo() string { + if m != nil { + return m.StderrFifo + } + return "" +} + +func (m *StartRequest) GetTlsCert() string { + if m != nil { + return m.TlsCert + } + return "" +} + +func (m *StartRequest) GetTlsKey() string { + if m != nil { + return m.TlsKey + } + return "" +} + +func (m *StartRequest) GetTlsCa() string { + if m != nil { + return m.TlsCa + } + return "" +} + +type StartResponse struct { + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *StartResponse) Reset() { *m = StartResponse{} } +func (m *StartResponse) String() string { return proto.CompactTextString(m) } +func (*StartResponse) ProtoMessage() {} +func (*StartResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_docklog_6cb3edde8d295224, []int{1} +} +func (m *StartResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StartResponse.Unmarshal(m, b) +} +func (m *StartResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StartResponse.Marshal(b, m, deterministic) +} +func (dst *StartResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_StartResponse.Merge(dst, src) +} +func (m *StartResponse) XXX_Size() int { + return xxx_messageInfo_StartResponse.Size(m) +} +func (m *StartResponse) XXX_DiscardUnknown() { + xxx_messageInfo_StartResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_StartResponse proto.InternalMessageInfo + +type StopRequest struct { + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *StopRequest) Reset() { *m = StopRequest{} } +func (m *StopRequest) String() string { return proto.CompactTextString(m) } +func (*StopRequest) ProtoMessage() {} +func (*StopRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_docklog_6cb3edde8d295224, []int{2} +} +func (m *StopRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StopRequest.Unmarshal(m, b) +} +func (m *StopRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StopRequest.Marshal(b, m, deterministic) +} +func (dst *StopRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_StopRequest.Merge(dst, src) +} +func (m *StopRequest) XXX_Size() int { + return xxx_messageInfo_StopRequest.Size(m) +} +func (m *StopRequest) XXX_DiscardUnknown() { + xxx_messageInfo_StopRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_StopRequest proto.InternalMessageInfo + +type StopResponse struct { + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *StopResponse) Reset() { *m = StopResponse{} } +func (m *StopResponse) String() string { return proto.CompactTextString(m) } +func (*StopResponse) ProtoMessage() {} +func (*StopResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_docklog_6cb3edde8d295224, []int{3} +} +func (m *StopResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StopResponse.Unmarshal(m, b) +} +func (m *StopResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StopResponse.Marshal(b, m, deterministic) +} +func (dst *StopResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_StopResponse.Merge(dst, src) +} +func (m *StopResponse) XXX_Size() int { + return xxx_messageInfo_StopResponse.Size(m) +} +func (m *StopResponse) XXX_DiscardUnknown() { + xxx_messageInfo_StopResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_StopResponse proto.InternalMessageInfo + +func init() { + proto.RegisterType((*StartRequest)(nil), "hashicorp.nomad.drivers.docker.docklog.proto.StartRequest") + proto.RegisterType((*StartResponse)(nil), "hashicorp.nomad.drivers.docker.docklog.proto.StartResponse") + proto.RegisterType((*StopRequest)(nil), "hashicorp.nomad.drivers.docker.docklog.proto.StopRequest") + proto.RegisterType((*StopResponse)(nil), "hashicorp.nomad.drivers.docker.docklog.proto.StopResponse") +} + +// Reference imports to suppress errors if they are not otherwise used. +var _ context.Context +var _ grpc.ClientConn + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the grpc package it is being compiled against. +const _ = grpc.SupportPackageIsVersion4 + +// DocklogClient is the client API for Docklog service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. +type DocklogClient interface { + Start(ctx context.Context, in *StartRequest, opts ...grpc.CallOption) (*StartResponse, error) + Stop(ctx context.Context, in *StopRequest, opts ...grpc.CallOption) (*StopResponse, error) +} + +type docklogClient struct { + cc *grpc.ClientConn +} + +func NewDocklogClient(cc *grpc.ClientConn) DocklogClient { + return &docklogClient{cc} +} + +func (c *docklogClient) Start(ctx context.Context, in *StartRequest, opts ...grpc.CallOption) (*StartResponse, error) { + out := new(StartResponse) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.drivers.docker.docklog.proto.Docklog/Start", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *docklogClient) Stop(ctx context.Context, in *StopRequest, opts ...grpc.CallOption) (*StopResponse, error) { + out := new(StopResponse) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.drivers.docker.docklog.proto.Docklog/Stop", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +// DocklogServer is the server API for Docklog service. +type DocklogServer interface { + Start(context.Context, *StartRequest) (*StartResponse, error) + Stop(context.Context, *StopRequest) (*StopResponse, error) +} + +func RegisterDocklogServer(s *grpc.Server, srv DocklogServer) { + s.RegisterService(&_Docklog_serviceDesc, srv) +} + +func _Docklog_Start_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(StartRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(DocklogServer).Start(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/hashicorp.nomad.drivers.docker.docklog.proto.Docklog/Start", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(DocklogServer).Start(ctx, req.(*StartRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _Docklog_Stop_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(StopRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(DocklogServer).Stop(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/hashicorp.nomad.drivers.docker.docklog.proto.Docklog/Stop", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(DocklogServer).Stop(ctx, req.(*StopRequest)) + } + return interceptor(ctx, in, info, handler) +} + +var _Docklog_serviceDesc = grpc.ServiceDesc{ + ServiceName: "hashicorp.nomad.drivers.docker.docklog.proto.Docklog", + HandlerType: (*DocklogServer)(nil), + Methods: []grpc.MethodDesc{ + { + MethodName: "Start", + Handler: _Docklog_Start_Handler, + }, + { + MethodName: "Stop", + Handler: _Docklog_Stop_Handler, + }, + }, + Streams: []grpc.StreamDesc{}, + Metadata: "docklog.proto", +} + +func init() { proto.RegisterFile("docklog.proto", fileDescriptor_docklog_6cb3edde8d295224) } + +var fileDescriptor_docklog_6cb3edde8d295224 = []byte{ + // 311 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x90, 0xb1, 0x4e, 0xf3, 0x40, + 0x10, 0x84, 0x7f, 0xe7, 0x8f, 0xed, 0xb0, 0x49, 0x40, 0x3a, 0x09, 0x61, 0xd2, 0x00, 0xae, 0x28, + 0xd0, 0x15, 0x50, 0x11, 0x3a, 0x40, 0x48, 0x88, 0x2e, 0xe9, 0x68, 0x2c, 0xe3, 0x3b, 0x93, 0x53, + 0x8c, 0xd7, 0xec, 0x6d, 0x90, 0x52, 0x21, 0xf1, 0x02, 0xbc, 0x1f, 0x4f, 0x83, 0x7c, 0x36, 0x16, + 0x94, 0x49, 0x65, 0xef, 0xcc, 0x37, 0xa7, 0x9d, 0x85, 0xb1, 0xc2, 0x6c, 0x59, 0xe0, 0xb3, 0xac, + 0x08, 0x19, 0xc5, 0xd9, 0x22, 0xb5, 0x0b, 0x93, 0x21, 0x55, 0xb2, 0xc4, 0x97, 0x54, 0x49, 0x45, + 0xe6, 0x4d, 0x93, 0x95, 0x35, 0xa6, 0x49, 0xfe, 0xa1, 0xe3, 0x2f, 0x0f, 0x46, 0x73, 0x4e, 0x89, + 0x67, 0xfa, 0x75, 0xa5, 0x2d, 0x8b, 0x09, 0x0c, 0x74, 0xa9, 0x2a, 0x34, 0x25, 0x47, 0xde, 0xb1, + 0x77, 0xba, 0x33, 0xeb, 0x66, 0x71, 0x02, 0xa3, 0x0c, 0x4b, 0x4e, 0x4d, 0xa9, 0x29, 0x31, 0x2a, + 0xea, 0x39, 0x7f, 0xd8, 0x69, 0xf7, 0x4a, 0x1c, 0xc1, 0xd0, 0xb2, 0xc2, 0x15, 0x27, 0xb9, 0xc9, + 0x31, 0xfa, 0xef, 0x08, 0x68, 0xa4, 0x3b, 0x93, 0x63, 0x0b, 0x68, 0xa2, 0x06, 0xe8, 0x77, 0x80, + 0x26, 0x72, 0xc0, 0x21, 0x0c, 0xb8, 0xb0, 0x49, 0xa6, 0x89, 0x23, 0xdf, 0xb9, 0x21, 0x17, 0xf6, + 0x46, 0x13, 0x8b, 0x03, 0xa8, 0x7f, 0x93, 0xa5, 0x5e, 0x47, 0x81, 0x73, 0x02, 0x2e, 0xec, 0x83, + 0x5e, 0x8b, 0x7d, 0x08, 0x5c, 0x26, 0x8d, 0x42, 0xa7, 0xfb, 0x75, 0x22, 0x8d, 0xf7, 0x60, 0xdc, + 0x76, 0xb3, 0x15, 0x96, 0x56, 0xc7, 0x63, 0x18, 0xce, 0x19, 0xab, 0xb6, 0x6b, 0xbc, 0x5b, 0x77, + 0xaf, 0xc7, 0xc6, 0x3e, 0xff, 0xec, 0x41, 0x78, 0xdb, 0x9c, 0x47, 0x7c, 0x78, 0xe0, 0xbb, 0xb0, + 0x98, 0xca, 0x4d, 0x2e, 0x2a, 0x7f, 0x5f, 0x73, 0x72, 0xb5, 0x55, 0xb6, 0xdd, 0xf6, 0x9f, 0x78, + 0x87, 0x7e, 0xbd, 0xa0, 0xb8, 0xdc, 0xf4, 0x99, 0xae, 0xe3, 0x64, 0xba, 0x4d, 0xf4, 0x67, 0x81, + 0xeb, 0xf0, 0xd1, 0x77, 0xfa, 0x53, 0xe0, 0x3e, 0x17, 0xdf, 0x01, 0x00, 0x00, 0xff, 0xff, 0xd7, + 0xfa, 0x4e, 0xd3, 0x6d, 0x02, 0x00, 0x00, +} diff --git a/drivers/docker/docklog/proto/docklog.proto b/drivers/docker/docklog/proto/docklog.proto new file mode 100644 index 000000000..ef77ea1c2 --- /dev/null +++ b/drivers/docker/docklog/proto/docklog.proto @@ -0,0 +1,25 @@ +syntax = "proto3"; +package hashicorp.nomad.drivers.docker.docklog.proto; +option go_package = "proto"; + +service Docklog { + rpc Start(StartRequest) returns (StartResponse) {} + rpc Stop(StopRequest) returns (StopResponse) {} +} + +message StartRequest { + string endpoint = 1; + string container_id = 2; + string stdout_fifo = 3; + string stderr_fifo = 4; + string tls_cert = 5; + string tls_key = 6; + string tls_ca = 7; +} + +message StartResponse { +} + +message StopRequest {} + +message StopResponse {} diff --git a/drivers/docker/docklog/server.go b/drivers/docker/docklog/server.go new file mode 100644 index 000000000..2602e2163 --- /dev/null +++ b/drivers/docker/docklog/server.go @@ -0,0 +1,39 @@ +package docklog + +import ( + "golang.org/x/net/context" + + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/drivers/docker/docklog/proto" +) + +// docklogServer is the server side translation between the protobuf and native interfaces +type docklogServer struct { + broker *plugin.GRPCBroker + impl Docklog +} + +// Start proxies the protobuf Start RPC to the Start fun of the Docklog interface +func (s *docklogServer) Start(ctx context.Context, req *proto.StartRequest) (*proto.StartResponse, error) { + opts := &StartOpts{ + Endpoint: req.Endpoint, + ContainerID: req.ContainerId, + Stdout: req.StdoutFifo, + Stderr: req.StderrFifo, + + TLSCert: req.TlsCert, + TLSKey: req.TlsKey, + TLSCA: req.TlsCa, + } + err := s.impl.Start(opts) + if err != nil { + return nil, err + } + resp := &proto.StartResponse{} + return resp, nil +} + +// Stop proxies the protobuf Stop RPC to the Stop fun of the Docklog interface +func (s *docklogServer) Stop(ctx context.Context, req *proto.StopRequest) (*proto.StopResponse, error) { + return &proto.StopResponse{}, s.impl.Stop() +} From 3183b33d247f7bc573dade2e2ef64c89ffc2572d Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Tue, 9 Oct 2018 23:01:20 -0400 Subject: [PATCH 122/152] client: review comments and fixup/skip tests --- client/alloc_endpoint_test.go | 2 + client/alloc_watcher_e2e_test.go | 1 + client/allocrunner/alloc_runner.go | 5 - client/allocrunner/alloc_runner_test.go | 23 +++-- .../allocrunner/taskrunner/driver_handle.go | 26 ++--- client/allocrunner/taskrunner/lifecycle.go | 20 ++-- client/allocrunner/taskrunner/task_runner.go | 26 +---- .../taskrunner/task_runner_getters.go | 2 +- client/client.go | 98 +++++++++---------- client/client_test.go | 3 + client/fingerprint_manager.go | 88 +++++++++++++---- client/fingerprint_manager_test.go | 32 +++++- client/fs_endpoint_test.go | 8 ++ client/testing.go | 9 +- command/agent/plugin_internal_catalog.go | 12 +++ command/agent/plugins.go | 2 - drivers/rawexec/driver.go | 12 +-- 17 files changed, 228 insertions(+), 141 deletions(-) create mode 100644 command/agent/plugin_internal_catalog.go diff --git a/client/alloc_endpoint_test.go b/client/alloc_endpoint_test.go index c111f209b..b616a7c9a 100644 --- a/client/alloc_endpoint_test.go +++ b/client/alloc_endpoint_test.go @@ -76,6 +76,7 @@ func TestAllocations_GarbageCollectAll_ACL(t *testing.T) { } func TestAllocations_GarbageCollect(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) client := TestClient(t, func(c *config.Config) { @@ -174,6 +175,7 @@ func TestAllocations_GarbageCollect_ACL(t *testing.T) { } func TestAllocations_Stats(t *testing.T) { + t.Skip("missing exec driver plugin implementation") t.Parallel() require := require.New(t) client := TestClient(t, nil) diff --git a/client/alloc_watcher_e2e_test.go b/client/alloc_watcher_e2e_test.go index df6b0b774..efc1aee71 100644 --- a/client/alloc_watcher_e2e_test.go +++ b/client/alloc_watcher_e2e_test.go @@ -19,6 +19,7 @@ import ( // TestPrevAlloc_StreamAllocDir_TLS asserts ephemeral disk migrations still // work when TLS is enabled. func TestPrevAlloc_StreamAllocDir_TLS(t *testing.T) { + t.Skip("missing mock driver plugin implementation") const ( caFn = "../helper/tlsutil/testdata/global-ca.pem" serverCertFn = "../helper/tlsutil/testdata/global-server.pem" diff --git a/client/allocrunner/alloc_runner.go b/client/allocrunner/alloc_runner.go index 6f67e0f1f..2e13092fc 100644 --- a/client/allocrunner/alloc_runner.go +++ b/client/allocrunner/alloc_runner.go @@ -80,9 +80,6 @@ type allocRunner struct { // and if necessary migrate its alloc dir. prevAllocWatcher allocwatcher.PrevAllocWatcher - // pluginLoader is used to load plugins. - pluginLoader loader.PluginCatalog - // pluginSingletonLoader is a plugin loader that will returns singleton // instances of the plugins. pluginSingletonLoader loader.PluginCatalog @@ -109,7 +106,6 @@ func NewAllocRunner(config *Config) (*allocRunner, error) { stateUpdater: config.StateUpdater, allocBroadcaster: cstructs.NewAllocBroadcaster(), prevAllocWatcher: config.PrevAllocWatcher, - pluginLoader: config.PluginLoader, pluginSingletonLoader: config.PluginSingletonLoader, } @@ -143,7 +139,6 @@ func (ar *allocRunner) initTaskRunners(tasks []*structs.Task) error { StateUpdater: ar, Consul: ar.consulClient, VaultClient: ar.vaultClient, - PluginLoader: ar.pluginLoader, PluginSingletonLoader: ar.pluginSingletonLoader, } diff --git a/client/allocrunner/alloc_runner_test.go b/client/allocrunner/alloc_runner_test.go index 925375ab3..0638d0382 100644 --- a/client/allocrunner/alloc_runner_test.go +++ b/client/allocrunner/alloc_runner_test.go @@ -7,28 +7,35 @@ import ( "github.com/hashicorp/nomad/client/state" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/nomad/mock" + "github.com/hashicorp/nomad/plugins/shared/catalog" + "github.com/hashicorp/nomad/plugins/shared/loader" + "github.com/hashicorp/nomad/plugins/shared/singleton" "github.com/stretchr/testify/require" ) // TestAllocRunner_AllocState_Initialized asserts that getting TaskStates via // AllocState() are initialized even before the AllocRunner has run. func TestAllocRunner_AllocState_Initialized(t *testing.T) { + t.Skip("missing exec driver plugin implementation") t.Parallel() alloc := mock.Alloc() logger := testlog.HCLogger(t) conf := &Config{ - Alloc: alloc, - Logger: logger, - ClientConfig: config.TestClientConfig(), - StateDB: state.NoopDB{}, - Consul: nil, - Vault: nil, - StateUpdater: nil, - PrevAllocWatcher: nil, + Alloc: alloc, + Logger: logger, + ClientConfig: config.TestClientConfig(), + StateDB: state.NoopDB{}, + Consul: nil, + Vault: nil, + StateUpdater: nil, + PrevAllocWatcher: nil, + PluginSingletonLoader: &loader.MockCatalog{}, } + pluginLoader := catalog.TestPluginLoader(t) + conf.PluginSingletonLoader = singleton.NewSingletonLoader(logger, pluginLoader) ar, err := NewAllocRunner(conf) require.NoError(t, err) diff --git a/client/allocrunner/taskrunner/driver_handle.go b/client/allocrunner/taskrunner/driver_handle.go index 9aea784d9..0d3c07658 100644 --- a/client/allocrunner/taskrunner/driver_handle.go +++ b/client/allocrunner/taskrunner/driver_handle.go @@ -4,14 +4,14 @@ import ( "context" "time" - "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/plugins/drivers" ) -func NewDriverHandle(driver drivers.DriverPlugin, taskID string, task *structs.Task, net *cstructs.DriverNetwork) interfaces.DriverHandle { - return &driverHandleImpl{ +// NewDriverHandle returns a handle for task operations on a specific task +func NewDriverHandle(driver drivers.DriverPlugin, taskID string, task *structs.Task, net *cstructs.DriverNetwork) *DriverHandle { + return &DriverHandle{ driver: driver, net: net, taskID: taskID, @@ -19,38 +19,40 @@ func NewDriverHandle(driver drivers.DriverPlugin, taskID string, task *structs.T } } -type driverHandleImpl struct { +// DriverHandle encapsulates a driver plugin client and task identifier and exposes +// an api to perform driver operations on the task +type DriverHandle struct { driver drivers.DriverPlugin net *cstructs.DriverNetwork task *structs.Task taskID string } -func (h *driverHandleImpl) ID() string { +func (h *DriverHandle) ID() string { return h.taskID } -func (h *driverHandleImpl) WaitCh(ctx context.Context) (<-chan *drivers.ExitResult, error) { +func (h *DriverHandle) WaitCh(ctx context.Context) (<-chan *drivers.ExitResult, error) { return h.driver.WaitTask(ctx, h.taskID) } -func (h *driverHandleImpl) Update(task *structs.Task) error { +func (h *DriverHandle) Update(task *structs.Task) error { return nil } -func (h *driverHandleImpl) Kill() error { +func (h *DriverHandle) Kill() error { return h.driver.StopTask(h.taskID, h.task.KillTimeout, h.task.KillSignal) } -func (h *driverHandleImpl) Stats() (*cstructs.TaskResourceUsage, error) { +func (h *DriverHandle) Stats() (*cstructs.TaskResourceUsage, error) { return h.driver.TaskStats(h.taskID) } -func (h *driverHandleImpl) Signal(s string) error { +func (h *DriverHandle) Signal(s string) error { return h.driver.SignalTask(h.taskID, s) } -func (h *driverHandleImpl) Exec(timeout time.Duration, cmd string, args []string) ([]byte, int, error) { +func (h *DriverHandle) Exec(timeout time.Duration, cmd string, args []string) ([]byte, int, error) { command := append([]string{cmd}, args...) res, err := h.driver.ExecTask(h.taskID, command, timeout) if err != nil { @@ -59,6 +61,6 @@ func (h *driverHandleImpl) Exec(timeout time.Duration, cmd string, args []string return res.Stdout, res.ExitResult.ExitCode, res.ExitResult.Err } -func (h *driverHandleImpl) Network() *cstructs.DriverNetwork { +func (h *DriverHandle) Network() *cstructs.DriverNetwork { return h.net } diff --git a/client/allocrunner/taskrunner/lifecycle.go b/client/allocrunner/taskrunner/lifecycle.go index 551f8eb10..497fdbc93 100644 --- a/client/allocrunner/taskrunner/lifecycle.go +++ b/client/allocrunner/taskrunner/lifecycle.go @@ -52,7 +52,6 @@ func (tr *TaskRunner) Signal(event *structs.TaskEvent, s string) error { tr.EmitEvent(event) // Send the signal - return handle.Signal(s) } @@ -90,23 +89,24 @@ func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error // Block until task has exited. waitCh, err := handle.WaitCh(ctx) - // The task may have already been cleaned up + // The error should be nil or TaskNotFound, if it's something else then a + // failure in the driver or transport layer occured if err != nil && err != drivers.ErrTaskNotFound { tr.logger.Error("failed to wait on task. Resources may have been leaked", "error", err) return err } - if waitCh != nil { + if err == nil { <-waitCh - } - // Store that the task has been destroyed and any associated error. - tr.UpdateState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskKilled).SetKillError(destroyErr)) + // Store that the task has been destroyed and any associated error. + tr.UpdateState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskKilled).SetKillError(destroyErr)) - if destroyErr != nil { - return destroyErr - } else if err := ctx.Err(); err != nil { - return err + if destroyErr != nil { + return destroyErr + } else if err := ctx.Err(); err != nil { + return err + } } return nil diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index 87f060e22..38934b3a2 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -28,7 +28,6 @@ import ( "github.com/hashicorp/nomad/plugins/shared" "github.com/hashicorp/nomad/plugins/shared/hclspec" "github.com/hashicorp/nomad/plugins/shared/loader" - "github.com/zclconf/go-cty/cty" ) const ( @@ -77,10 +76,6 @@ type TaskRunner struct { // stateDB is for persisting localState and taskState stateDB cstate.StateDB - // persistedHash is the hash of the last persisted state for skipping - // unnecessary writes - persistedHash []byte - // ctx is the task runner's context representing the tasks's lifecycle. // Canceling the context will cause the task to be destroyed. ctx context.Context @@ -114,7 +109,7 @@ type TaskRunner struct { handleLock sync.Mutex // handle to the running driver - handle tinterfaces.DriverHandle + handle *DriverHandle // network is the configuration for the driver network if one was created network *cstructs.DriverNetwork @@ -161,9 +156,6 @@ type TaskRunner struct { resourceUsage *cstructs.TaskResourceUsage resourceUsageLock sync.Mutex - // PluginLoader is used to load plugins. - pluginLoader loader.PluginCatalog - // PluginSingletonLoader is a plugin loader that will returns singleton // instances of the plugins. pluginSingletonLoader loader.PluginCatalog @@ -189,9 +181,6 @@ type Config struct { // StateUpdater is used to emit updated task state StateUpdater interfaces.TaskStateHandler - // PluginLoader is used to load plugins. - PluginLoader loader.PluginCatalog - // PluginSingletonLoader is a plugin loader that will returns singleton // instances of the plugins. PluginSingletonLoader loader.PluginCatalog @@ -227,7 +216,6 @@ func NewTaskRunner(config *Config) (*TaskRunner, error) { ctxCancel: trCancel, triggerUpdateCh: make(chan struct{}, triggerUpdateChCap), waitCh: make(chan struct{}), - pluginLoader: config.PluginLoader, pluginSingletonLoader: config.PluginSingletonLoader, } @@ -437,16 +425,15 @@ func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { // runDriver runs the driver and waits for it to exit func (tr *TaskRunner) runDriver() error { + // TODO(nickethier): make sure this uses alloc.AllocatedResources once #4750 is rebased taskConfig := drivers.NewTaskConfig(tr.task, tr.taskDir, tr.envBuilder.Build()) taskConfig.ID = tr.buildID() taskConfig.StdoutPath = tr.logmonHookConfig.stdoutFifo taskConfig.StderrPath = tr.logmonHookConfig.stderrFifo + // TODO: load variables evalCtx := &hcl.EvalContext{ Functions: shared.GetStdlibFuncs(), - Variables: map[string]cty.Value{ - "NOMAD_ENV_bin": cty.StringVal("/bin/consul"), - }, } val, diag := shared.ParseHclInterface(tr.task.Config, tr.taskSchema, evalCtx) @@ -487,12 +474,7 @@ func (tr *TaskRunner) runDriver() error { func (tr *TaskRunner) updateDriverHandle(taskID string) { tr.handleLock.Lock() defer tr.handleLock.Unlock() - tr.handle = &driverHandleImpl{ - driver: tr.driver, - net: tr.network, - taskID: taskID, - task: tr.Task(), - } + tr.handle = NewDriverHandle(tr.driver, taskID, tr.Task(), tr.network) } // initDriver creates the driver for the task diff --git a/client/allocrunner/taskrunner/task_runner_getters.go b/client/allocrunner/taskrunner/task_runner_getters.go index 2affbcc3e..8a5661be0 100644 --- a/client/allocrunner/taskrunner/task_runner_getters.go +++ b/client/allocrunner/taskrunner/task_runner_getters.go @@ -58,7 +58,7 @@ func (tr *TaskRunner) getDriverHandle() interfaces.DriverHandle { } // setDriverHanlde sets the driver handle and creates a new result proxy. -func (tr *TaskRunner) setDriverHandle(handle interfaces.DriverHandle) { +func (tr *TaskRunner) setDriverHandle(handle *DriverHandle) { tr.handleLock.Lock() defer tr.handleLock.Unlock() tr.handle = handle diff --git a/client/client.go b/client/client.go index fd9935de9..f1bc1ae47 100644 --- a/client/client.go +++ b/client/client.go @@ -1039,64 +1039,64 @@ func (c *Client) updateNodeFromFingerprint(response *cstructs.FingerprintRespons func (c *Client) updateNodeFromDriver(name string, info *structs.DriverInfo) *structs.Node { c.configLock.Lock() defer c.configLock.Unlock() + if info == nil { + return c.configCopy.Node + } var hasChanged bool hadDriver := c.config.Node.Drivers[name] != nil - if info != nil { - if !hadDriver { - // If the driver info has not yet been set, do that here + if !hadDriver { + // If the driver info has not yet been set, do that here + hasChanged = true + c.config.Node.Drivers[name] = info + for attrName, newVal := range info.Attributes { + c.config.Node.Attributes[attrName] = newVal + } + } else { + oldVal := c.config.Node.Drivers[name] + // The driver info has already been set, fix it up + if oldVal.Detected != info.Detected { hasChanged = true - c.config.Node.Drivers[name] = info - for attrName, newVal := range info.Attributes { + c.config.Node.Drivers[name].Detected = info.Detected + } + + if oldVal.Healthy != info.Healthy || oldVal.HealthDescription != info.HealthDescription { + hasChanged = true + if info.HealthDescription != "" { + event := &structs.NodeEvent{ + Subsystem: "Driver", + Message: info.HealthDescription, + Timestamp: time.Now(), + Details: map[string]string{"driver": name}, + } + c.triggerNodeEvent(event) + } + } + + for attrName, newVal := range info.Attributes { + oldVal := c.config.Node.Drivers[name].Attributes[attrName] + if oldVal == newVal { + continue + } + + hasChanged = true + if newVal == "" { + delete(c.config.Node.Attributes, attrName) + } else { c.config.Node.Attributes[attrName] = newVal } - } else { - oldVal := c.config.Node.Drivers[name] - // The driver info has already been set, fix it up - if oldVal.Detected != info.Detected { - hasChanged = true - c.config.Node.Drivers[name].Detected = info.Detected - } - - if oldVal.Healthy != info.Healthy || oldVal.HealthDescription != info.HealthDescription { - hasChanged = true - if info.HealthDescription != "" { - event := &structs.NodeEvent{ - Subsystem: "Driver", - Message: info.HealthDescription, - Timestamp: time.Now(), - Details: map[string]string{"driver": name}, - } - c.triggerNodeEvent(event) - } - } - - for attrName, newVal := range info.Attributes { - oldVal := c.config.Node.Drivers[name].Attributes[attrName] - if oldVal == newVal { - continue - } - - hasChanged = true - if newVal == "" { - delete(c.config.Node.Attributes, attrName) - } else { - c.config.Node.Attributes[attrName] = newVal - } - } - } - - // COMPAT Remove in Nomad 0.10 - // We maintain the driver enabled attribute until all drivers expose - // their attributes as DriverInfo - driverName := fmt.Sprintf("driver.%s", name) - if info.Detected { - c.config.Node.Attributes[driverName] = "1" - } else { - delete(c.config.Node.Attributes, driverName) } + } + // COMPAT Remove in Nomad 0.10 + // We maintain the driver enabled attribute until all drivers expose + // their attributes as DriverInfo + driverName := fmt.Sprintf("driver.%s", name) + if info.Detected { + c.config.Node.Attributes[driverName] = "1" + } else { + delete(c.config.Node.Attributes, driverName) } if hasChanged { diff --git a/client/client_test.go b/client/client_test.go index a8a6c6df4..5b293a261 100644 --- a/client/client_test.go +++ b/client/client_test.go @@ -159,6 +159,7 @@ func TestClient_Fingerprint(t *testing.T) { } func TestClient_Fingerprint_Periodic(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() c1 := TestClient(t, func(c *config.Config) { @@ -399,6 +400,7 @@ func TestClient_Heartbeat(t *testing.T) { } func TestClient_UpdateAllocStatus(t *testing.T) { + t.Skip("missing exec driver plugin implementation") t.Parallel() s1, _ := testServer(t, nil) defer s1.Shutdown() @@ -662,6 +664,7 @@ func TestClient_Init(t *testing.T) { } func TestClient_BlockedAllocations(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() s1, _ := testServer(t, nil) defer s1.Shutdown() diff --git a/client/fingerprint_manager.go b/client/fingerprint_manager.go index 7254235c1..c6846676f 100644 --- a/client/fingerprint_manager.go +++ b/client/fingerprint_manager.go @@ -118,13 +118,9 @@ func (fp *FingerprintManager) Run() error { var availDrivers []string var skippedDrivers []string - var registeredDrivers []string for _, pl := range fp.singletonLoader.Catalog()[base.PluginTypeDriver] { - registeredDrivers = append(registeredDrivers, pl.Name) - } - - for _, name := range registeredDrivers { + name := pl.Name // Skip fingerprinting drivers that are not in the whitelist if it is // enabled. if _, ok := whitelistDrivers[name]; whitelistDriversEnabled && !ok { @@ -197,7 +193,7 @@ func (fm *FingerprintManager) setupDrivers(driverNames []string) error { driver, ok := plug.Plugin().(drivers.DriverPlugin) if !ok { - return fmt.Errorf("registered driver plugin %q does not implement DriverPlugin interface") + return fmt.Errorf("registered driver plugin %q does not implement DriverPlugin interface", name) } // Pass true for whether the health check is periodic here, so that the @@ -209,6 +205,7 @@ func (fm *FingerprintManager) setupDrivers(driverNames []string) error { ctx, cancel := context.WithCancel(context.Background()) fingerCh, err := driver.Fingerprint(ctx) if err != nil { + cancel() return err } @@ -218,10 +215,14 @@ func (fm *FingerprintManager) setupDrivers(driverNames []string) error { // attributes. go fm.watchDriverFingerprint(fingerCh, name, cancel) + if fm.logger.IsTrace() { + fm.logger.Trace("initial driver fingerprint", "driver", name, "fingerprint", finger) + } // Log the fingerprinters which have been applied if finger.Health != drivers.HealthStateUndetected { availDrivers = append(availDrivers, name) } + fm.processDriverFingerprint(finger, name) } fm.logger.Debug("detected drivers", "drivers", availDrivers) @@ -282,20 +283,69 @@ func (fm *FingerprintManager) watchDriverFingerprint(fpChan <-chan *drivers.Fing case <-fm.shutdownCh: cancel() return - case fp := <-fpChan: - di := &structs.DriverInfo{ - Attributes: fp.Attributes, - Detected: fp.Health != drivers.HealthStateUndetected, - Healthy: fp.Health == drivers.HealthStateHealthy, - HealthDescription: fp.HealthDescription, - UpdateTime: time.Now(), + case fp, ok := <-fpChan: + // if the channel is closed attempt to open a new one + if !ok { + newFpChan, newCancel, err := fm.retryDriverFingerprint(name) + if err != nil { + fm.logger.Warn("failed to fingerprint driver, retrying in 30s", "error", err) + fm.nodeLock.Lock() + n := fm.updateNodeFromDriver(name, &structs.DriverInfo{ + Healthy: false, + HealthDescription: "failed to fingerprint driver", + UpdateTime: time.Now(), + }) + if n != nil { + fm.node = n + } + fm.nodeLock.Unlock() + time.Sleep(30 * time.Second) + } else { + cancel() + fpChan = newFpChan + cancel = newCancel + } + continue + } else { + fm.processDriverFingerprint(fp, name) } - fm.nodeLock.Lock() - n := fm.updateNodeFromDriver(name, di) - if n != nil { - fm.node = n - } - fm.nodeLock.Unlock() } } } + +func (fm *FingerprintManager) processDriverFingerprint(fp *drivers.Fingerprint, driverName string) { + di := &structs.DriverInfo{ + Attributes: fp.Attributes, + Detected: fp.Health != drivers.HealthStateUndetected, + Healthy: fp.Health == drivers.HealthStateHealthy, + HealthDescription: fp.HealthDescription, + UpdateTime: time.Now(), + } + fm.nodeLock.Lock() + n := fm.updateNodeFromDriver(driverName, di) + if n != nil { + fm.node = n + } + fm.nodeLock.Unlock() +} + +func (fm *FingerprintManager) retryDriverFingerprint(driverName string) (<-chan *drivers.Fingerprint, context.CancelFunc, error) { + plug, err := fm.singletonLoader.Dispense(driverName, base.PluginTypeDriver, fm.logger) + if err != nil { + return nil, nil, err + } + + driver, ok := plug.Plugin().(drivers.DriverPlugin) + if !ok { + return nil, nil, fmt.Errorf("registered driver plugin %q does not implement DriverPlugin interface", driverName) + } + + ctx, cancel := context.WithCancel(context.Background()) + fingerCh, err := driver.Fingerprint(ctx) + if err != nil { + cancel() + return nil, nil, err + } + + return fingerCh, cancel, nil +} diff --git a/client/fingerprint_manager_test.go b/client/fingerprint_manager_test.go index 00e139689..1ccdd5a7c 100644 --- a/client/fingerprint_manager_test.go +++ b/client/fingerprint_manager_test.go @@ -8,9 +8,13 @@ import ( "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/testutil" "github.com/stretchr/testify/require" + + // registering raw_exec driver plugin used in testing + _ "github.com/hashicorp/nomad/drivers/rawexec" ) func TestFingerprintManager_Run_MockDriver(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) testClient := TestClient(t, nil) @@ -19,6 +23,7 @@ func TestFingerprintManager_Run_MockDriver(t *testing.T) { defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -46,6 +51,7 @@ func TestFingerprintManager_Run_ResourcesFingerprint(t *testing.T) { defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -67,12 +73,17 @@ func TestFingerprintManager_Run_ResourcesFingerprint(t *testing.T) { func TestFingerprintManager_Fingerprint_Run(t *testing.T) { t.Parallel() require := require.New(t) - testClient := TestClient(t, nil) + testClient := TestClient(t, func(c *config.Config) { + c.Options = map[string]string{ + "driver.raw_exec.enable": "true", + } + }) testClient.logger = testlog.HCLogger(t) defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -92,6 +103,7 @@ func TestFingerprintManager_Fingerprint_Run(t *testing.T) { } func TestFingerprintManager_Fingerprint_Periodic(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) testClient := TestClient(t, func(c *config.Config) { @@ -105,6 +117,7 @@ func TestFingerprintManager_Fingerprint_Periodic(t *testing.T) { defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -153,6 +166,7 @@ func TestFingerprintManager_Fingerprint_Periodic(t *testing.T) { // This is a temporary measure to check that a driver has both attributes on a // node set as well as DriverInfo. func TestFingerprintManager_HealthCheck_Driver(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) testClient := TestClient(t, func(c *config.Config) { @@ -167,6 +181,7 @@ func TestFingerprintManager_HealthCheck_Driver(t *testing.T) { defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -254,6 +269,7 @@ func TestFingerprintManager_HealthCheck_Driver(t *testing.T) { } func TestFingerprintManager_HealthCheck_Periodic(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) testClient := TestClient(t, func(c *config.Config) { @@ -267,6 +283,7 @@ func TestFingerprintManager_HealthCheck_Periodic(t *testing.T) { defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -349,6 +366,7 @@ func TestFingerprintManager_HealthCheck_Periodic(t *testing.T) { } func TestFimgerprintManager_Run_InWhitelist(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) @@ -363,6 +381,7 @@ func TestFimgerprintManager_Run_InWhitelist(t *testing.T) { defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -393,6 +412,7 @@ func TestFingerprintManager_Run_InBlacklist(t *testing.T) { defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -425,6 +445,7 @@ func TestFingerprintManager_Run_Combination(t *testing.T) { defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -458,6 +479,7 @@ func TestFingerprintManager_Run_WhitelistDrivers(t *testing.T) { defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -480,6 +502,7 @@ func TestFingerprintManager_Run_AllDriversBlacklisted(t *testing.T) { testClient := TestClient(t, func(c *config.Config) { c.Options = map[string]string{ + "driver.raw_exec.enable": "1", "driver.whitelist": " foo,bar,baz ", } }) @@ -488,6 +511,7 @@ func TestFingerprintManager_Run_AllDriversBlacklisted(t *testing.T) { defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -502,8 +526,8 @@ func TestFingerprintManager_Run_AllDriversBlacklisted(t *testing.T) { node := testClient.config.Node require.NotContains(node.Attributes, "driver.raw_exec") - require.NotContains(node.Attributes, "driver.exec") - require.NotContains(node.Attributes, "driver.docker") + //require.NotContains(node.Attributes, "driver.exec") + //require.NotContains(node.Attributes, "driver.docker") } func TestFingerprintManager_Run_DriversWhiteListBlacklistCombination(t *testing.T) { @@ -522,6 +546,7 @@ func TestFingerprintManager_Run_DriversWhiteListBlacklistCombination(t *testing. defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, @@ -555,6 +580,7 @@ func TestFingerprintManager_Run_DriversInBlacklist(t *testing.T) { defer testClient.Shutdown() fm := NewFingerprintManager( + testClient.config.PluginSingletonLoader, testClient.GetConfig, testClient.config.Node, testClient.shutdownCh, diff --git a/client/fs_endpoint_test.go b/client/fs_endpoint_test.go index b9e141acb..4986338e4 100644 --- a/client/fs_endpoint_test.go +++ b/client/fs_endpoint_test.go @@ -74,6 +74,7 @@ func TestFS_Stat_NoAlloc(t *testing.T) { } func TestFS_Stat(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) @@ -212,6 +213,7 @@ func TestFS_List_NoAlloc(t *testing.T) { } func TestFS_List(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) @@ -523,6 +525,7 @@ func TestFS_Stream_ACL(t *testing.T) { } func TestFS_Stream(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) @@ -633,6 +636,7 @@ func (r *ReadWriteCloseChecker) Close() error { } func TestFS_Stream_Follow(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) @@ -730,6 +734,7 @@ OUTER: } func TestFS_Stream_Limit(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) @@ -902,6 +907,7 @@ OUTER: // TestFS_Logs_TaskPending asserts that trying to stream logs for tasks which // have not started returns a 404 error. func TestFS_Logs_TaskPending(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) @@ -1130,6 +1136,7 @@ func TestFS_Logs_ACL(t *testing.T) { } func TestFS_Logs(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) @@ -1231,6 +1238,7 @@ OUTER: } func TestFS_Logs_Follow(t *testing.T) { + t.Skip("missing mock driver plugin implementation") t.Parallel() require := require.New(t) diff --git a/client/testing.go b/client/testing.go index 35d17b8d9..941094990 100644 --- a/client/testing.go +++ b/client/testing.go @@ -25,14 +25,15 @@ func TestClient(t testing.T, cb func(c *config.Config)) *Client { logger := testlog.HCLogger(t) conf.Logger = logger - // Set the plugin loaders - conf.PluginLoader = catalog.TestPluginLoader(t) - conf.PluginSingletonLoader = singleton.NewSingletonLoader(logger, conf.PluginLoader) - if cb != nil { cb(conf) } + // Set the plugin loaders + if conf.PluginLoader == nil { + conf.PluginLoader = catalog.TestPluginLoaderWithOptions(t, "", conf.Options, nil) + conf.PluginSingletonLoader = singleton.NewSingletonLoader(logger, conf.PluginLoader) + } catalog := consul.NewMockCatalog(logger) mockService := consulApi.NewMockConsulServiceClient(t, logger) client, err := NewClient(conf, catalog, mockService) diff --git a/command/agent/plugin_internal_catalog.go b/command/agent/plugin_internal_catalog.go new file mode 100644 index 000000000..034dea2cf --- /dev/null +++ b/command/agent/plugin_internal_catalog.go @@ -0,0 +1,12 @@ +package agent + +import ( + + // Each internal plugin has an init func which registers itself with the + // plugin catalog. Since the plugin implementations are not imported by the + // client or server they must be referenced here so plugin registration + // occures. + + // raw_exec driver + _ "github.com/hashicorp/nomad/drivers/rawexec" +) diff --git a/command/agent/plugins.go b/command/agent/plugins.go index 5d53129d2..6637d69f9 100644 --- a/command/agent/plugins.go +++ b/command/agent/plugins.go @@ -6,8 +6,6 @@ import ( "github.com/hashicorp/nomad/plugins/shared/catalog" "github.com/hashicorp/nomad/plugins/shared/loader" "github.com/hashicorp/nomad/plugins/shared/singleton" - - _ "github.com/hashicorp/nomad/drivers/rawexec" ) // setupPlugins is used to setup the plugin loaders. diff --git a/drivers/rawexec/driver.go b/drivers/rawexec/driver.go index 2691e321c..ae077a11f 100644 --- a/drivers/rawexec/driver.go +++ b/drivers/rawexec/driver.go @@ -23,6 +23,8 @@ import ( "golang.org/x/net/context" ) +// When the package is loaded the driver is registered as an internal plugin +// with the plugin catalog func init() { catalog.RegisterDeferredConfig(loader.PluginID{ Name: pluginName, @@ -32,15 +34,13 @@ func init() { Factory: func(l hclog.Logger) interface{} { return NewRawExecDriver(l) }, }, func(opts map[string]string) (map[string]interface{}, error) { - fmt.Println(opts) conf := map[string]interface{}{} - if v, ok := opts["driver.raw_exec.enable"]; ok && v == "true" { - conf["enabled"] = true + if v, err := strconv.ParseBool(opts["driver.raw_exec.enable"]); err == nil { + conf["enabled"] = v } - if v, ok := opts["driver.raw_exec.no_cgroups"]; ok && v == "true" { - conf["no_cgroups"] = true + if v, err := strconv.ParseBool(opts["driver.raw_exec.no_cgroups"]); err == nil { + conf["no_cgroups"] = v } - return conf, nil }, ) From 9686e1b258d7155b4652f4f4fcb38932956661fd Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Tue, 9 Oct 2018 23:05:22 -0400 Subject: [PATCH 123/152] client: fix broked tests from refactoring --- .../taskrunner/restarts/restarts_test.go | 26 ++++++++++--------- 1 file changed, 14 insertions(+), 12 deletions(-) diff --git a/client/allocrunner/taskrunner/restarts/restarts_test.go b/client/allocrunner/taskrunner/restarts/restarts_test.go index a671df83f..c8faa67ec 100644 --- a/client/allocrunner/taskrunner/restarts/restarts_test.go +++ b/client/allocrunner/taskrunner/restarts/restarts_test.go @@ -5,8 +5,8 @@ import ( "testing" "time" - cstructs "github.com/hashicorp/nomad/client/driver/structs" "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/drivers" ) func testPolicy(success bool, mode string) *structs.RestartPolicy { @@ -25,8 +25,10 @@ func withinJitter(expected, actual time.Duration) bool { expected.Nanoseconds()) <= jitter } -func testWaitResult(exit int) *cstructs.WaitResult { - return cstructs.NewWaitResult(exit, 0, nil) +func testExitResult(exit int) *drivers.ExitResult { + return &drivers.ExitResult{ + ExitCode: exit, + } } func TestClient_RestartTracker_ModeDelay(t *testing.T) { @@ -34,7 +36,7 @@ func TestClient_RestartTracker_ModeDelay(t *testing.T) { p := testPolicy(true, structs.RestartPolicyModeDelay) rt := NewRestartTracker(p, structs.JobTypeService) for i := 0; i < p.Attempts; i++ { - state, when := rt.SetWaitResult(testWaitResult(127)).GetState() + state, when := rt.SetExitResult(testExitResult(127)).GetState() if state != structs.TaskRestarting { t.Fatalf("NextRestart() returned %v, want %v", state, structs.TaskRestarting) } @@ -45,7 +47,7 @@ func TestClient_RestartTracker_ModeDelay(t *testing.T) { // Follow up restarts should cause delay. for i := 0; i < 3; i++ { - state, when := rt.SetWaitResult(testWaitResult(127)).GetState() + state, when := rt.SetExitResult(testExitResult(127)).GetState() if state != structs.TaskRestarting { t.Fail() } @@ -60,7 +62,7 @@ func TestClient_RestartTracker_ModeFail(t *testing.T) { p := testPolicy(true, structs.RestartPolicyModeFail) rt := NewRestartTracker(p, structs.JobTypeSystem) for i := 0; i < p.Attempts; i++ { - state, when := rt.SetWaitResult(testWaitResult(127)).GetState() + state, when := rt.SetExitResult(testExitResult(127)).GetState() if state != structs.TaskRestarting { t.Fatalf("NextRestart() returned %v, want %v", state, structs.TaskRestarting) } @@ -70,7 +72,7 @@ func TestClient_RestartTracker_ModeFail(t *testing.T) { } // Next restart should cause fail - if state, _ := rt.SetWaitResult(testWaitResult(127)).GetState(); state != structs.TaskNotRestarting { + if state, _ := rt.SetExitResult(testExitResult(127)).GetState(); state != structs.TaskNotRestarting { t.Fatalf("NextRestart() returned %v; want %v", state, structs.TaskNotRestarting) } } @@ -79,7 +81,7 @@ func TestClient_RestartTracker_NoRestartOnSuccess(t *testing.T) { t.Parallel() p := testPolicy(false, structs.RestartPolicyModeDelay) rt := NewRestartTracker(p, structs.JobTypeBatch) - if state, _ := rt.SetWaitResult(testWaitResult(0)).GetState(); state != structs.TaskTerminated { + if state, _ := rt.SetExitResult(testExitResult(0)).GetState(); state != structs.TaskTerminated { t.Fatalf("NextRestart() returned %v, expected: %v", state, structs.TaskTerminated) } } @@ -91,28 +93,28 @@ func TestClient_RestartTracker_ZeroAttempts(t *testing.T) { // Test with a non-zero exit code rt := NewRestartTracker(p, structs.JobTypeService) - if state, when := rt.SetWaitResult(testWaitResult(1)).GetState(); state != structs.TaskNotRestarting { + if state, when := rt.SetExitResult(testExitResult(1)).GetState(); state != structs.TaskNotRestarting { t.Fatalf("expect no restart, got restart/delay: %v/%v", state, when) } // Even with a zero (successful) exit code non-batch jobs should exit // with TaskNotRestarting rt = NewRestartTracker(p, structs.JobTypeService) - if state, when := rt.SetWaitResult(testWaitResult(0)).GetState(); state != structs.TaskNotRestarting { + if state, when := rt.SetExitResult(testExitResult(0)).GetState(); state != structs.TaskNotRestarting { t.Fatalf("expect no restart, got restart/delay: %v/%v", state, when) } // Batch jobs with a zero exit code and 0 attempts *do* exit cleanly // with Terminated rt = NewRestartTracker(p, structs.JobTypeBatch) - if state, when := rt.SetWaitResult(testWaitResult(0)).GetState(); state != structs.TaskTerminated { + if state, when := rt.SetExitResult(testExitResult(0)).GetState(); state != structs.TaskTerminated { t.Fatalf("expect terminated, got restart/delay: %v/%v", state, when) } // Batch jobs with a non-zero exit code and 0 attempts exit with // TaskNotRestarting rt = NewRestartTracker(p, structs.JobTypeBatch) - if state, when := rt.SetWaitResult(testWaitResult(1)).GetState(); state != structs.TaskNotRestarting { + if state, when := rt.SetExitResult(testExitResult(1)).GetState(); state != structs.TaskNotRestarting { t.Fatalf("expect no restart, got restart/delay: %v/%v", state, when) } } From ed3cdaf3d1ba1b80ef30193d01839aab1a6728f8 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 10 Oct 2018 20:07:52 -0400 Subject: [PATCH 124/152] plugin/driver: add Copy funcs --- plugins/drivers/driver.go | 32 ++++++++++++++++++++++++++++++++ plugins/drivers/task_handle.go | 7 +++++++ plugins/drivers/utils.go | 15 --------------- 3 files changed, 39 insertions(+), 15 deletions(-) diff --git a/plugins/drivers/driver.go b/plugins/drivers/driver.go index 7d503f451..f64d3dd5b 100644 --- a/plugins/drivers/driver.go +++ b/plugins/drivers/driver.go @@ -8,6 +8,7 @@ import ( "github.com/hashicorp/nomad/client/allocdir" cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/helper" "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/shared/hclspec" @@ -110,6 +111,17 @@ type TaskConfig struct { StderrPath string } +func (tc *TaskConfig) Copy() *TaskConfig { + if tc == nil { + return nil + } + c := new(TaskConfig) + *c = *tc + c.Env = helper.CopyMapStringString(c.Env) + c.Resources = tc.Resources.Copy() + return c +} + func (tc *TaskConfig) EnvList() []string { l := make([]string, 0, len(tc.Env)) for k, v := range tc.Env { @@ -161,6 +173,20 @@ type Resources struct { LinuxResources *LinuxResources } +func (r *Resources) Copy() *Resources { + if r == nil { + return nil + } + res := new(Resources) + if r.NomadResources != nil { + res.NomadResources = r.NomadResources.Copy() + } + if r.LinuxResources != nil { + res.LinuxResources = r.LinuxResources.Copy() + } + return res +} + type LinuxResources struct { CPUPeriod int64 CPUQuota int64 @@ -171,6 +197,12 @@ type LinuxResources struct { CpusetMems string } +func (r *LinuxResources) Copy() *LinuxResources { + res := new(LinuxResources) + *res = *r + return res +} + type DeviceConfig struct { TaskPath string HostPath string diff --git a/plugins/drivers/task_handle.go b/plugins/drivers/task_handle.go index 745ddea07..e71288952 100644 --- a/plugins/drivers/task_handle.go +++ b/plugins/drivers/task_handle.go @@ -27,3 +27,10 @@ func (h *TaskHandle) GetDriverState(v interface{}) error { return base.MsgPackDecode(h.driverState, v) } + +func (h *TaskHandle) Copy() *TaskHandle { + handle := new(TaskHandle) + *handle = *h + handle.Config = h.Config.Copy() + return handle +} diff --git a/plugins/drivers/utils.go b/plugins/drivers/utils.go index ba629abd5..ddf97fb6b 100644 --- a/plugins/drivers/utils.go +++ b/plugins/drivers/utils.go @@ -4,8 +4,6 @@ import ( "time" "github.com/golang/protobuf/ptypes" - "github.com/hashicorp/nomad/client/allocdir" - "github.com/hashicorp/nomad/client/driver/env" cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" "github.com/hashicorp/nomad/plugins/drivers/proto" @@ -47,19 +45,6 @@ func healthStateFromProto(pb proto.FingerprintResponse_HealthState) HealthState return HealthStateUndetected } -// NewTaskConfig builds a TaskConfig from a Task struct -func NewTaskConfig(task *structs.Task, taskDir *allocdir.TaskDir, env *env.TaskEnv) *TaskConfig { - return &TaskConfig{ - Name: task.Name, - Resources: &Resources{ - NomadResources: task.Resources, - }, - Env: env.Map(), - User: task.User, - AllocDir: taskDir.AllocDir, - } -} - func taskConfigFromProto(pb *proto.TaskConfig) *TaskConfig { if pb == nil { return &TaskConfig{} From 047fad2953d6ca29bb4fad6795ebab0997c992f9 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 10 Oct 2018 20:08:57 -0400 Subject: [PATCH 125/152] client: simplify driver plugin logic from review comments --- client/allocrunner/taskrunner/lifecycle.go | 21 ++-- client/allocrunner/taskrunner/state/state.go | 4 +- client/allocrunner/taskrunner/task_runner.go | 54 +++++----- client/fingerprint_manager.go | 103 ++++++++++--------- client/fingerprint_manager_test.go | 1 + 5 files changed, 95 insertions(+), 88 deletions(-) diff --git a/client/allocrunner/taskrunner/lifecycle.go b/client/allocrunner/taskrunner/lifecycle.go index 497fdbc93..7a3397696 100644 --- a/client/allocrunner/taskrunner/lifecycle.go +++ b/client/allocrunner/taskrunner/lifecycle.go @@ -91,22 +91,23 @@ func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error // The error should be nil or TaskNotFound, if it's something else then a // failure in the driver or transport layer occured - if err != nil && err != drivers.ErrTaskNotFound { + if err != nil { + if err == drivers.ErrTaskNotFound { + return nil + } tr.logger.Error("failed to wait on task. Resources may have been leaked", "error", err) return err } - if err == nil { - <-waitCh + <-waitCh - // Store that the task has been destroyed and any associated error. - tr.UpdateState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskKilled).SetKillError(destroyErr)) + // Store that the task has been destroyed and any associated error. + tr.UpdateState(structs.TaskStateDead, structs.NewTaskEvent(structs.TaskKilled).SetKillError(destroyErr)) - if destroyErr != nil { - return destroyErr - } else if err := ctx.Err(); err != nil { - return err - } + if destroyErr != nil { + return destroyErr + } else if err := ctx.Err(); err != nil { + return err } return nil diff --git a/client/allocrunner/taskrunner/state/state.go b/client/allocrunner/taskrunner/state/state.go index 1b4f07aba..94aedcc95 100644 --- a/client/allocrunner/taskrunner/state/state.go +++ b/client/allocrunner/taskrunner/state/state.go @@ -30,8 +30,8 @@ func (s *LocalState) Copy() *LocalState { // Create a copy c := &LocalState{ Hooks: make(map[string]*HookState, len(s.Hooks)), - DriverNetwork: s.DriverNetwork, - TaskHandle: s.TaskHandle, + DriverNetwork: s.DriverNetwork.Copy(), + TaskHandle: s.TaskHandle.Copy(), } // Copy the hooks diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index 38934b3a2..a9d10dc19 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -9,6 +9,7 @@ import ( metrics "github.com/armon/go-metrics" log "github.com/hashicorp/go-hclog" + multierror "github.com/hashicorp/go-multierror" "github.com/hashicorp/hcl2/hcl" "github.com/hashicorp/hcl2/hcldec" "github.com/hashicorp/nomad/client/allocdir" @@ -111,9 +112,6 @@ type TaskRunner struct { // handle to the running driver handle *DriverHandle - // network is the configuration for the driver network if one was created - network *cstructs.DriverNetwork - // task is the task being run task *structs.Task taskLock sync.RWMutex @@ -426,10 +424,7 @@ func (tr *TaskRunner) shouldRestart() (bool, time.Duration) { func (tr *TaskRunner) runDriver() error { // TODO(nickethier): make sure this uses alloc.AllocatedResources once #4750 is rebased - taskConfig := drivers.NewTaskConfig(tr.task, tr.taskDir, tr.envBuilder.Build()) - taskConfig.ID = tr.buildID() - taskConfig.StdoutPath = tr.logmonHookConfig.stdoutFifo - taskConfig.StderrPath = tr.logmonHookConfig.stderrFifo + taskConfig := tr.buildTaskConfig() // TODO: load variables evalCtx := &hcl.EvalContext{ @@ -438,16 +433,11 @@ func (tr *TaskRunner) runDriver() error { val, diag := shared.ParseHclInterface(tr.task.Config, tr.taskSchema, evalCtx) if diag.HasErrors() { - errStr := "failed to parse config" - for _, err := range diag.Errs() { - errStr = fmt.Sprintf("%s\n* %s", errStr, err.Error()) - } - return errors.New(errStr) + return multierror.Append(errors.New("failed to parse config"), diag.Errs()...) } if err := taskConfig.EncodeDriverConfig(val); err != nil { - tr.logger.Warn("failed to encode driver config", "error", err) - return err + return fmt.Errorf("failed to encode driver config: %v", err) } //TODO mounts and devices @@ -456,27 +446,19 @@ func (tr *TaskRunner) runDriver() error { // Start the job handle, net, err := tr.driver.StartTask(taskConfig) if err != nil { - tr.logger.Warn("driver start failed", "error", err) - return err + return fmt.Errorf("driver start failed: %v", err) } - tr.network = net tr.localStateLock.Lock() tr.localState.TaskHandle = handle tr.localStateLock.Unlock() - tr.updateDriverHandle(taskConfig.ID) + tr.setDriverHandle(NewDriverHandle(tr.driver, taskConfig.ID, tr.Task(), net)) // Emit an event that we started tr.UpdateState(structs.TaskStateRunning, structs.NewTaskEvent(structs.TaskStarted)) return nil } -func (tr *TaskRunner) updateDriverHandle(taskID string) { - tr.handleLock.Lock() - defer tr.handleLock.Unlock() - tr.handle = NewDriverHandle(tr.driver, taskID, tr.Task(), tr.network) -} - // initDriver creates the driver for the task /*func (tr *TaskRunner) initDriver() error { // Create a task-specific event emitter callback to expose minimal @@ -527,7 +509,10 @@ func (tr *TaskRunner) initDriver() error { if err != nil { return err } - spec, _ := hclspec.Convert(schema) + spec, diag := hclspec.Convert(schema) + if diag.HasErrors() { + return multierror.Append(errors.New("failed to convert task schema"), diag.Errs()...) + } tr.taskSchema = spec caps, err := tr.driver.Capabilities() @@ -575,9 +560,22 @@ func (tr *TaskRunner) persistLocalState() error { return tr.stateDB.PutTaskRunnerLocalState(tr.allocID, tr.taskName, tr.localState) } -// buildID builds a consistent unique ID for the task from the alloc ID, task name and restart attempt -func (tr *TaskRunner) buildID() string { - return fmt.Sprintf("%s/%s/%d", tr.allocID, tr.taskName, tr.restartTracker.GetCount()) +// buildTaskConfig builds a drivers.TaskConfig with an unique ID for the task. +// The ID is consistently built from the alloc ID, task name and restart attempt. +func (tr *TaskRunner) buildTaskConfig() *drivers.TaskConfig { + return &drivers.TaskConfig{ + ID: fmt.Sprintf("%s/%s/%d", tr.allocID, tr.taskName, tr.restartTracker.GetCount()), + Name: tr.task.Name, + Resources: &drivers.Resources{ + NomadResources: tr.task.Resources, + //TODO Calculate the LinuxResources + }, + Env: tr.envBuilder.Build().Map(), + User: tr.task.User, + AllocDir: tr.taskDir.AllocDir, + StdoutPath: tr.logmonHookConfig.stdoutFifo, + StderrPath: tr.logmonHookConfig.stderrFifo, + } } // XXX If the objects don't exists since the client shutdown before the task diff --git a/client/fingerprint_manager.go b/client/fingerprint_manager.go index c6846676f..4866f0e2e 100644 --- a/client/fingerprint_manager.go +++ b/client/fingerprint_manager.go @@ -16,6 +16,16 @@ import ( "github.com/hashicorp/nomad/plugins/shared/loader" ) +const ( + // driverFPBackoffBaseline is the baseline time for exponential backoff while + // fingerprinting a driver. + driverFPBackoffBaseline = 5 * time.Second + + // driverFPBackoffLimit is the limit of the exponential backoff for fingerprinting + // a driver. + driverFPBackoffLimit = 2 * time.Minute +) + // FingerprintManager runs a client fingerprinters on a continuous basis, and // updates the client when the node has changed type FingerprintManager struct { @@ -186,29 +196,11 @@ func (fm *FingerprintManager) setupDrivers(driverNames []string) error { var availDrivers []string for _, name := range driverNames { // TODO: driver reattach - plug, err := fm.singletonLoader.Dispense(name, base.PluginTypeDriver, fm.logger) + fingerCh, cancel, err := fm.dispenseDriverFingerprint(name) if err != nil { return err } - driver, ok := plug.Plugin().(drivers.DriverPlugin) - if !ok { - return fmt.Errorf("registered driver plugin %q does not implement DriverPlugin interface", name) - } - - // Pass true for whether the health check is periodic here, so that the - // fingerprinter will not set the initial health check status (this is set - // below, with an empty health status so that a node event is not - // triggered) - // Later, the periodic health checker will update this value for drivers - // where health checks are enabled. - ctx, cancel := context.WithCancel(context.Background()) - fingerCh, err := driver.Fingerprint(ctx) - if err != nil { - cancel() - return err - } - finger := <-fingerCh // Start a periodic watcher to detect changes to a drivers health and @@ -278,41 +270,56 @@ func (fm *FingerprintManager) fingerprint(name string, f fingerprint.Fingerprint // watchDrivers facilitates the different periods between fingerprint and // health checking a driver func (fm *FingerprintManager) watchDriverFingerprint(fpChan <-chan *drivers.Fingerprint, name string, cancel context.CancelFunc) { + var backoff time.Duration + var i int for { + if backoff > 0 { + time.Sleep(backoff) + } + select { case <-fm.shutdownCh: cancel() return case fp, ok := <-fpChan: - // if the channel is closed attempt to open a new one - if !ok { - newFpChan, newCancel, err := fm.retryDriverFingerprint(name) - if err != nil { - fm.logger.Warn("failed to fingerprint driver, retrying in 30s", "error", err) - fm.nodeLock.Lock() - n := fm.updateNodeFromDriver(name, &structs.DriverInfo{ - Healthy: false, - HealthDescription: "failed to fingerprint driver", - UpdateTime: time.Now(), - }) - if n != nil { - fm.node = n - } - fm.nodeLock.Unlock() - time.Sleep(30 * time.Second) - } else { - cancel() - fpChan = newFpChan - cancel = newCancel - } - continue - } else { + if ok { fm.processDriverFingerprint(fp, name) + continue } + // if the channel is closed attempt to open a new one + newFpChan, newCancel, err := fm.dispenseDriverFingerprint(name) + if err != nil { + fm.logger.Warn("failed to fingerprint driver, retrying in 30s", "error", err) + di := &structs.DriverInfo{ + Healthy: false, + HealthDescription: "failed to fingerprint driver", + UpdateTime: time.Now(), + } + if n := fm.updateNodeFromDriver(name, di); n != nil { + fm.setNode(n) + } + + // Calculate the new backoff + backoff = (1 << (2 * uint64(i))) * driverFPBackoffBaseline + if backoff > driverFPBackoffLimit { + backoff = driverFPBackoffLimit + } + i++ + continue + } + cancel() + fpChan = newFpChan + cancel = newCancel + + // Reset backoff + backoff = 0 + i = 0 } } } +// processDriverFringerprint converts a Fingerprint from a driver into a DriverInfo +// struct and updates the Node with it func (fm *FingerprintManager) processDriverFingerprint(fp *drivers.Fingerprint, driverName string) { di := &structs.DriverInfo{ Attributes: fp.Attributes, @@ -321,15 +328,15 @@ func (fm *FingerprintManager) processDriverFingerprint(fp *drivers.Fingerprint, HealthDescription: fp.HealthDescription, UpdateTime: time.Now(), } - fm.nodeLock.Lock() - n := fm.updateNodeFromDriver(driverName, di) - if n != nil { - fm.node = n + if n := fm.updateNodeFromDriver(driverName, di); n != nil { + fm.setNode(n) } - fm.nodeLock.Unlock() } -func (fm *FingerprintManager) retryDriverFingerprint(driverName string) (<-chan *drivers.Fingerprint, context.CancelFunc, error) { +// dispenseDriverFingerprint dispenses a driver plugin for the given driver name +// and requests a fingerprint channel. The channel and a context cancel function +// is returned to the caller +func (fm *FingerprintManager) dispenseDriverFingerprint(driverName string) (<-chan *drivers.Fingerprint, context.CancelFunc, error) { plug, err := fm.singletonLoader.Dispense(driverName, base.PluginTypeDriver, fm.logger) if err != nil { return nil, nil, err diff --git a/client/fingerprint_manager_test.go b/client/fingerprint_manager_test.go index 1ccdd5a7c..a21271376 100644 --- a/client/fingerprint_manager_test.go +++ b/client/fingerprint_manager_test.go @@ -526,6 +526,7 @@ func TestFingerprintManager_Run_AllDriversBlacklisted(t *testing.T) { node := testClient.config.Node require.NotContains(node.Attributes, "driver.raw_exec") + // TODO(nickethier): uncomment after missing driver implementations added //require.NotContains(node.Attributes, "driver.exec") //require.NotContains(node.Attributes, "driver.docker") } From 8cf669b5aa0975eda6b863b0981c74177f3a7817 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Wed, 10 Oct 2018 22:11:45 -0400 Subject: [PATCH 126/152] taskrunner: return error on waitCh --- client/allocrunner/taskrunner/lifecycle.go | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/client/allocrunner/taskrunner/lifecycle.go b/client/allocrunner/taskrunner/lifecycle.go index 7a3397696..a561ec8da 100644 --- a/client/allocrunner/taskrunner/lifecycle.go +++ b/client/allocrunner/taskrunner/lifecycle.go @@ -33,8 +33,9 @@ func (tr *TaskRunner) Restart(ctx context.Context, event *structs.TaskEvent, fai // Drain the wait channel or wait for the request context to be canceled waitCh, err := handle.WaitCh(ctx) if err != nil { - tr.logger.Error("failed to kill task. Resources may have been leaked", "error", err) + return err } + <-waitCh return nil } From c2d2bfa06ca306f9b8697f510017f2a33df82187 Mon Sep 17 00:00:00 2001 From: Preetha Appan Date: Thu, 11 Oct 2018 09:36:02 -0500 Subject: [PATCH 127/152] Get raw exec tests compiling and passing again --- drivers/rawexec/driver_test.go | 87 +++++++++++++++++------------ drivers/rawexec/driver_unix_test.go | 28 ++++++---- plugins/drivers/utils.go | 8 ++- 3 files changed, 75 insertions(+), 48 deletions(-) diff --git a/drivers/rawexec/driver_test.go b/drivers/rawexec/driver_test.go index dad5810c5..35c095899 100644 --- a/drivers/rawexec/driver_test.go +++ b/drivers/rawexec/driver_test.go @@ -12,12 +12,15 @@ import ( "testing" "time" + "github.com/hashicorp/hcl2/hcl" ctestutil "github.com/hashicorp/nomad/client/testutil" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/helper/testtask" "github.com/hashicorp/nomad/helper/uuid" basePlug "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/drivers" + "github.com/hashicorp/nomad/plugins/shared" + "github.com/hashicorp/nomad/plugins/shared/hclspec" "github.com/hashicorp/nomad/testutil" "github.com/stretchr/testify/require" ) @@ -111,14 +114,16 @@ func TestRawExecDriver_StartWait(t *testing.T) { ID: uuid.Generate(), Name: "test", } - task.EncodeDriverConfig(&TaskConfig{ - Command: "go", - Args: []string{"version"}, - }) + + taskConfig := map[string]interface{}{} + taskConfig["command"] = "go" + taskConfig["args"] = []string{"version"} + + encodeDriverHelper(require, task, taskConfig) cleanup := harness.MkAllocDir(task, false) defer cleanup() - handle, err := harness.StartTask(task) + handle, _, err := harness.StartTask(task) require.NoError(err) ch, err := harness.WaitTask(context.Background(), handle.Config.ID) @@ -138,14 +143,17 @@ func TestRawExecDriver_StartWaitStop(t *testing.T) { ID: uuid.Generate(), Name: "test", } - task.EncodeDriverConfig(&TaskConfig{ - Command: testtask.Path(), - Args: []string{"sleep", "100s"}, - }) + + taskConfig := map[string]interface{}{} + taskConfig["command"] = testtask.Path() + taskConfig["args"] = []string{"sleep", "100s"} + + encodeDriverHelper(require, task, taskConfig) + cleanup := harness.MkAllocDir(task, false) defer cleanup() - handle, err := harness.StartTask(task) + handle, _, err := harness.StartTask(task) require.NoError(err) ch, err := harness.WaitTask(context.Background(), handle.Config.ID) @@ -196,15 +204,17 @@ func TestRawExecDriver_StartWaitRecoverWaitStop(t *testing.T) { ID: uuid.Generate(), Name: "sleep", } - task.EncodeDriverConfig(&TaskConfig{ - Command: testtask.Path(), - Args: []string{"sleep", "100s"}, - }) + taskConfig := map[string]interface{}{} + taskConfig["command"] = testtask.Path() + taskConfig["args"] = []string{"sleep", "100s"} + + encodeDriverHelper(require, task, taskConfig) + testtask.SetTaskConfigEnv(task) cleanup := harness.MkAllocDir(task, false) defer cleanup() - handle, err := harness.StartTask(task) + handle, _, err := harness.StartTask(task) require.NoError(err) ch, err := harness.WaitTask(context.Background(), task.ID) @@ -277,17 +287,14 @@ func TestRawExecDriver_Start_Wait_AllocDir(t *testing.T) { exp := []byte("win") file := "output.txt" outPath := fmt.Sprintf(`%s/%s`, task.TaskDir().SharedAllocDir, file) - task.EncodeDriverConfig(&TaskConfig{ - Command: testtask.Path(), - Args: []string{ - "sleep", "1s", "write", - string(exp), outPath, - }, - }) + taskConfig := map[string]interface{}{} + taskConfig["command"] = testtask.Path() + taskConfig["args"] = []string{"sleep", "1s", "write", string(exp), outPath} + encodeDriverHelper(require, task, taskConfig) testtask.SetTaskConfigEnv(task) - _, err := harness.StartTask(task) + _, _, err := harness.StartTask(task) require.NoError(err) // Task should terminate quickly @@ -331,14 +338,13 @@ func TestRawExecDriver_Start_Kill_Wait_Cgroup(t *testing.T) { cleanup := harness.MkAllocDir(task, false) defer cleanup() - task.EncodeDriverConfig(&TaskConfig{ - Command: testtask.Path(), - Args: []string{"fork/exec", pidFile, "pgrp", "0", "sleep", "20s"}, - }) - + taskConfig := map[string]interface{}{} + taskConfig["command"] = testtask.Path() + taskConfig["args"] = []string{"fork/exec", pidFile, "pgrp", "0", "sleep", "20s"} + encodeDriverHelper(require, task, taskConfig) testtask.SetTaskConfigEnv(task) - _, err := harness.StartTask(task) + _, _, err := harness.StartTask(task) require.NoError(err) // Find the process @@ -419,14 +425,13 @@ func TestRawExecDriver_Exec(t *testing.T) { cleanup := harness.MkAllocDir(task, false) defer cleanup() - task.EncodeDriverConfig(&TaskConfig{ - Command: testtask.Path(), - Args: []string{"sleep", "9000s"}, - }) - + taskConfig := map[string]interface{}{} + taskConfig["command"] = testtask.Path() + taskConfig["args"] = []string{"sleep", "9000s"} + encodeDriverHelper(require, task, taskConfig) testtask.SetTaskConfigEnv(task) - _, err := harness.StartTask(task) + _, _, err := harness.StartTask(task) require.NoError(err) // Exec a command that should work @@ -443,3 +448,15 @@ func TestRawExecDriver_Exec(t *testing.T) { require.NoError(harness.DestroyTask(task.ID, true)) } + +func encodeDriverHelper(require *require.Assertions, task *drivers.TaskConfig, taskConfig map[string]interface{}) { + evalCtx := &hcl.EvalContext{ + Functions: shared.GetStdlibFuncs(), + } + spec, diag := hclspec.Convert(taskConfigSpec) + require.False(diag.HasErrors()) + taskConfigCtyVal, diag := shared.ParseHclInterface(taskConfig, spec, evalCtx) + require.False(diag.HasErrors()) + err := task.EncodeDriverConfig(taskConfigCtyVal) + require.Nil(err) +} diff --git a/drivers/rawexec/driver_unix_test.go b/drivers/rawexec/driver_unix_test.go index 7558df44f..b23e021d3 100644 --- a/drivers/rawexec/driver_unix_test.go +++ b/drivers/rawexec/driver_unix_test.go @@ -4,12 +4,13 @@ package rawexec import ( "context" + "runtime" + "testing" + "fmt" "io/ioutil" "path/filepath" - "runtime" "strings" - "testing" "time" "github.com/hashicorp/nomad/helper/testlog" @@ -39,13 +40,14 @@ func TestRawExecDriver_User(t *testing.T) { cleanup := harness.MkAllocDir(task, false) defer cleanup() - task.EncodeDriverConfig(&TaskConfig{ - Command: testtask.Path(), - Args: []string{"sleep", "45s"}, - }) + taskConfig := map[string]interface{}{} + taskConfig["command"] = testtask.Path() + taskConfig["args"] = []string{"sleep", "45s"} + encodeDriverHelper(require, task, taskConfig) testtask.SetTaskConfigEnv(task) - _, err := harness.StartTask(task) + + _, _, err := harness.StartTask(task) require.Error(err) msg := "unknown user alice" require.Contains(err.Error(), msg) @@ -69,10 +71,12 @@ func TestRawExecDriver_Signal(t *testing.T) { cleanup := harness.MkAllocDir(task, true) defer cleanup() - task.EncodeDriverConfig(&TaskConfig{ - Command: "/bin/bash", - Args: []string{"test.sh"}, - }) + taskConfig := map[string]interface{}{} + taskConfig["command"] = "/bin/bash" + taskConfig["args"] = []string{"test.sh"} + + encodeDriverHelper(require, task, taskConfig) + testtask.SetTaskConfigEnv(task) testFile := filepath.Join(task.TaskDir().Dir, "test.sh") testData := []byte(` @@ -88,7 +92,7 @@ done require.NoError(ioutil.WriteFile(testFile, testData, 0777)) testtask.SetTaskConfigEnv(task) - _, err := harness.StartTask(task) + _, _, err := harness.StartTask(task) require.NoError(err) go func() { diff --git a/plugins/drivers/utils.go b/plugins/drivers/utils.go index ddf97fb6b..dd0da0614 100644 --- a/plugins/drivers/utils.go +++ b/plugins/drivers/utils.go @@ -136,8 +136,10 @@ func resourcesFromProto(pb *proto.Resources) *Resources { } func resourcesToProto(r *Resources) *proto.Resources { + if r == nil { + return nil + } var pb proto.Resources - if r.NomadResources != nil { pb.RawResources = &proto.RawResources{ Cpu: int64(r.NomadResources.CPU), @@ -404,3 +406,7 @@ func resourceUsageFromProto(pb *proto.TaskResourceUsage) *cstructs.ResourceUsage MemoryStats: &memory, } } + +func BytesToMB(bytes int64) int64 { + return bytes / (1024 * 1024) +} From 15e8477c4e4736f540655f0849bc37de477f86e9 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Fri, 5 Oct 2018 15:12:03 -0400 Subject: [PATCH 128/152] rkt: start rkt driver plugin --- drivers/rkt/driver.go | 612 ++++++++++++++++++++++++++++++++++++++++++ drivers/rkt/state.go | 78 ++++++ 2 files changed, 690 insertions(+) create mode 100644 drivers/rkt/driver.go create mode 100644 drivers/rkt/state.go diff --git a/drivers/rkt/driver.go b/drivers/rkt/driver.go new file mode 100644 index 000000000..5a68c1afe --- /dev/null +++ b/drivers/rkt/driver.go @@ -0,0 +1,612 @@ +package rkt + +import ( + "bytes" + "context" + "fmt" + "net" + "os" + "os/exec" + "path/filepath" + "regexp" + "runtime" + "strings" + "syscall" + "time" + + hclog "github.com/hashicorp/go-hclog" + version "github.com/hashicorp/go-version" + "github.com/hashicorp/nomad/client/config" + "github.com/hashicorp/nomad/client/driver/env" + "github.com/hashicorp/nomad/client/driver/executor" + dstructs "github.com/hashicorp/nomad/client/driver/structs" + "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/drivers" + "github.com/hashicorp/nomad/plugins/drivers/utils" + "github.com/hashicorp/nomad/plugins/shared/hclspec" +) + +const ( + // pluginName is the name of the plugin + pluginName = "rkt" + + // fingerprintPeriod is the interval at which the driver will send fingerprint responses + fingerprintPeriod = 30 * time.Second + + // minRktVersion is the earliest supported version of rkt. rkt added support + // for CPU and memory isolators in 0.14.0. We cannot support an earlier + // version to maintain an uniform interface across all drivers + minRktVersion = "1.27.0" + + // rktCmd is the command rkt is installed as. + rktCmd = "rkt" + + // rktNetworkDeadline is how long to wait for container network to start + rktNetworkDeadline = 1 * time.Minute +) + +var ( + // pluginInfo is the response returned for the PluginInfo RPC + pluginInfo = &base.PluginInfoResponse{ + Type: base.PluginTypeDriver, + PluginApiVersion: "0.0.1", + PluginVersion: "0.1.0", + Name: pluginName, + } + + // configSpec is the hcl specification returned by the ConfigSchema RPC + configSpec = hclspec.NewObject(map[string]*hclspec.Spec{ + "volumes_enabled": hclspec.NewDefault( + hclspec.NewAttr("volumes_enabled", "bool", false), + hclspec.NewLiteral("true"), + ), + }) + + // taskConfigSpec is the hcl specification for the driver config section of + // a task within a job. It is returned in the TaskConfigSchema RPC + taskConfigSpec = hclspec.NewObject(map[string]*hclspec.Spec{ + "image": hclspec.NewAttr("image", "string", true), + "command": hclspec.NewAttr("command", "string", true), + "args": hclspec.NewAttr("command", "list(string)", false), + "trust_prefix": hclspec.NewAttr("trust_prefix", "string", false), + "dns_servers": hclspec.NewAttr("dns_servers", "list(string)", false), + "dns_search_domains": hclspec.NewAttr("dns_search_domains", "list(string)", false), + "net": hclspec.NewAttr("net", "list(string)", false), + "port_map": hclspec.NewAttr("port_map", "map(string)", false), + "volumes": hclspec.NewAttr("volumes", "list(string)", false), + "insecure_options": hclspec.NewAttr("insecure_options", "list(string)", false), + "no_overlay": hclspec.NewAttr("no_overlay", "bool", false), + "debug": hclspec.NewAttr("debug", "bool", false), + "group": hclspec.NewAttr("group", "string", false), + }) + + // capabilities is returned by the Capabilities RPC and indicates what + // optional features this driver supports + capabilities = &drivers.Capabilities{ + SendSignals: true, + Exec: true, + FSIsolation: drivers.FSIsolationChroot, + } + + reRktVersion = regexp.MustCompile(`rkt [vV]ersion[:]? (\d[.\d]+)`) + reAppcVersion = regexp.MustCompile(`appc [vV]ersion[:]? (\d[.\d]+)`) +) + +// Config is the client configuration for the driver +type Config struct { + // VolumesEnabled allows tasks to bind host paths (volumes) inside their + // container. Binding relative paths is always allowed and will be resolved + // relative to the allocation's directory. + VolumesEnabled bool `codec:"volumes_enabled"` +} + +// TaskConfig is the driver configuration of a task within a job +type TaskConfig struct { + ImageName string `codec:"image"` + Command string `codec:"command"` + Args []string `codec:"args"` + TrustPrefix string `codec:"trust_prefix"` + DNSServers []string `codec:"dns_servers"` // DNS Server for containers + DNSSearchDomains []string `codec:"dns_search_domains"` // DNS Search domains for containers + Net []string `codec:"net"` // Networks for the containers + PortMap map[string]string `codec:"port_map"` // A map of host port and the port name defined in the image manifest file + Volumes []string `codec:"volumes"` // Host-Volumes to mount in, syntax: /path/to/host/directory:/destination/path/in/container[:readOnly] + InsecureOptions []string `codec:"insecure_options"` // list of args for --insecure-options + + NoOverlay bool `codec:"no_overlay"` // disable overlayfs for rkt run + Debug bool `codec:"debug"` // Enable debug option for rkt command + Group string `codec:"group"` // Group override for the container +} + +// Driver is a driver for running images via Rkt +// We attempt to chose sane defaults for now, with more configuration available +// planned in the future +type Driver struct { + // eventer is used to handle multiplexing of TaskEvents calls such that an + // event can be broadcast to all callers + eventer *utils.Eventer + + // config is the driver configuration set by the SetConfig RPC + config *Config + + // tasks is the in memory datastore mapping taskIDs to execDriverHandles + tasks *taskStore + + // ctx is the context for the driver. It is passed to other subsystems to + // coordinate shutdown + ctx context.Context + + // signalShutdown is called when the driver is shutting down and cancels the + // ctx passed to any subsystems + signalShutdown context.CancelFunc + + // logger will log to the plugin output which is usually an 'executor.out' + // file located in the root of the TaskDir + logger hclog.Logger +} + +func NewDriver(logger hclog.Logger) *Driver { + + return nil +} + +func (d *Driver) PluginInfo() (*base.PluginInfoResponse, error) { + return pluginInfo, nil +} + +func (d *Driver) ConfigSchema() (*hclspec.Spec, error) { + return configSpec, nil +} + +func (d *Driver) SetConfig(data []byte) error { + var config Config + if err := base.MsgPackDecode(data, &config); err != nil { + return err + } + + d.config = &config + return nil +} + +func (d *Driver) TaskConfigSchema() (*hclspec.Spec, error) { + return taskConfigSpec, nil +} + +func (d *Driver) Capabilities() (*drivers.Capabilities, error) { + return capabilities, nil +} + +func (d *Driver) Fingerprint(ctx context.Context) (<-chan *drivers.Fingerprint, error) { + ch := make(chan *drivers.Fingerprint) + go d.handleFingerprint(ctx, ch) + return ch, nil + +} +func (d *Driver) handleFingerprint(ctx context.Context, ch chan *drivers.Fingerprint) { + ticker := time.NewTimer(0) + for { + select { + case <-ctx.Done(): + return + case <-d.ctx.Done(): + return + case <-ticker.C: + ticker.Reset(fingerprintPeriod) + ch <- d.buildFingerprint() + } + } +} + +func (d *Driver) buildFingerprint() *drivers.Fingerprint { + fingerprint := &drivers.Fingerprint{ + Attributes: map[string]string{}, + Health: drivers.HealthStateHealthy, + HealthDescription: "healthy", + } + + // Only enable if we are root when running on non-windows systems. + if runtime.GOOS != "windows" && syscall.Geteuid() != 0 { + d.logger.Debug("must run as root user, disabling") + fingerprint.Health = drivers.HealthStateUndetected + fingerprint.HealthDescription = "driver must run as root user" + return fingerprint + } + + outBytes, err := exec.Command(rktCmd, "version").Output() + if err != nil { + fingerprint.Health = drivers.HealthStateUndetected + fingerprint.HealthDescription = fmt.Sprintf("failed to executor %s version: %v", rktCmd, err) + return fingerprint + } + out := strings.TrimSpace(string(outBytes)) + + rktMatches := reRktVersion.FindStringSubmatch(out) + appcMatches := reAppcVersion.FindStringSubmatch(out) + if len(rktMatches) != 2 || len(appcMatches) != 2 { + fingerprint.Health = drivers.HealthStateUndetected + fingerprint.HealthDescription = "unable to parse rkt version string" + return fingerprint + } + + minVersion, _ := version.NewVersion(minRktVersion) + currentVersion, _ := version.NewVersion(rktMatches[1]) + if currentVersion.LessThan(minVersion) { + // Do not allow ancient rkt versions + fingerprint.Health = drivers.HealthStateUndetected + fingerprint.HealthDescription = fmt.Sprintf("unsuported rkt version %s", currentVersion) + d.logger.Warn("unsupported rkt version please upgrade to >= "+minVersion.String(), + "rkt_version", currentVersion) + return fingerprint + } + + fingerprint.Attributes["driver.rkt"] = "1" + fingerprint.Attributes["driver.rkt.version"] = rktMatches[1] + fingerprint.Attributes["driver.rkt.appc.version"] = appcMatches[1] + if d.config.VolumesEnabled { + fingerprint.Attributes["driver.rkt.volumes.enabled"] = "1" + } + + return fingerprint + +} + +func (d *Driver) RecoverTask(handle *drivers.TaskHandle) error { + panic("not implemented") +} + +func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) { + if _, ok := d.tasks.Get(cfg.ID); ok { + return nil, fmt.Errorf("task with ID '%s' already started", cfg.ID) + } + + var driverConfig TaskConfig + if err := cfg.DecodeDriverConfig(&driverConfig); err != nil { + return nil, fmt.Errorf("failed to decode driver config: %v", err) + } + + handle := drivers.NewTaskHandle(pluginName) + handle.Config = cfg + + panic("not implemented") + + // ACI image + img := driverConfig.ImageName + + // Global arguments given to both prepare and run-prepared + globalArgs := make([]string, 0, 50) + + // Add debug option to rkt command. + debug := driverConfig.Debug + + // Add the given trust prefix + trustPrefix := driverConfig.TrustPrefix + insecure := false + if trustPrefix != "" { + var outBuf, errBuf bytes.Buffer + cmd := exec.Command(rktCmd, "trust", "--skip-fingerprint-review=true", fmt.Sprintf("--prefix=%s", trustPrefix), fmt.Sprintf("--debug=%t", debug)) + cmd.Stdout = &outBuf + cmd.Stderr = &errBuf + if err := cmd.Run(); err != nil { + return nil, fmt.Errorf("Error running rkt trust: %s\n\nOutput: %s\n\nError: %s", + err, outBuf.String(), errBuf.String()) + } + d.logger.Debug("added trust prefix", "trust_prefix", trustPrefix) + } else { + // Disable signature verification if the trust command was not run. + insecure = true + } + + // if we have a selective insecure_options, prefer them + // insecure options are rkt's global argument, so we do this before the actual "run" + if len(driverConfig.InsecureOptions) > 0 { + globalArgs = append(globalArgs, fmt.Sprintf("--insecure-options=%s", strings.Join(driverConfig.InsecureOptions, ","))) + } else if insecure { + globalArgs = append(globalArgs, "--insecure-options=all") + } + + // debug is rkt's global argument, so add it before the actual "run" + globalArgs = append(globalArgs, fmt.Sprintf("--debug=%t", debug)) + + prepareArgs := make([]string, 0, 50) + runArgs := make([]string, 0, 50) + + prepareArgs = append(prepareArgs, globalArgs...) + prepareArgs = append(prepareArgs, "prepare") + runArgs = append(runArgs, globalArgs...) + runArgs = append(runArgs, "run-prepared") + + // disable overlayfs + if driverConfig.NoOverlay { + prepareArgs = append(prepareArgs, "--no-overlay=true") + } + + // Convert underscores to dashes in task names for use in volume names #2358 + sanitizedName := strings.Replace(cfg.Name, "_", "-", -1) + + // Mount /alloc + allocVolName := fmt.Sprintf("%s-%s-alloc", cfg.ID, sanitizedName) + prepareArgs = append(prepareArgs, fmt.Sprintf("--volume=%s,kind=host,source=%s", allocVolName, cfg.TaskDir().SharedAllocDir)) + prepareArgs = append(prepareArgs, fmt.Sprintf("--mount=volume=%s,target=%s", allocVolName, cfg.Env[env.AllocDir])) + + // Mount /local + localVolName := fmt.Sprintf("%s-%s-local", cfg.ID, sanitizedName) + prepareArgs = append(prepareArgs, fmt.Sprintf("--volume=%s,kind=host,source=%s", localVolName, cfg.TaskDir().LocalDir)) + prepareArgs = append(prepareArgs, fmt.Sprintf("--mount=volume=%s,target=%s", localVolName, cfg.Env[env.TaskLocalDir])) + + // Mount /secrets + secretsVolName := fmt.Sprintf("%s-%s-secrets", cfg.ID, sanitizedName) + prepareArgs = append(prepareArgs, fmt.Sprintf("--volume=%s,kind=host,source=%s", secretsVolName, cfg.TaskDir().SecretsDir)) + prepareArgs = append(prepareArgs, fmt.Sprintf("--mount=volume=%s,target=%s", secretsVolName, cfg.Env[env.SecretsDir])) + + // Mount arbitrary volumes if enabled + if len(driverConfig.Volumes) > 0 { + if !d.config.VolumesEnabled { + return nil, fmt.Errorf("volumes_enabled is false; cannot use rkt volumes: %+q", driverConfig.Volumes) + } + for i, rawvol := range driverConfig.Volumes { + parts := strings.Split(rawvol, ":") + readOnly := "false" + // job spec: + // volumes = ["/host/path:/container/path[:readOnly]"] + // the third parameter is optional, mount is read-write by default + if len(parts) == 3 { + if parts[2] == "readOnly" { + d.logger.Debug("mounting volume as readOnly", "volume", strings.Join(parts[:2], parts[1])) + readOnly = "true" + } else { + d.logger.Warn("unknown volume parameter ignored for mount", "parameter", parts[2], "mount", parts[0]) + } + } else if len(parts) != 2 { + return nil, fmt.Errorf("invalid rkt volume: %q", rawvol) + } + volName := fmt.Sprintf("%s-%s-%d", cfg.ID, sanitizedName, i) + prepareArgs = append(prepareArgs, fmt.Sprintf("--volume=%s,kind=host,source=%s,readOnly=%s", volName, parts[0], readOnly)) + prepareArgs = append(prepareArgs, fmt.Sprintf("--mount=volume=%s,target=%s", volName, parts[1])) + } + } + + // Inject environment variables + for k, v := range cfg.Env { + prepareArgs = append(prepareArgs, fmt.Sprintf("--set-env=%s=%s", k, v)) + } + + // Image is set here, because the commands that follow apply to it + prepareArgs = append(prepareArgs, img) + + // Check if the user has overridden the exec command. + if driverConfig.Command != "" { + prepareArgs = append(prepareArgs, fmt.Sprintf("--exec=%v", driverConfig.Command)) + } + + // Add memory isolator + prepareArgs = append(prepareArgs, fmt.Sprintf("--memory=%v", int64(cfg.Resources.MemoryLimitBytes))) + + // Add CPU isolator + prepareArgs = append(prepareArgs, fmt.Sprintf("--cpu-shares=%v", int64(cfg.Resources.CPUShares))) + + // Add DNS servers + if len(driverConfig.DNSServers) == 1 && (driverConfig.DNSServers[0] == "host" || driverConfig.DNSServers[0] == "none") { + // Special case single item lists with the special values "host" or "none" + runArgs = append(runArgs, fmt.Sprintf("--dns=%s", driverConfig.DNSServers[0])) + } else { + for _, ip := range driverConfig.DNSServers { + if err := net.ParseIP(ip); err == nil { + msg := fmt.Errorf("invalid ip address for container dns server %q", ip) + d.logger.Debug("error parsing DNS server", "error", msg) + return nil, msg + } + runArgs = append(runArgs, fmt.Sprintf("--dns=%s", ip)) + } + } + + // set DNS search domains + for _, domain := range driverConfig.DNSSearchDomains { + runArgs = append(runArgs, fmt.Sprintf("--dns-search=%s", domain)) + } + + // set network + network := strings.Join(driverConfig.Net, ",") + if network != "" { + runArgs = append(runArgs, fmt.Sprintf("--net=%s", network)) + } + + // Setup port mapping and exposed ports + /*if len(task.Resources.Networks) == 0 { + d.logger.Println("[DEBUG] driver.rkt: No network interfaces are available") + if len(driverConfig.PortMap) > 0 { + return nil, fmt.Errorf("Trying to map ports but no network interface is available") + } + } else if network == "host" { + // Port mapping is skipped when host networking is used. + d.logger.Println("[DEBUG] driver.rkt: Ignoring port_map when using --net=host") + } else { + // TODO add support for more than one network + network := task.Resources.Networks[0] + for _, port := range network.ReservedPorts { + var containerPort string + + mapped, ok := driverConfig.PortMap[port.Label] + if !ok { + // If the user doesn't have a mapped port using port_map, driver stops running container. + return nil, fmt.Errorf("port_map is not set. When you defined port in the resources, you need to configure port_map.") + } + containerPort = mapped + + hostPortStr := strconv.Itoa(port.Value) + + d.logger.Printf("[DEBUG] driver.rkt: exposed port %s", containerPort) + // Add port option to rkt run arguments. rkt allows multiple port args + prepareArgs = append(prepareArgs, fmt.Sprintf("--port=%s:%s", containerPort, hostPortStr)) + } + + for _, port := range network.DynamicPorts { + // By default we will map the allocated port 1:1 to the container + var containerPort string + + if mapped, ok := driverConfig.PortMap[port.Label]; ok { + containerPort = mapped + } else { + // If the user doesn't have mapped a port using port_map, driver stops running container. + return nil, fmt.Errorf("port_map is not set. When you defined port in the resources, you need to configure port_map.") + } + + hostPortStr := strconv.Itoa(port.Value) + + d.logger.Printf("[DEBUG] driver.rkt: exposed port %s", containerPort) + // Add port option to rkt run arguments. rkt allows multiple port args + prepareArgs = append(prepareArgs, fmt.Sprintf("--port=%s:%s", containerPort, hostPortStr)) + } + + }*/ + + // If a user has been specified for the task, pass it through to the user + if cfg.User != "" { + prepareArgs = append(prepareArgs, fmt.Sprintf("--user=%s", cfg.User)) + } + + // There's no task-level parameter for groups so check the driver + // config for a custom group + if driverConfig.Group != "" { + prepareArgs = append(prepareArgs, fmt.Sprintf("--group=%s", driverConfig.Group)) + } + + // Add user passed arguments. + if len(driverConfig.Args) != 0 { + + // Need to start arguments with "--" + prepareArgs = append(prepareArgs, "--") + + for _, arg := range driverConfig.Args { + prepareArgs = append(prepareArgs, fmt.Sprintf("%v", arg)) + } + } + + pluginLogFile := filepath.Join(cfg.TaskDir().Dir, fmt.Sprintf("%s-executor.out", cfg.Name)) + executorConfig := &dstructs.ExecutorConfig{ + LogFile: pluginLogFile, + LogLevel: "debug", + } + + // TODO: best way to pass port ranges in from client config + execIntf, pluginClient, err := utils.CreateExecutor(os.Stderr, hclog.Debug, 14000, 14512, executorConfig) + if err != nil { + return nil, err + } + + absPath, err := GetAbsolutePath(rktCmd) + if err != nil { + return nil, err + } + + var outBuf, errBuf bytes.Buffer + cmd := exec.Command(rktCmd, prepareArgs...) + cmd.Stdout = &outBuf + cmd.Stderr = &errBuf + d.logger.Debug("preparing task", "pod", img, "task_name", cfg.Name, "args", prepareArgs) + if err := cmd.Run(); err != nil { + return nil, fmt.Errorf("Error preparing rkt pod: %s\n\nOutput: %s\n\nError: %s", + err, outBuf.String(), errBuf.String()) + } + uuid := strings.TrimSpace(outBuf.String()) + d.logger.Debug("task prepared", "pod", img, "task_name", cfg.Name, "uuid", uuid) + runArgs = append(runArgs, uuid) + + // The task's environment is set via --set-env flags above, but the rkt + // command itself needs an evironment with PATH set to find iptables. + eb := env.NewEmptyBuilder() + filter := strings.Split(d.config.ReadDefault("env.blacklist", config.DefaultEnvBlacklist), ",") + rktEnv := eb.SetHostEnvvars(filter).Build() + + // Enable ResourceLimits to place the executor in a parent cgroup of + // the rkt container. This allows stats collection via the executor to + // work just like it does for exec. + execCmd := &executor.ExecCommand{ + Cmd: absPath, + Args: runArgs, + ResourceLimits: true, + Resources: &executor.Resources{ + CPU: int(cfg.Resources.CPUShares), + MemoryMB: int(cfg.Resources.MemoryLimitBytes), + //IOPS: task.Resources.IOPS, + //DiskMB: task.Resources.DiskMB, + }, + Env: cfg.EnvList(), + TaskDir: cfg.TaskDir().Dir, + StdoutPath: cfg.StdoutPath, + StderrPath: cfg.StderrPath, + } + ps, err := execIntf.Launch(execCmd) + if err != nil { + pluginClient.Kill() + return nil, err + } + + d.logger.Debug("started task", "aci", img, "uuid", uuid, "task_name", cfg.Name, "args", runArgs) + h := &rktHandle{ + uuid: uuid, + env: rktEnv, + taskDir: ctx.TaskDir, + pluginClient: pluginClient, + executor: execIntf, + executorPid: ps.Pid, + logger: d.logger, + killTimeout: GetKillTimeout(task.KillTimeout, maxKill), + maxKillTimeout: maxKill, + shutdownSignal: task.KillSignal, + doneCh: make(chan struct{}), + waitCh: make(chan *dstructs.WaitResult, 1), + } + go h.run() + + // Do not attempt to retrieve driver network if one won't exist: + // - "host" means the container itself has no networking metadata + // - "none" means no network is configured + // https://coreos.com/rkt/docs/latest/networking/overview.html#no-loopback-only-networking + var driverNetwork *cstructs.DriverNetwork + if network != "host" && network != "none" { + d.logger.Printf("[DEBUG] driver.rkt: retrieving network information for pod %q (UUID: %s) for task %q", img, uuid, d.taskName) + driverNetwork, err = rktGetDriverNetwork(uuid, driverConfig.PortMap, d.logger) + if err != nil && !pluginClient.Exited() { + d.logger.Printf("[WARN] driver.rkt: network status retrieval for pod %q (UUID: %s) for task %q failed. Last error: %v", img, uuid, d.taskName, err) + + // If a portmap was given, this turns into a fatal error + if len(driverConfig.PortMap) != 0 { + pluginClient.Kill() + return nil, fmt.Errorf("Trying to map ports but driver could not determine network information") + } + } + } + +} + +func (d *Driver) WaitTask(ctx context.Context, taskID string) (<-chan *drivers.ExitResult, error) { + panic("not implemented") +} + +func (d *Driver) StopTask(taskID string, timeout time.Duration, signal string) error { + panic("not implemented") +} + +func (d *Driver) DestroyTask(taskID string, force bool) error { + panic("not implemented") +} + +func (d *Driver) InspectTask(taskID string) (*drivers.TaskStatus, error) { + panic("not implemented") +} + +func (d *Driver) TaskStats(taskID string) (*drivers.TaskStats, error) { + panic("not implemented") +} + +func (d *Driver) TaskEvents(context.Context) (<-chan *drivers.TaskEvent, error) { + panic("not implemented") +} + +func (d *Driver) SignalTask(taskID string, signal string) error { + panic("not implemented") +} + +func (d *Driver) ExecTask(taskID string, cmd []string, timeout time.Duration) (*drivers.ExecTaskResult, error) { + panic("not implemented") +} diff --git a/drivers/rkt/state.go b/drivers/rkt/state.go new file mode 100644 index 000000000..834154dd4 --- /dev/null +++ b/drivers/rkt/state.go @@ -0,0 +1,78 @@ +package rkt + +import ( + "sync" + "time" + + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/client/driver/executor" + "github.com/hashicorp/nomad/plugins/drivers" +) + +type taskStore struct { + store map[string]*rktTaskHandle + lock sync.RWMutex +} + +func newTaskStore() *taskStore { + return &taskStore{store: map[string]*rktTaskHandle{}} +} + +func (ts *taskStore) Set(id string, handle *rktTaskHandle) { + ts.lock.Lock() + defer ts.lock.Unlock() + ts.store[id] = handle +} + +func (ts *taskStore) Get(id string) (*rktTaskHandle, bool) { + ts.lock.RLock() + defer ts.lock.RUnlock() + t, ok := ts.store[id] + return t, ok +} + +func (ts *taskStore) Delete(id string) { + ts.lock.Lock() + defer ts.lock.Unlock() + delete(ts.store, id) +} + +type rktTaskHandle struct { + exec executor.Executor + pid int + pluginClient *plugin.Client + logger hclog.Logger + + // stateLock syncs access to all fields below + stateLock sync.RWMutex + + task *drivers.TaskConfig + procState drivers.TaskState + startedAt time.Time + completedAt time.Time + exitResult *drivers.ExitResult +} + +func (h *rktTaskHandle) IsRunning() bool { + return h.procState == drivers.TaskStateRunning +} + +func (h *rktTaskHandle) run() { + if h.exitResult == nil { + h.exitResult = &drivers.ExitResult{} + } + + ps, err := h.exec.Wait() + if err != nil { + h.exitResult.Err = err + h.procState = drivers.TaskStateUnknown + h.completedAt = time.Now() + return + } + h.procState = drivers.TaskStateExited + h.exitResult.ExitCode = ps.ExitCode + h.exitResult.Signal = ps.Signal + h.completedAt = ps.Time + +} From c7ac1186c93b0c01bd0dd0a5f500564683b1e569 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Thu, 11 Oct 2018 15:42:08 -0400 Subject: [PATCH 129/152] client: add test for driverfailure during fingerprinting --- client/fingerprint_manager.go | 3 +- client/fingerprint_manager_test.go | 57 ++++++++++++++++++++++++++++++ drivers/rawexec/driver.go | 1 + plugins/drivers/client.go | 23 +++++++----- 4 files changed, 74 insertions(+), 10 deletions(-) diff --git a/client/fingerprint_manager.go b/client/fingerprint_manager.go index 4866f0e2e..0ea99bba0 100644 --- a/client/fingerprint_manager.go +++ b/client/fingerprint_manager.go @@ -276,13 +276,12 @@ func (fm *FingerprintManager) watchDriverFingerprint(fpChan <-chan *drivers.Fing if backoff > 0 { time.Sleep(backoff) } - select { case <-fm.shutdownCh: cancel() return case fp, ok := <-fpChan: - if ok { + if ok && fp.Err == nil { fm.processDriverFingerprint(fp, name) continue } diff --git a/client/fingerprint_manager_test.go b/client/fingerprint_manager_test.go index a21271376..7b0804399 100644 --- a/client/fingerprint_manager_test.go +++ b/client/fingerprint_manager_test.go @@ -4,8 +4,11 @@ import ( "fmt" "testing" + log "github.com/hashicorp/go-hclog" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/shared/loader" "github.com/hashicorp/nomad/testutil" "github.com/stretchr/testify/require" @@ -565,6 +568,60 @@ func TestFingerprintManager_Run_DriversWhiteListBlacklistCombination(t *testing. require.NotContains(node.Drivers, "exec") } +func TestFingerprintManager_Run_DriverFailure(t *testing.T) { + t.Parallel() + require := require.New(t) + + testClient := TestClient(t, func(c *config.Config) { + c.Options = map[string]string{ + "driver.raw_exec.enable": "1", + } + }) + + testClient.logger = testlog.HCLogger(t) + defer testClient.Shutdown() + + singLoader := testClient.config.PluginSingletonLoader + + dispenseCalls := 0 + loader := &loader.MockCatalog{ + DispenseF: func(name, pluginType string, logger log.Logger) (loader.PluginInstance, error) { + if pluginType == base.PluginTypeDriver && name == "raw_exec" { + dispenseCalls++ + } + return singLoader.Dispense(name, pluginType, logger) + }, + ReattachF: singLoader.Reattach, + CatalogF: singLoader.Catalog, + } + + fm := NewFingerprintManager( + loader, + testClient.GetConfig, + testClient.config.Node, + testClient.shutdownCh, + testClient.updateNodeFromFingerprint, + testClient.updateNodeFromDriver, + testClient.logger, + ) + + fpChan, cancel, err := fm.dispenseDriverFingerprint("raw_exec") + require.NoError(err) + require.Equal(1, dispenseCalls) + + cancel() + go fm.watchDriverFingerprint(fpChan, "raw_exec", cancel) + + testutil.WaitForResult(func() (bool, error) { + if 2 != dispenseCalls { + return false, fmt.Errorf("expected dispenseCalls to be 2 but was %d", dispenseCalls) + } + return true, nil + }, func(err error) { + require.NoError(err) + }) +} + func TestFingerprintManager_Run_DriversInBlacklist(t *testing.T) { t.Parallel() require := require.New(t) diff --git a/drivers/rawexec/driver.go b/drivers/rawexec/driver.go index ae077a11f..7e8fc2643 100644 --- a/drivers/rawexec/driver.go +++ b/drivers/rawexec/driver.go @@ -196,6 +196,7 @@ func (r *RawExecDriver) Fingerprint(ctx context.Context) (<-chan *drivers.Finger } func (r *RawExecDriver) handleFingerprint(ctx context.Context, ch chan *drivers.Fingerprint) { + defer close(ch) ticker := time.NewTimer(0) for { select { diff --git a/plugins/drivers/client.go b/plugins/drivers/client.go index 10de472cc..1fabb91f4 100644 --- a/plugins/drivers/client.go +++ b/plugins/drivers/client.go @@ -67,37 +67,44 @@ func (d *driverPluginClient) Capabilities() (*Capabilities, error) { func (d *driverPluginClient) Fingerprint(ctx context.Context) (<-chan *Fingerprint, error) { req := &proto.FingerprintRequest{} - stream, err := d.client.Fingerprint(context.Background(), req) + stream, err := d.client.Fingerprint(ctx, req) if err != nil { return nil, err } ch := make(chan *Fingerprint) - go d.handleFingerprint(ch, stream) + go d.handleFingerprint(ctx, ch, stream) return ch, nil } -func (d *driverPluginClient) handleFingerprint(ch chan *Fingerprint, stream proto.Driver_FingerprintClient) { +func (d *driverPluginClient) handleFingerprint(ctx context.Context, ch chan *Fingerprint, stream proto.Driver_FingerprintClient) { defer close(ch) for { pb, err := stream.Recv() if err == io.EOF { - break + return } if err != nil { d.logger.Error("error receiving stream from Fingerprint driver RPC", "error", err) - ch <- &Fingerprint{Err: fmt.Errorf("error from RPC stream: %v", err)} - break + select { + case <-ctx.Done(): + case ch <- &Fingerprint{Err: fmt.Errorf("error from RPC stream: %v", err)}: + } + return } f := &Fingerprint{ Attributes: pb.Attributes, Health: healthStateFromProto(pb.Health), HealthDescription: pb.HealthDescription, } - ch <- f - } + select { + case <-ctx.Done(): + return + case ch <- f: + } + } } // RecoverTask does internal state recovery to be able to control the task of From d9f0cbf4a9fcd31786b7079918ddf34f601b2808 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Thu, 11 Oct 2018 15:44:00 -0400 Subject: [PATCH 130/152] client: log retry during driver fingerprint redispense --- client/fingerprint_manager.go | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/client/fingerprint_manager.go b/client/fingerprint_manager.go index 0ea99bba0..ac07097dc 100644 --- a/client/fingerprint_manager.go +++ b/client/fingerprint_manager.go @@ -271,7 +271,7 @@ func (fm *FingerprintManager) fingerprint(name string, f fingerprint.Fingerprint // health checking a driver func (fm *FingerprintManager) watchDriverFingerprint(fpChan <-chan *drivers.Fingerprint, name string, cancel context.CancelFunc) { var backoff time.Duration - var i int + var retry int for { if backoff > 0 { time.Sleep(backoff) @@ -288,7 +288,7 @@ func (fm *FingerprintManager) watchDriverFingerprint(fpChan <-chan *drivers.Fing // if the channel is closed attempt to open a new one newFpChan, newCancel, err := fm.dispenseDriverFingerprint(name) if err != nil { - fm.logger.Warn("failed to fingerprint driver, retrying in 30s", "error", err) + fm.logger.Warn("failed to fingerprint driver, retrying in 30s", "error", err, "retry", retry) di := &structs.DriverInfo{ Healthy: false, HealthDescription: "failed to fingerprint driver", @@ -299,11 +299,11 @@ func (fm *FingerprintManager) watchDriverFingerprint(fpChan <-chan *drivers.Fing } // Calculate the new backoff - backoff = (1 << (2 * uint64(i))) * driverFPBackoffBaseline + backoff = (1 << (2 * uint64(retry))) * driverFPBackoffBaseline if backoff > driverFPBackoffLimit { backoff = driverFPBackoffLimit } - i++ + retry++ continue } cancel() @@ -312,7 +312,7 @@ func (fm *FingerprintManager) watchDriverFingerprint(fpChan <-chan *drivers.Fing // Reset backoff backoff = 0 - i = 0 + retry = 0 } } } From 678072ecd1bb10fa4cde1ff1be14eec62058d4f1 Mon Sep 17 00:00:00 2001 From: Preetha Appan Date: Fri, 12 Oct 2018 12:37:28 -0500 Subject: [PATCH 131/152] RKT driver plugin and unit tests --- drivers/rkt/driver.go | 618 ++++++++++++++++++++++++------ drivers/rkt/driver_test.go | 758 +++++++++++++++++++++++++++++++++++++ drivers/rkt/handle.go | 63 +++ drivers/rkt/state.go | 47 +-- plugins/drivers/driver.go | 17 +- plugins/drivers/utils.go | 5 +- 6 files changed, 1332 insertions(+), 176 deletions(-) create mode 100644 drivers/rkt/driver_test.go create mode 100644 drivers/rkt/handle.go diff --git a/drivers/rkt/driver.go b/drivers/rkt/driver.go index 5a68c1afe..988f9d7b6 100644 --- a/drivers/rkt/driver.go +++ b/drivers/rkt/driver.go @@ -1,8 +1,9 @@ +// +build linux + package rkt import ( "bytes" - "context" "fmt" "net" "os" @@ -14,16 +15,29 @@ import ( "syscall" "time" - hclog "github.com/hashicorp/go-hclog" - version "github.com/hashicorp/go-version" + appcschema "github.com/appc/spec/schema" + "github.com/hashicorp/go-hclog" + "github.com/hashicorp/go-plugin" + "github.com/hashicorp/go-version" "github.com/hashicorp/nomad/client/config" "github.com/hashicorp/nomad/client/driver/env" - "github.com/hashicorp/nomad/client/driver/executor" dstructs "github.com/hashicorp/nomad/client/driver/structs" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/drivers/shared/eventer" "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/drivers" "github.com/hashicorp/nomad/plugins/drivers/utils" "github.com/hashicorp/nomad/plugins/shared/hclspec" + rktv1 "github.com/rkt/rkt/api/v1" + "golang.org/x/net/context" + + "encoding/json" + "io/ioutil" + "math/rand" + "strconv" + + "github.com/hashicorp/consul-template/signals" + "github.com/hashicorp/nomad/client/driver/executor" ) const ( @@ -63,11 +77,11 @@ var ( }) // taskConfigSpec is the hcl specification for the driver config section of - // a task within a job. It is returned in the TaskConfigSchema RPC + // a taskConfig within a job. It is returned in the TaskConfigSchema RPC taskConfigSpec = hclspec.NewObject(map[string]*hclspec.Spec{ "image": hclspec.NewAttr("image", "string", true), - "command": hclspec.NewAttr("command", "string", true), - "args": hclspec.NewAttr("command", "list(string)", false), + "command": hclspec.NewAttr("command", "string", false), + "args": hclspec.NewAttr("args", "list(string)", false), "trust_prefix": hclspec.NewAttr("trust_prefix", "string", false), "dns_servers": hclspec.NewAttr("dns_servers", "list(string)", false), "dns_search_domains": hclspec.NewAttr("dns_search_domains", "list(string)", false), @@ -85,11 +99,13 @@ var ( capabilities = &drivers.Capabilities{ SendSignals: true, Exec: true, - FSIsolation: drivers.FSIsolationChroot, + FSIsolation: cstructs.FSIsolationChroot, } reRktVersion = regexp.MustCompile(`rkt [vV]ersion[:]? (\d[.\d]+)`) reAppcVersion = regexp.MustCompile(`appc [vV]ersion[:]? (\d[.\d]+)`) + + _ drivers.DriverPlugin = (*RktDriver)(nil) ) // Config is the client configuration for the driver @@ -100,31 +116,42 @@ type Config struct { VolumesEnabled bool `codec:"volumes_enabled"` } -// TaskConfig is the driver configuration of a task within a job +// TaskConfig is the driver configuration of a taskConfig within a job type TaskConfig struct { - ImageName string `codec:"image"` - Command string `codec:"command"` - Args []string `codec:"args"` - TrustPrefix string `codec:"trust_prefix"` - DNSServers []string `codec:"dns_servers"` // DNS Server for containers - DNSSearchDomains []string `codec:"dns_search_domains"` // DNS Search domains for containers - Net []string `codec:"net"` // Networks for the containers - PortMap map[string]string `codec:"port_map"` // A map of host port and the port name defined in the image manifest file - Volumes []string `codec:"volumes"` // Host-Volumes to mount in, syntax: /path/to/host/directory:/destination/path/in/container[:readOnly] - InsecureOptions []string `codec:"insecure_options"` // list of args for --insecure-options + ImageName string `codec:"image" cty:"image"` + Command string `codec:"command" cty:"command"` + Args []string `codec:"args" cty:"args"` + TrustPrefix string `codec:"trust_prefix" cty:"trust_prefix"` + DNSServers []string `codec:"dns_servers" cty:"dns_servers"` // DNS Server for containers + DNSSearchDomains []string `codec:"dns_search_domains" cty:"dns_search_domains"` // DNS Search domains for containers + Net []string `codec:"net" cty:"net"` // Networks for the containers + PortMap map[string]string `codec:"port_map" cty:"port_map"` // A map of host port and the port name defined in the image manifest file + Volumes []string `codec:"volumes" cty:"volumes"` // Host-Volumes to mount in, syntax: /path/to/host/directory:/destination/path/in/container[:readOnly] + InsecureOptions []string `codec:"insecure_options" cty:"insecure_options"` // list of args for --insecure-options - NoOverlay bool `codec:"no_overlay"` // disable overlayfs for rkt run - Debug bool `codec:"debug"` // Enable debug option for rkt command - Group string `codec:"group"` // Group override for the container + NoOverlay bool `codec:"no_overlay" cty:"no_overlay"` // disable overlayfs for rkt run + Debug bool `codec:"debug" cty:"debug"` // Enable debug option for rkt command + Group string `codec:"group" cty:"group"` // Group override for the container } -// Driver is a driver for running images via Rkt +// RktTaskState is the state which is encoded in the handle returned in +// StartTask. This information is needed to rebuild the taskConfig state and handler +// during recovery. +type RktTaskState struct { + ReattachConfig *utils.ReattachConfig + TaskConfig *drivers.TaskConfig + Pid int + StartedAt time.Time + UUID string +} + +// RktDriver is a driver for running images via Rkt // We attempt to chose sane defaults for now, with more configuration available // planned in the future -type Driver struct { +type RktDriver struct { // eventer is used to handle multiplexing of TaskEvents calls such that an // event can be broadcast to all callers - eventer *utils.Eventer + eventer *eventer.Eventer // config is the driver configuration set by the SetConfig RPC config *Config @@ -145,20 +172,28 @@ type Driver struct { logger hclog.Logger } -func NewDriver(logger hclog.Logger) *Driver { - - return nil +func NewRktDriver(logger hclog.Logger) drivers.DriverPlugin { + ctx, cancel := context.WithCancel(context.Background()) + logger = logger.Named(pluginName) + return &RktDriver{ + eventer: eventer.NewEventer(ctx, logger), + config: &Config{}, + tasks: newTaskStore(), + ctx: ctx, + signalShutdown: cancel, + logger: logger, + } } -func (d *Driver) PluginInfo() (*base.PluginInfoResponse, error) { +func (d *RktDriver) PluginInfo() (*base.PluginInfoResponse, error) { return pluginInfo, nil } -func (d *Driver) ConfigSchema() (*hclspec.Spec, error) { +func (d *RktDriver) ConfigSchema() (*hclspec.Spec, error) { return configSpec, nil } -func (d *Driver) SetConfig(data []byte) error { +func (d *RktDriver) SetConfig(data []byte) error { var config Config if err := base.MsgPackDecode(data, &config); err != nil { return err @@ -168,21 +203,21 @@ func (d *Driver) SetConfig(data []byte) error { return nil } -func (d *Driver) TaskConfigSchema() (*hclspec.Spec, error) { +func (d *RktDriver) TaskConfigSchema() (*hclspec.Spec, error) { return taskConfigSpec, nil } -func (d *Driver) Capabilities() (*drivers.Capabilities, error) { +func (d *RktDriver) Capabilities() (*drivers.Capabilities, error) { return capabilities, nil } -func (d *Driver) Fingerprint(ctx context.Context) (<-chan *drivers.Fingerprint, error) { +func (r *RktDriver) Fingerprint(ctx context.Context) (<-chan *drivers.Fingerprint, error) { ch := make(chan *drivers.Fingerprint) - go d.handleFingerprint(ctx, ch) + go r.handleFingerprint(ctx, ch) return ch, nil - } -func (d *Driver) handleFingerprint(ctx context.Context, ch chan *drivers.Fingerprint) { + +func (d *RktDriver) handleFingerprint(ctx context.Context, ch chan *drivers.Fingerprint) { ticker := time.NewTimer(0) for { select { @@ -197,7 +232,7 @@ func (d *Driver) handleFingerprint(ctx context.Context, ch chan *drivers.Fingerp } } -func (d *Driver) buildFingerprint() *drivers.Fingerprint { +func (d *RktDriver) buildFingerprint() *drivers.Fingerprint { fingerprint := &drivers.Fingerprint{ Attributes: map[string]string{}, Health: drivers.HealthStateHealthy, @@ -250,24 +285,75 @@ func (d *Driver) buildFingerprint() *drivers.Fingerprint { } -func (d *Driver) RecoverTask(handle *drivers.TaskHandle) error { - panic("not implemented") +func (d *RktDriver) RecoverTask(handle *drivers.TaskHandle) error { + if handle == nil { + return fmt.Errorf("error: handle cannot be nil") + } + + var taskState RktTaskState + if err := handle.GetDriverState(&taskState); err != nil { + d.logger.Error("failed to decode taskConfig state from handle", "error", err, "task_id", handle.Config.ID) + return fmt.Errorf("failed to decode taskConfig state from handle: %v", err) + } + + plugRC, err := utils.ReattachConfigToGoPlugin(taskState.ReattachConfig) + if err != nil { + d.logger.Error("failed to build ReattachConfig from taskConfig state", "error", err, "task_id", handle.Config.ID) + return fmt.Errorf("failed to build ReattachConfig from taskConfig state: %v", err) + } + + pluginConfig := &plugin.ClientConfig{ + Reattach: plugRC, + } + + execImpl, pluginClient, err := utils.CreateExecutorWithConfig(pluginConfig, os.Stderr) + if err != nil { + d.logger.Error("failed to reattach to executor", "error", err, "task_id", handle.Config.ID) + return fmt.Errorf("failed to reattach to executor: %v", err) + } + + // The taskConfig's environment is set via --set-env flags in Start, but the rkt + // command itself needs an environment with PATH set to find iptables. + // TODO (preetha) need to figure out how to read env.blacklist + eb := env.NewEmptyBuilder() + filter := strings.Split(config.DefaultEnvBlacklist, ",") + rktEnv := eb.SetHostEnvvars(filter).Build() + + h := &rktTaskHandle{ + exec: execImpl, + env: rktEnv, + pid: taskState.Pid, + uuid: taskState.UUID, + pluginClient: pluginClient, + taskConfig: taskState.TaskConfig, + procState: drivers.TaskStateRunning, + startedAt: taskState.StartedAt, + exitResult: &drivers.ExitResult{}, + } + + d.tasks.Set(taskState.TaskConfig.ID, h) + + go h.run() + return nil } -func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) { +func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cstructs.DriverNetwork, error) { if _, ok := d.tasks.Get(cfg.ID); ok { - return nil, fmt.Errorf("task with ID '%s' already started", cfg.ID) + return nil, nil, fmt.Errorf("taskConfig with ID '%s' already started", cfg.ID) } var driverConfig TaskConfig + d.logger.Debug("msgpack encoding", "data", string(cfg.GetRawConfig())) + if err := cfg.DecodeDriverConfig(&driverConfig); err != nil { - return nil, fmt.Errorf("failed to decode driver config: %v", err) + return nil, nil, fmt.Errorf("failed to decode driver config: %v", err) } handle := drivers.NewTaskHandle(pluginName) handle.Config = cfg - panic("not implemented") + // todo(preetha) - port map in client v1 is a slice of maps that get merged. figure out if the caller will do this + //driverConfig.PortMap // ACI image img := driverConfig.ImageName @@ -287,7 +373,7 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) cmd.Stdout = &outBuf cmd.Stderr = &errBuf if err := cmd.Run(); err != nil { - return nil, fmt.Errorf("Error running rkt trust: %s\n\nOutput: %s\n\nError: %s", + return nil, nil, fmt.Errorf("Error running rkt trust: %s\n\nOutput: %s\n\nError: %s", err, outBuf.String(), errBuf.String()) } d.logger.Debug("added trust prefix", "trust_prefix", trustPrefix) @@ -320,28 +406,28 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) prepareArgs = append(prepareArgs, "--no-overlay=true") } - // Convert underscores to dashes in task names for use in volume names #2358 + // Convert underscores to dashes in taskConfig names for use in volume names #2358 sanitizedName := strings.Replace(cfg.Name, "_", "-", -1) // Mount /alloc allocVolName := fmt.Sprintf("%s-%s-alloc", cfg.ID, sanitizedName) prepareArgs = append(prepareArgs, fmt.Sprintf("--volume=%s,kind=host,source=%s", allocVolName, cfg.TaskDir().SharedAllocDir)) - prepareArgs = append(prepareArgs, fmt.Sprintf("--mount=volume=%s,target=%s", allocVolName, cfg.Env[env.AllocDir])) + prepareArgs = append(prepareArgs, fmt.Sprintf("--mount=volume=%s,target=%s", allocVolName, "/alloc")) // Mount /local localVolName := fmt.Sprintf("%s-%s-local", cfg.ID, sanitizedName) prepareArgs = append(prepareArgs, fmt.Sprintf("--volume=%s,kind=host,source=%s", localVolName, cfg.TaskDir().LocalDir)) - prepareArgs = append(prepareArgs, fmt.Sprintf("--mount=volume=%s,target=%s", localVolName, cfg.Env[env.TaskLocalDir])) + prepareArgs = append(prepareArgs, fmt.Sprintf("--mount=volume=%s,target=%s", localVolName, "/local")) // Mount /secrets secretsVolName := fmt.Sprintf("%s-%s-secrets", cfg.ID, sanitizedName) prepareArgs = append(prepareArgs, fmt.Sprintf("--volume=%s,kind=host,source=%s", secretsVolName, cfg.TaskDir().SecretsDir)) - prepareArgs = append(prepareArgs, fmt.Sprintf("--mount=volume=%s,target=%s", secretsVolName, cfg.Env[env.SecretsDir])) + prepareArgs = append(prepareArgs, fmt.Sprintf("--mount=volume=%s,target=%s", secretsVolName, "/secrets")) // Mount arbitrary volumes if enabled if len(driverConfig.Volumes) > 0 { if !d.config.VolumesEnabled { - return nil, fmt.Errorf("volumes_enabled is false; cannot use rkt volumes: %+q", driverConfig.Volumes) + return nil, nil, fmt.Errorf("volumes_enabled is false; cannot use rkt volumes: %+q", driverConfig.Volumes) } for i, rawvol := range driverConfig.Volumes { parts := strings.Split(rawvol, ":") @@ -357,7 +443,7 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) d.logger.Warn("unknown volume parameter ignored for mount", "parameter", parts[2], "mount", parts[0]) } } else if len(parts) != 2 { - return nil, fmt.Errorf("invalid rkt volume: %q", rawvol) + return nil, nil, fmt.Errorf("invalid rkt volume: %q", rawvol) } volName := fmt.Sprintf("%s-%s-%d", cfg.ID, sanitizedName, i) prepareArgs = append(prepareArgs, fmt.Sprintf("--volume=%s,kind=host,source=%s,readOnly=%s", volName, parts[0], readOnly)) @@ -379,10 +465,10 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) } // Add memory isolator - prepareArgs = append(prepareArgs, fmt.Sprintf("--memory=%v", int64(cfg.Resources.MemoryLimitBytes))) + prepareArgs = append(prepareArgs, fmt.Sprintf("--memory=%v", int64(cfg.Resources.LinuxResources.MemoryLimitBytes))) // Add CPU isolator - prepareArgs = append(prepareArgs, fmt.Sprintf("--cpu-shares=%v", int64(cfg.Resources.CPUShares))) + prepareArgs = append(prepareArgs, fmt.Sprintf("--cpu-shares=%v", int64(cfg.Resources.LinuxResources.CPUShares))) // Add DNS servers if len(driverConfig.DNSServers) == 1 && (driverConfig.DNSServers[0] == "host" || driverConfig.DNSServers[0] == "none") { @@ -391,9 +477,9 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) } else { for _, ip := range driverConfig.DNSServers { if err := net.ParseIP(ip); err == nil { - msg := fmt.Errorf("invalid ip address for container dns server %q", ip) - d.logger.Debug("error parsing DNS server", "error", msg) - return nil, msg + wrappedErr := fmt.Errorf("invalid ip address for container dns server %q", ip) + d.logger.Debug("error parsing DNS server", "error", wrappedErr) + return nil, nil, wrappedErr } runArgs = append(runArgs, fmt.Sprintf("--dns=%s", ip)) } @@ -411,30 +497,29 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) } // Setup port mapping and exposed ports - /*if len(task.Resources.Networks) == 0 { - d.logger.Println("[DEBUG] driver.rkt: No network interfaces are available") + if len(cfg.Resources.NomadResources.Networks) == 0 { + d.logger.Debug("no network interfaces are available") if len(driverConfig.PortMap) > 0 { - return nil, fmt.Errorf("Trying to map ports but no network interface is available") + return nil, nil, fmt.Errorf("Trying to map ports but no network interface is available") } } else if network == "host" { // Port mapping is skipped when host networking is used. - d.logger.Println("[DEBUG] driver.rkt: Ignoring port_map when using --net=host") + d.logger.Debug("driver.rkt: Ignoring port_map when using --net=host") } else { - // TODO add support for more than one network - network := task.Resources.Networks[0] + network := cfg.Resources.NomadResources.Networks[0] for _, port := range network.ReservedPorts { var containerPort string mapped, ok := driverConfig.PortMap[port.Label] if !ok { // If the user doesn't have a mapped port using port_map, driver stops running container. - return nil, fmt.Errorf("port_map is not set. When you defined port in the resources, you need to configure port_map.") + return nil, nil, fmt.Errorf("port_map is not set. When you defined port in the resources, you need to configure port_map.") } containerPort = mapped hostPortStr := strconv.Itoa(port.Value) - d.logger.Printf("[DEBUG] driver.rkt: exposed port %s", containerPort) + d.logger.Debug("driver.rkt: exposed port", "containerPort", containerPort) // Add port option to rkt run arguments. rkt allows multiple port args prepareArgs = append(prepareArgs, fmt.Sprintf("--port=%s:%s", containerPort, hostPortStr)) } @@ -447,24 +532,24 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) containerPort = mapped } else { // If the user doesn't have mapped a port using port_map, driver stops running container. - return nil, fmt.Errorf("port_map is not set. When you defined port in the resources, you need to configure port_map.") + return nil, nil, fmt.Errorf("port_map is not set. When you defined port in the resources, you need to configure port_map.") } hostPortStr := strconv.Itoa(port.Value) - d.logger.Printf("[DEBUG] driver.rkt: exposed port %s", containerPort) + d.logger.Debug("driver.rkt: exposed port", "containerPort", containerPort) // Add port option to rkt run arguments. rkt allows multiple port args prepareArgs = append(prepareArgs, fmt.Sprintf("--port=%s:%s", containerPort, hostPortStr)) } - }*/ + } - // If a user has been specified for the task, pass it through to the user + // If a user has been specified for the taskConfig, pass it through to the user if cfg.User != "" { prepareArgs = append(prepareArgs, fmt.Sprintf("--user=%s", cfg.User)) } - // There's no task-level parameter for groups so check the driver + // There's no taskConfig-level parameter for groups so check the driver // config for a custom group if driverConfig.Group != "" { prepareArgs = append(prepareArgs, fmt.Sprintf("--group=%s", driverConfig.Group)) @@ -488,33 +573,35 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) } // TODO: best way to pass port ranges in from client config - execIntf, pluginClient, err := utils.CreateExecutor(os.Stderr, hclog.Debug, 14000, 14512, executorConfig) + execImpl, pluginClient, err := utils.CreateExecutor(os.Stderr, hclog.Debug, 14000, 14512, executorConfig) if err != nil { - return nil, err + return nil, nil, err } absPath, err := GetAbsolutePath(rktCmd) if err != nil { - return nil, err + return nil, nil, err } var outBuf, errBuf bytes.Buffer cmd := exec.Command(rktCmd, prepareArgs...) cmd.Stdout = &outBuf cmd.Stderr = &errBuf - d.logger.Debug("preparing task", "pod", img, "task_name", cfg.Name, "args", prepareArgs) + d.logger.Debug("preparing taskConfig", "pod", img, "task_name", cfg.Name, "args", prepareArgs) if err := cmd.Run(); err != nil { - return nil, fmt.Errorf("Error preparing rkt pod: %s\n\nOutput: %s\n\nError: %s", + return nil, nil, fmt.Errorf("Error preparing rkt pod: %s\n\nOutput: %s\n\nError: %s", err, outBuf.String(), errBuf.String()) } uuid := strings.TrimSpace(outBuf.String()) - d.logger.Debug("task prepared", "pod", img, "task_name", cfg.Name, "uuid", uuid) + d.logger.Debug("taskConfig prepared", "pod", img, "task_name", cfg.Name, "uuid", uuid) runArgs = append(runArgs, uuid) - // The task's environment is set via --set-env flags above, but the rkt - // command itself needs an evironment with PATH set to find iptables. + // The taskConfig's environment is set via --set-env flags above, but the rkt + // command itself needs an environment with PATH set to find iptables. + + // TODO (preetha) need to figure out how to pass env.blacklist from client config eb := env.NewEmptyBuilder() - filter := strings.Split(d.config.ReadDefault("env.blacklist", config.DefaultEnvBlacklist), ",") + filter := strings.Split(config.DefaultEnvBlacklist, ",") rktEnv := eb.SetHostEnvvars(filter).Build() // Enable ResourceLimits to place the executor in a parent cgroup of @@ -525,37 +612,50 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) Args: runArgs, ResourceLimits: true, Resources: &executor.Resources{ - CPU: int(cfg.Resources.CPUShares), - MemoryMB: int(cfg.Resources.MemoryLimitBytes), - //IOPS: task.Resources.IOPS, - //DiskMB: task.Resources.DiskMB, + CPU: int(cfg.Resources.LinuxResources.CPUShares), + MemoryMB: int(drivers.BytesToMB(cfg.Resources.LinuxResources.MemoryLimitBytes)), + DiskMB: cfg.Resources.NomadResources.DiskMB, }, Env: cfg.EnvList(), TaskDir: cfg.TaskDir().Dir, StdoutPath: cfg.StdoutPath, StderrPath: cfg.StderrPath, } - ps, err := execIntf.Launch(execCmd) + ps, err := execImpl.Launch(execCmd) if err != nil { pluginClient.Kill() - return nil, err + return nil, nil, err } - d.logger.Debug("started task", "aci", img, "uuid", uuid, "task_name", cfg.Name, "args", runArgs) - h := &rktHandle{ - uuid: uuid, - env: rktEnv, - taskDir: ctx.TaskDir, - pluginClient: pluginClient, - executor: execIntf, - executorPid: ps.Pid, - logger: d.logger, - killTimeout: GetKillTimeout(task.KillTimeout, maxKill), - maxKillTimeout: maxKill, - shutdownSignal: task.KillSignal, - doneCh: make(chan struct{}), - waitCh: make(chan *dstructs.WaitResult, 1), + d.logger.Debug("started taskConfig", "aci", img, "uuid", uuid, "task_name", cfg.Name, "args", runArgs) + h := &rktTaskHandle{ + exec: execImpl, + env: rktEnv, + pid: ps.Pid, + uuid: uuid, + pluginClient: pluginClient, + taskConfig: cfg, + procState: drivers.TaskStateRunning, + startedAt: time.Now().Round(time.Millisecond), + logger: d.logger, } + + rktDriverState := RktTaskState{ + ReattachConfig: utils.ReattachConfigFromGoPlugin(pluginClient.ReattachConfig()), + Pid: ps.Pid, + TaskConfig: cfg, + StartedAt: h.startedAt, + UUID: uuid, + } + + if err := handle.SetDriverState(&rktDriverState); err != nil { + d.logger.Error("failed to start task, error setting driver state", "error", err) + execImpl.Shutdown("", 0) + pluginClient.Kill() + return nil, nil, fmt.Errorf("failed to set driver state: %v", err) + } + + d.tasks.Set(cfg.ID, h) go h.run() // Do not attempt to retrieve driver network if one won't exist: @@ -564,49 +664,333 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, error) // https://coreos.com/rkt/docs/latest/networking/overview.html#no-loopback-only-networking var driverNetwork *cstructs.DriverNetwork if network != "host" && network != "none" { - d.logger.Printf("[DEBUG] driver.rkt: retrieving network information for pod %q (UUID: %s) for task %q", img, uuid, d.taskName) + d.logger.Debug("retrieving network information for pod", "pod", img, "UUID", uuid, "task", cfg.Name) driverNetwork, err = rktGetDriverNetwork(uuid, driverConfig.PortMap, d.logger) if err != nil && !pluginClient.Exited() { - d.logger.Printf("[WARN] driver.rkt: network status retrieval for pod %q (UUID: %s) for task %q failed. Last error: %v", img, uuid, d.taskName, err) + d.logger.Warn("network status retrieval for pod failed", "pod", img, "UUID", uuid, "task", cfg.Name, "error", err) // If a portmap was given, this turns into a fatal error if len(driverConfig.PortMap) != 0 { pluginClient.Kill() - return nil, fmt.Errorf("Trying to map ports but driver could not determine network information") + return nil, nil, fmt.Errorf("Trying to map ports but driver could not determine network information") } } } + return handle, driverNetwork, nil + } -func (d *Driver) WaitTask(ctx context.Context, taskID string) (<-chan *drivers.ExitResult, error) { - panic("not implemented") +func (d *RktDriver) WaitTask(ctx context.Context, taskID string) (<-chan *drivers.ExitResult, error) { + handle, ok := d.tasks.Get(taskID) + if !ok { + return nil, drivers.ErrTaskNotFound + } + + ch := make(chan *drivers.ExitResult) + go d.handleWait(ctx, handle, ch) + + return ch, nil } -func (d *Driver) StopTask(taskID string, timeout time.Duration, signal string) error { - panic("not implemented") +func (d *RktDriver) StopTask(taskID string, timeout time.Duration, signal string) error { + handle, ok := d.tasks.Get(taskID) + if !ok { + return drivers.ErrTaskNotFound + } + + if err := handle.exec.Shutdown(signal, timeout); err != nil { + if handle.pluginClient.Exited() { + return nil + } + return fmt.Errorf("executor Shutdown failed: %v", err) + } + + return nil } -func (d *Driver) DestroyTask(taskID string, force bool) error { - panic("not implemented") +func (d *RktDriver) DestroyTask(taskID string, force bool) error { + handle, ok := d.tasks.Get(taskID) + if !ok { + return drivers.ErrTaskNotFound + } + + if handle.IsRunning() && !force { + return fmt.Errorf("cannot destroy running task") + } + + if !handle.pluginClient.Exited() { + if handle.IsRunning() { + if err := handle.exec.Shutdown("", 0); err != nil { + handle.logger.Error("destroying executor failed", "err", err) + } + } + + handle.pluginClient.Kill() + } + + d.tasks.Delete(taskID) + return nil } -func (d *Driver) InspectTask(taskID string) (*drivers.TaskStatus, error) { - panic("not implemented") +func (d *RktDriver) InspectTask(taskID string) (*drivers.TaskStatus, error) { + handle, ok := d.tasks.Get(taskID) + if !ok { + return nil, drivers.ErrTaskNotFound + } + + handle.stateLock.RLock() + defer handle.stateLock.RUnlock() + + status := &drivers.TaskStatus{ + ID: handle.taskConfig.ID, + Name: handle.taskConfig.Name, + State: handle.procState, + StartedAt: handle.startedAt, + CompletedAt: handle.completedAt, + ExitResult: handle.exitResult, + DriverAttributes: map[string]string{ + "pid": strconv.Itoa(handle.pid), + }, + } + + return status, nil } -func (d *Driver) TaskStats(taskID string) (*drivers.TaskStats, error) { - panic("not implemented") +func (d *RktDriver) TaskStats(taskID string) (*cstructs.TaskResourceUsage, error) { + handle, ok := d.tasks.Get(taskID) + if !ok { + return nil, drivers.ErrTaskNotFound + } + + return handle.exec.Stats() } -func (d *Driver) TaskEvents(context.Context) (<-chan *drivers.TaskEvent, error) { - panic("not implemented") +func (d *RktDriver) TaskEvents(ctx context.Context) (<-chan *drivers.TaskEvent, error) { + return d.eventer.TaskEvents(ctx) } -func (d *Driver) SignalTask(taskID string, signal string) error { - panic("not implemented") +func (d *RktDriver) SignalTask(taskID string, signal string) error { + handle, ok := d.tasks.Get(taskID) + if !ok { + return drivers.ErrTaskNotFound + } + + sig := os.Interrupt + if s, ok := signals.SignalLookup[signal]; ok { + d.logger.Warn("signal to send to task unknown, using SIGINT", "signal", signal, "task_id", handle.taskConfig.ID) + sig = s + } + return handle.exec.Signal(sig) } -func (d *Driver) ExecTask(taskID string, cmd []string, timeout time.Duration) (*drivers.ExecTaskResult, error) { - panic("not implemented") +func (d *RktDriver) ExecTask(taskID string, cmdArgs []string, timeout time.Duration) (*drivers.ExecTaskResult, error) { + if len(cmdArgs) == 0 { + return nil, fmt.Errorf("error cmd must have atleast one value") + } + handle, ok := d.tasks.Get(taskID) + if !ok { + return nil, drivers.ErrTaskNotFound + } + // enter + UUID + cmd + args... + cmd := cmdArgs[0] + args := cmdArgs[1:] + enterArgs := make([]string, 3+len(args)) + enterArgs[0] = "enter" + enterArgs[1] = handle.uuid + enterArgs[2] = handle.env.ReplaceEnv(cmd) + copy(enterArgs[3:], handle.env.ParseAndReplace(args)) + out, exitCode, err := handle.exec.Exec(time.Now().Add(timeout), rktCmd, enterArgs) + if err != nil { + return nil, err + } + + return &drivers.ExecTaskResult{ + Stdout: out, + ExitResult: &drivers.ExitResult{ + ExitCode: exitCode, + }, + }, nil + +} + +// GetAbsolutePath returns the absolute path of the passed binary by resolving +// it in the path and following symlinks. +func GetAbsolutePath(bin string) (string, error) { + lp, err := exec.LookPath(bin) + if err != nil { + return "", fmt.Errorf("failed to resolve path to %q executable: %v", bin, err) + } + + return filepath.EvalSymlinks(lp) +} + +func rktGetDriverNetwork(uuid string, driverConfigPortMap map[string]string, logger hclog.Logger) (*cstructs.DriverNetwork, error) { + deadline := time.Now().Add(rktNetworkDeadline) + var lastErr error + try := 0 + + for time.Now().Before(deadline) { + try++ + if status, err := rktGetStatus(uuid, logger); err == nil { + for _, net := range status.Networks { + if !net.IP.IsGlobalUnicast() { + continue + } + + // Get the pod manifest so we can figure out which ports are exposed + var portmap map[string]int + manifest, err := rktGetManifest(uuid) + if err == nil { + portmap, err = rktManifestMakePortMap(manifest, driverConfigPortMap) + if err != nil { + lastErr = fmt.Errorf("could not create manifest-based portmap: %v", err) + return nil, lastErr + } + } else { + lastErr = fmt.Errorf("could not get pod manifest: %v", err) + return nil, lastErr + } + + // This is a successful landing; log if its not the first attempt. + if try > 1 { + logger.Debug("retrieved network info for pod", "uuid", uuid, "attempt", try) + } + return &cstructs.DriverNetwork{ + PortMap: portmap, + IP: status.Networks[0].IP.String(), + }, nil + } + + if len(status.Networks) == 0 { + lastErr = fmt.Errorf("no networks found") + } else { + lastErr = fmt.Errorf("no good driver networks out of %d returned", len(status.Networks)) + } + } else { + lastErr = fmt.Errorf("getting status failed: %v", err) + } + + waitTime := getJitteredNetworkRetryTime() + logger.Debug("failed getting network info for pod, sleeping", "uuid", uuid, "attempt", try, "err", lastErr, "wait", waitTime) + time.Sleep(waitTime) + } + return nil, fmt.Errorf("timed out, last error: %v", lastErr) +} + +// Given a rkt/appc pod manifest and driver portmap configuration, create +// a driver portmap. +func rktManifestMakePortMap(manifest *appcschema.PodManifest, configPortMap map[string]string) (map[string]int, error) { + if len(manifest.Apps) == 0 { + return nil, fmt.Errorf("manifest has no apps") + } + if len(manifest.Apps) != 1 { + return nil, fmt.Errorf("manifest has multiple apps!") + } + app := manifest.Apps[0] + if app.App == nil { + return nil, fmt.Errorf("specified app has no App object") + } + + portMap := make(map[string]int) + for svc, name := range configPortMap { + for _, port := range app.App.Ports { + if port.Name.String() == name { + portMap[svc] = int(port.Port) + } + } + } + return portMap, nil +} + +// Retrieve pod status for the pod with the given UUID. +func rktGetStatus(uuid string, logger hclog.Logger) (*rktv1.Pod, error) { + statusArgs := []string{ + "status", + "--format=json", + uuid, + } + var outBuf, errBuf bytes.Buffer + cmd := exec.Command(rktCmd, statusArgs...) + cmd.Stdout = &outBuf + cmd.Stderr = &errBuf + if err := cmd.Run(); err != nil { + if outBuf.Len() > 0 { + logger.Debug("status output for UUID", "uuid", uuid, elide(outBuf)) + } + if errBuf.Len() == 0 { + return nil, err + } + logger.Debug("status error output", "uuid", uuid, "error", elide(errBuf)) + return nil, fmt.Errorf("%s. stderr: %q", err, elide(errBuf)) + } + var status rktv1.Pod + if err := json.Unmarshal(outBuf.Bytes(), &status); err != nil { + return nil, err + } + return &status, nil +} + +// Retrieves a pod manifest +func rktGetManifest(uuid string) (*appcschema.PodManifest, error) { + statusArgs := []string{ + "cat-manifest", + uuid, + } + var outBuf bytes.Buffer + cmd := exec.Command(rktCmd, statusArgs...) + cmd.Stdout = &outBuf + cmd.Stderr = ioutil.Discard + if err := cmd.Run(); err != nil { + return nil, err + } + var manifest appcschema.PodManifest + if err := json.Unmarshal(outBuf.Bytes(), &manifest); err != nil { + return nil, err + } + return &manifest, nil +} + +// Create a time with a 0 to 100ms jitter for rktGetDriverNetwork retries +func getJitteredNetworkRetryTime() time.Duration { + return time.Duration(900+rand.Intn(100)) * time.Millisecond +} + +// Conditionally elide a buffer to an arbitrary length +func elideToLen(inBuf bytes.Buffer, length int) bytes.Buffer { + if inBuf.Len() > length { + inBuf.Truncate(length) + inBuf.WriteString("...") + } + return inBuf +} + +// Conditionally elide a buffer to an 80 character string +func elide(inBuf bytes.Buffer) string { + tempBuf := elideToLen(inBuf, 80) + return tempBuf.String() +} + +func (d *RktDriver) handleWait(ctx context.Context, handle *rktTaskHandle, ch chan *drivers.ExitResult) { + defer close(ch) + var result *drivers.ExitResult + ps, err := handle.exec.Wait() + if err != nil { + result = &drivers.ExitResult{ + Err: fmt.Errorf("executor: error waiting on process: %v", err), + } + } else { + result = &drivers.ExitResult{ + ExitCode: ps.ExitCode, + Signal: ps.Signal, + } + } + + select { + case <-ctx.Done(): + return + case <-d.ctx.Done(): + return + case ch <- result: + } } diff --git a/drivers/rkt/driver_test.go b/drivers/rkt/driver_test.go new file mode 100644 index 000000000..faa31a644 --- /dev/null +++ b/drivers/rkt/driver_test.go @@ -0,0 +1,758 @@ +// +build linux + +package rkt + +import ( + "fmt" + "io/ioutil" + "path/filepath" + "sync" + "testing" + "time" + + "os" + + "bytes" + + "github.com/hashicorp/hcl2/hcl" + ctestutil "github.com/hashicorp/nomad/client/testutil" + "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/helper/testtask" + "github.com/hashicorp/nomad/helper/uuid" + "github.com/hashicorp/nomad/nomad/structs" + basePlug "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/drivers" + "github.com/hashicorp/nomad/plugins/shared" + "github.com/hashicorp/nomad/plugins/shared/hclspec" + "github.com/hashicorp/nomad/testutil" + "github.com/stretchr/testify/require" + "golang.org/x/net/context" +) + +func TestRktVersionRegex(t *testing.T) { + ctestutil.RktCompatible(t) + t.Parallel() + + inputRkt := "rkt version 0.8.1" + inputAppc := "appc version 1.2.0" + expectedRkt := "0.8.1" + expectedAppc := "1.2.0" + rktMatches := reRktVersion.FindStringSubmatch(inputRkt) + appcMatches := reAppcVersion.FindStringSubmatch(inputAppc) + if rktMatches[1] != expectedRkt { + fmt.Printf("Test failed; got %q; want %q\n", rktMatches[1], expectedRkt) + } + if appcMatches[1] != expectedAppc { + fmt.Printf("Test failed; got %q; want %q\n", appcMatches[1], expectedAppc) + } +} + +// Tests setting driver config options +func TestRktDriver_SetConfig(t *testing.T) { + t.Parallel() + require := require.New(t) + + d := NewRktDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + // Enable Volumes + config := &Config{ + VolumesEnabled: true, + } + + var data []byte + require.NoError(basePlug.MsgPackEncode(&data, config)) + require.NoError(harness.SetConfig(data)) + require.Exactly(config, d.(*RktDriver).config) + + config.VolumesEnabled = false + data = []byte{} + require.NoError(basePlug.MsgPackEncode(&data, config)) + require.NoError(harness.SetConfig(data)) + require.Exactly(config, d.(*RktDriver).config) + +} + +// Verifies using a trust prefix and passing dns servers and search domains +// Also verifies sending sigterm correctly stops the driver instance +func TestRktDriver_Start_Wait_Stop_DNS(t *testing.T) { + ctestutil.RktCompatible(t) + if !testutil.IsTravis() { + t.Parallel() + } + + require := require.New(t) + d := NewRktDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + Name: "etcd", + Resources: &drivers.Resources{ + NomadResources: &structs.Resources{ + MemoryMB: 128, + CPU: 100, + }, + LinuxResources: &drivers.LinuxResources{ + MemoryLimitBytes: 134217728, + CPUShares: 100, + }, + }, + } + + taskConfig := map[string]interface{}{ + "trust_prefix": "coreos.com/etcd", + "image": "coreos.com/etcd:v2.0.4", + "command": "/etcd", + "dns_servers": []string{"8.8.8.8", "8.8.4.4"}, + "dns_search_domains": []string{"example.com", "example.org", "example.net"}, + "net": []string{"host"}, + } + + encodeDriverHelper(require, task, taskConfig) + testtask.SetTaskConfigEnv(task) + cleanup := harness.MkAllocDir(task, true) + defer cleanup() + + handle, driverNet, err := harness.StartTask(task) + require.NoError(err) + require.Nil(driverNet) + + // Wait for task to start + ch, err := harness.WaitTask(context.Background(), handle.Config.ID) + require.NoError(err) + + var wg sync.WaitGroup + wg.Add(1) + + // Block on the channel returned by wait task + go func() { + defer wg.Done() + result := <-ch + require.Equal(15, result.Signal) + }() + + // Wait until task started + require.NoError(harness.WaitUntilStarted(task.ID, 1*time.Second)) + + // Add to the wait group + wg.Add(1) + + // Stop the task + go func() { + defer wg.Done() + err := harness.StopTask(task.ID, 1*time.Second, "SIGTERM") + require.NoError(err) + }() + + // Wait on the wait group + waitCh := make(chan struct{}) + go func() { + defer close(waitCh) + wg.Wait() + }() + + // Verify that the task exited + select { + case <-waitCh: + status, err := harness.InspectTask(task.ID) + require.NoError(err) + require.Equal(drivers.TaskStateExited, status.State) + case <-time.After(2 * time.Second): + require.Fail("timeout waiting for task to shutdown") + } + + require.NoError(harness.DestroyTask(task.ID, true)) + +} + +// Verifies waiting on task to exit cleanly +func TestRktDriver_Start_Wait_Stop(t *testing.T) { + ctestutil.RktCompatible(t) + if !testutil.IsTravis() { + t.Parallel() + } + + require := require.New(t) + d := NewRktDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + Name: "etcd", + Resources: &drivers.Resources{ + NomadResources: &structs.Resources{ + MemoryMB: 128, + CPU: 100, + }, + LinuxResources: &drivers.LinuxResources{ + MemoryLimitBytes: 134217728, + CPUShares: 100, + }, + }, + } + + taskConfig := map[string]interface{}{ + "trust_prefix": "coreos.com/etcd", + "image": "coreos.com/etcd:v2.0.4", + "command": "/etcd", + "args": []string{"--version"}, + "net": []string{"none"}, + "debug": true, + } + + encodeDriverHelper(require, task, taskConfig) + cleanup := harness.MkAllocDir(task, true) + defer cleanup() + + handle, _, err := harness.StartTask(task) + require.NoError(err) + + // Wait on the task, it should exit since we are only asking for etcd version here + ch, err := harness.WaitTask(context.Background(), handle.Config.ID) + require.NoError(err) + result := <-ch + require.Nil(result.Err) + + require.Zero(result.ExitCode) + + require.NoError(harness.DestroyTask(task.ID, true)) + +} + +// Verifies that skipping trust_prefix works +func TestRktDriver_Start_Wait_Skip_Trust(t *testing.T) { + ctestutil.RktCompatible(t) + if !testutil.IsTravis() { + t.Parallel() + } + + require := require.New(t) + d := NewRktDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + Name: "etcd", + Resources: &drivers.Resources{ + NomadResources: &structs.Resources{ + MemoryMB: 128, + CPU: 100, + }, + LinuxResources: &drivers.LinuxResources{ + MemoryLimitBytes: 134217728, + CPUShares: 100, + }, + }, + } + + taskConfig := map[string]interface{}{ + "image": "coreos.com/etcd:v2.0.4", + "command": "/etcd", + "args": []string{"--version"}, + "net": []string{"none"}, + "debug": true, + } + + encodeDriverHelper(require, task, taskConfig) + testtask.SetTaskConfigEnv(task) + + cleanup := harness.MkAllocDir(task, true) + defer cleanup() + + handle, _, err := harness.StartTask(task) + require.NoError(err) + + // Wait on the task, it should exit since we are only asking for etcd version here + ch, err := harness.WaitTask(context.Background(), handle.Config.ID) + require.NoError(err) + result := <-ch + require.Nil(result.Err) + require.Zero(result.ExitCode) + + require.NoError(harness.DestroyTask(task.ID, true)) + +} + +// Verifies that an invalid trust prefix returns expected error +func TestRktDriver_InvalidTrustPrefix(t *testing.T) { + ctestutil.RktCompatible(t) + if !testutil.IsTravis() { + t.Parallel() + } + + require := require.New(t) + d := NewRktDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + Name: "etcd", + Resources: &drivers.Resources{ + NomadResources: &structs.Resources{ + MemoryMB: 128, + CPU: 100, + }, + LinuxResources: &drivers.LinuxResources{ + MemoryLimitBytes: 134217728, + CPUShares: 100, + }, + }, + } + + taskConfig := map[string]interface{}{ + "trust_prefix": "example.com/invalid", + "image": "coreos.com/etcd:v2.0.4", + "command": "/etcd", + "args": []string{"--version"}, + "net": []string{"none"}, + "debug": true, + } + + encodeDriverHelper(require, task, taskConfig) + testtask.SetTaskConfigEnv(task) + + cleanup := harness.MkAllocDir(task, true) + defer cleanup() + + _, _, err := harness.StartTask(task) + require.Error(err) + expectedErr := "Error running rkt trust" + require.Contains(err.Error(), expectedErr) + +} + +// Verifies reattaching to a running container +// This test manipulates the harness's internal state map +// to remove the task and then reattaches to it +func TestRktDriver_StartWaitRecoverWaitStop(t *testing.T) { + ctestutil.RktCompatible(t) + if !testutil.IsTravis() { + t.Parallel() + } + + require := require.New(t) + d := NewRktDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + Name: "etcd", + Resources: &drivers.Resources{ + NomadResources: &structs.Resources{ + MemoryMB: 128, + CPU: 100, + }, + LinuxResources: &drivers.LinuxResources{ + MemoryLimitBytes: 134217728, + CPUShares: 100, + }, + }, + } + + taskConfig := map[string]interface{}{ + "image": "coreos.com/etcd:v2.0.4", + "command": "/etcd", + } + + encodeDriverHelper(require, task, taskConfig) + + cleanup := harness.MkAllocDir(task, true) + defer cleanup() + + handle, _, err := harness.StartTask(task) + require.NoError(err) + + ch, err := harness.WaitTask(context.Background(), task.ID) + require.NoError(err) + + var waitDone bool + var wg sync.WaitGroup + wg.Add(1) + go func() { + defer wg.Done() + result := <-ch + require.Error(result.Err) + waitDone = true + }() + + originalStatus, err := d.InspectTask(task.ID) + require.NoError(err) + + d.(*RktDriver).tasks.Delete(task.ID) + + wg.Wait() + require.True(waitDone) + _, err = d.InspectTask(task.ID) + require.Equal(drivers.ErrTaskNotFound, err) + + err = d.RecoverTask(handle) + require.NoError(err) + + status, err := d.InspectTask(task.ID) + require.NoError(err) + require.Exactly(originalStatus, status) + + ch, err = harness.WaitTask(context.Background(), task.ID) + require.NoError(err) + + wg.Add(1) + waitDone = false + go func() { + defer wg.Done() + result := <-ch + require.NoError(result.Err) + require.NotZero(result.ExitCode) + require.Equal(9, result.Signal) + waitDone = true + }() + + time.Sleep(300 * time.Millisecond) + require.NoError(d.StopTask(task.ID, 0, "SIGKILL")) + wg.Wait() + require.NoError(d.DestroyTask(task.ID, false)) + require.True(waitDone) + +} + +// Verifies mounting a volume from the host machine and writing +// some data to it from inside the container +func TestRktDriver_Start_Wait_Volume(t *testing.T) { + ctestutil.RktCompatible(t) + if !testutil.IsTravis() { + t.Parallel() + } + + require := require.New(t) + d := NewRktDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + // enable volumes + config := &Config{VolumesEnabled: true} + + var data []byte + require.NoError(basePlug.MsgPackEncode(&data, config)) + require.NoError(harness.SetConfig(data)) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + Name: "rkttest_alpine", + Resources: &drivers.Resources{ + NomadResources: &structs.Resources{ + MemoryMB: 128, + CPU: 100, + }, + LinuxResources: &drivers.LinuxResources{ + MemoryLimitBytes: 134217728, + CPUShares: 100, + }, + }, + } + exp := []byte{'w', 'i', 'n'} + file := "output.txt" + tmpvol, err := ioutil.TempDir("", "nomadtest_rktdriver_volumes") + if err != nil { + t.Fatalf("error creating temporary dir: %v", err) + } + defer os.RemoveAll(tmpvol) + hostpath := filepath.Join(tmpvol, file) + + taskConfig := map[string]interface{}{ + "image": "docker://redis:3.2-alpine", + "command": "/bin/sh", + "args": []string{ + "-c", + fmt.Sprintf("echo -n %s > /foo/%s", string(exp), file), + }, + "net": []string{"none"}, + "volumes": []string{fmt.Sprintf("%s:/foo", tmpvol)}, + } + + encodeDriverHelper(require, task, taskConfig) + testtask.SetTaskConfigEnv(task) + + cleanup := harness.MkAllocDir(task, true) + defer cleanup() + + _, _, err = harness.StartTask(task) + require.NoError(err) + + // Task should terminate quickly + waitCh, err := harness.WaitTask(context.Background(), task.ID) + require.NoError(err) + + select { + case res := <-waitCh: + require.NoError(res.Err) + require.True(res.Successful(), fmt.Sprintf("exit code %v", res.ExitCode)) + case <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second): + require.Fail("WaitTask timeout") + } + + // Check that data was written to the shared alloc directory. + act, err := ioutil.ReadFile(hostpath) + require.NoError(err) + require.Exactly(exp, act) + require.NoError(harness.DestroyTask(task.ID, true)) +} + +// Verifies port mapping +func TestRktDriver_PortMapping(t *testing.T) { + ctestutil.RktCompatible(t) + + require := require.New(t) + d := NewRktDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + Name: "redis", + Resources: &drivers.Resources{ + NomadResources: &structs.Resources{ + MemoryMB: 128, + CPU: 100, + Networks: []*structs.NetworkResource{ + { + IP: "127.0.0.1", + ReservedPorts: []structs.Port{{Label: "main", Value: 8080}}, + }, + }, + }, + LinuxResources: &drivers.LinuxResources{ + MemoryLimitBytes: 134217728, + CPUShares: 100, + }, + }, + } + + taskConfig := map[string]interface{}{ + "image": "docker://redis:3.2-alpine", + "port_map": map[string]string{ + "main": "6379-tcp", + }, + "debug": "true", + } + + encodeDriverHelper(require, task, taskConfig) + + cleanup := harness.MkAllocDir(task, true) + defer cleanup() + + _, driverNetwork, err := harness.StartTask(task) + require.NoError(err) + require.NotNil(driverNetwork) + require.NoError(harness.DestroyTask(task.ID, true)) +} + +// This test starts a redis container, setting user and group. +// It verifies that running ps inside the container shows the expected user and group +func TestRktDriver_UserGroup(t *testing.T) { + ctestutil.RktCompatible(t) + if !testutil.IsTravis() { + t.Parallel() + } + + require := require.New(t) + d := NewRktDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + User: "nobody", + Name: "rkttest_alpine", + Resources: &drivers.Resources{ + NomadResources: &structs.Resources{ + MemoryMB: 128, + CPU: 100, + }, + LinuxResources: &drivers.LinuxResources{ + MemoryLimitBytes: 134217728, + CPUShares: 100, + }, + }, + } + + taskConfig := map[string]interface{}{ + "image": "docker://redis:3.2-alpine", + "group": "nogroup", + "command": "sleep", + "args": []string{"9000"}, + "net": []string{"none"}, + } + + encodeDriverHelper(require, task, taskConfig) + testtask.SetTaskConfigEnv(task) + + cleanup := harness.MkAllocDir(task, true) + defer cleanup() + + _, _, err := harness.StartTask(task) + require.NoError(err) + + expected := []byte("\nnobody nogroup /bin/sleep 9000\n") + testutil.WaitForResult(func() (bool, error) { + res, err := d.ExecTask(task.ID, []string{"ps", "-o", "user,group,args"}, time.Second) + require.NoError(err) + require.Zero(res.ExitResult.ExitCode) + require.True(res.ExitResult.Successful()) + raw := res.Stdout + return bytes.Contains(raw, expected), fmt.Errorf("expected %q but found:\n%s", expected, raw) + }, func(err error) { + t.Fatalf("err: %v", err) + }) + + require.NoError(harness.DestroyTask(task.ID, true)) +} + +// Verifies executing both correct and incorrect commands inside the container +func TestRktDriver_Exec(t *testing.T) { + ctestutil.RktCompatible(t) + if !testutil.IsTravis() { + t.Parallel() + } + + require := require.New(t) + d := NewRktDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + Name: "etcd", + Resources: &drivers.Resources{ + NomadResources: &structs.Resources{ + MemoryMB: 128, + CPU: 100, + }, + LinuxResources: &drivers.LinuxResources{ + MemoryLimitBytes: 134217728, + CPUShares: 100, + }, + }, + } + + taskConfig := map[string]interface{}{ + "trust_prefix": "coreos.com/etcd", + "image": "coreos.com/etcd:v2.0.4", + "net": []string{"none"}, + } + + encodeDriverHelper(require, task, taskConfig) + testtask.SetTaskConfigEnv(task) + + cleanup := harness.MkAllocDir(task, true) + defer cleanup() + + _, _, err := harness.StartTask(task) + require.NoError(err) + + // Run command that should succeed + expected := []byte("etcd version") + testutil.WaitForResult(func() (bool, error) { + res, err := d.ExecTask(task.ID, []string{"/etcd", "--version"}, time.Second) + require.NoError(err) + require.True(res.ExitResult.Successful()) + raw := res.Stdout + return bytes.Contains(raw, expected), fmt.Errorf("expected %q but found:\n%s", expected, raw) + }, func(err error) { + t.Fatalf("err: %v", err) + }) + + // Run command that should fail + expected = []byte("flag provided but not defined") + testutil.WaitForResult(func() (bool, error) { + res, err := d.ExecTask(task.ID, []string{"/etcd", "--cgdfgdfg"}, time.Second) + require.False(res.ExitResult.Successful()) + fmt.Println(err) + fmt.Println(res.ExitResult.ExitCode) + raw := res.Stdout + return bytes.Contains(raw, expected), fmt.Errorf("expected %q but found:\n%s", expected, raw) + }, func(err error) { + t.Fatalf("err: %v", err) + }) + + require.NoError(harness.DestroyTask(task.ID, true)) +} + +// Verifies executing both correct and incorrect commands inside the container +func TestRktDriver_Stats(t *testing.T) { + ctestutil.RktCompatible(t) + if !testutil.IsTravis() { + t.Parallel() + } + + require := require.New(t) + d := NewRktDriver(testlog.HCLogger(t)) + harness := drivers.NewDriverHarness(t, d) + + task := &drivers.TaskConfig{ + ID: uuid.Generate(), + Name: "etcd", + Resources: &drivers.Resources{ + NomadResources: &structs.Resources{ + MemoryMB: 128, + CPU: 100, + }, + LinuxResources: &drivers.LinuxResources{ + MemoryLimitBytes: 134217728, + CPUShares: 100, + }, + }, + } + + taskConfig := map[string]interface{}{ + "trust_prefix": "coreos.com/etcd", + "image": "coreos.com/etcd:v2.0.4", + "net": []string{"none"}, + } + + encodeDriverHelper(require, task, taskConfig) + testtask.SetTaskConfigEnv(task) + + cleanup := harness.MkAllocDir(task, true) + defer cleanup() + + _, _, err := harness.StartTask(task) + require.NoError(err) + + // Run command that should succeed + expected := []byte("etcd version") + testutil.WaitForResult(func() (bool, error) { + res, err := d.ExecTask(task.ID, []string{"/etcd", "--version"}, time.Second) + require.NoError(err) + require.True(res.ExitResult.Successful()) + raw := res.Stdout + return bytes.Contains(raw, expected), fmt.Errorf("expected %q but found:\n%s", expected, raw) + }, func(err error) { + t.Fatalf("err: %v", err) + }) + + // Run command that should fail + expected = []byte("flag provided but not defined") + testutil.WaitForResult(func() (bool, error) { + res, err := d.ExecTask(task.ID, []string{"/etcd", "--cgdfgdfg"}, time.Second) + require.False(res.ExitResult.Successful()) + fmt.Println(err) + fmt.Println(res.ExitResult.ExitCode) + raw := res.Stdout + return bytes.Contains(raw, expected), fmt.Errorf("expected %q but found:\n%s", expected, raw) + }, func(err error) { + t.Fatalf("err: %v", err) + }) + + require.NoError(harness.DestroyTask(task.ID, true)) +} + +func encodeDriverHelper(require *require.Assertions, task *drivers.TaskConfig, taskConfig map[string]interface{}) { + evalCtx := &hcl.EvalContext{ + Functions: shared.GetStdlibFuncs(), + } + spec, diag := hclspec.Convert(taskConfigSpec) + require.False(diag.HasErrors()) + taskConfigCtyVal, diag := shared.ParseHclInterface(taskConfig, spec, evalCtx) + if diag.HasErrors() { + fmt.Println("conversion error", diag.Error()) + } + require.False(diag.HasErrors()) + err := task.EncodeDriverConfig(taskConfigCtyVal) + require.Nil(err) +} diff --git a/drivers/rkt/handle.go b/drivers/rkt/handle.go new file mode 100644 index 000000000..1466ccacb --- /dev/null +++ b/drivers/rkt/handle.go @@ -0,0 +1,63 @@ +// +build linux + +package rkt + +import ( + "sync" + "time" + + hclog "github.com/hashicorp/go-hclog" + plugin "github.com/hashicorp/go-plugin" + "github.com/hashicorp/nomad/client/driver/env" + "github.com/hashicorp/nomad/client/driver/executor" + "github.com/hashicorp/nomad/plugins/drivers" +) + +type rktTaskHandle struct { + exec executor.Executor + env *env.TaskEnv + uuid string + pid int + pluginClient *plugin.Client + logger hclog.Logger + + // stateLock syncs access to all fields below + stateLock sync.RWMutex + + taskConfig *drivers.TaskConfig + procState drivers.TaskState + startedAt time.Time + completedAt time.Time + exitResult *drivers.ExitResult +} + +func (h *rktTaskHandle) IsRunning() bool { + return h.procState == drivers.TaskStateRunning +} + +func (h *rktTaskHandle) run() { + + // since run is called immediately after the handle is created this + // ensures the exitResult is initialized so we avoid a nil pointer + // thus it does not need to be included in the lock + if h.exitResult == nil { + h.exitResult = &drivers.ExitResult{} + } + + ps, err := h.exec.Wait() + h.stateLock.Lock() + defer h.stateLock.Unlock() + + if err != nil { + h.exitResult.Err = err + h.procState = drivers.TaskStateUnknown + h.completedAt = time.Now() + return + } + h.procState = drivers.TaskStateExited + h.exitResult.ExitCode = ps.ExitCode + h.exitResult.Signal = ps.Signal + h.completedAt = ps.Time + + // TODO: detect if the taskConfig OOMed +} diff --git a/drivers/rkt/state.go b/drivers/rkt/state.go index 834154dd4..1146db6c7 100644 --- a/drivers/rkt/state.go +++ b/drivers/rkt/state.go @@ -1,13 +1,9 @@ +// +build linux + package rkt import ( "sync" - "time" - - hclog "github.com/hashicorp/go-hclog" - plugin "github.com/hashicorp/go-plugin" - "github.com/hashicorp/nomad/client/driver/executor" - "github.com/hashicorp/nomad/plugins/drivers" ) type taskStore struct { @@ -37,42 +33,3 @@ func (ts *taskStore) Delete(id string) { defer ts.lock.Unlock() delete(ts.store, id) } - -type rktTaskHandle struct { - exec executor.Executor - pid int - pluginClient *plugin.Client - logger hclog.Logger - - // stateLock syncs access to all fields below - stateLock sync.RWMutex - - task *drivers.TaskConfig - procState drivers.TaskState - startedAt time.Time - completedAt time.Time - exitResult *drivers.ExitResult -} - -func (h *rktTaskHandle) IsRunning() bool { - return h.procState == drivers.TaskStateRunning -} - -func (h *rktTaskHandle) run() { - if h.exitResult == nil { - h.exitResult = &drivers.ExitResult{} - } - - ps, err := h.exec.Wait() - if err != nil { - h.exitResult.Err = err - h.procState = drivers.TaskStateUnknown - h.completedAt = time.Now() - return - } - h.procState = drivers.TaskStateExited - h.exitResult.ExitCode = ps.ExitCode - h.exitResult.Signal = ps.Signal - h.completedAt = ps.Time - -} diff --git a/plugins/drivers/driver.go b/plugins/drivers/driver.go index f64d3dd5b..c50a3042e 100644 --- a/plugins/drivers/driver.go +++ b/plugins/drivers/driver.go @@ -13,7 +13,6 @@ import ( "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/shared/hclspec" "github.com/zclconf/go-cty/cty" - "github.com/zclconf/go-cty/cty/gocty" "github.com/zclconf/go-cty/cty/msgpack" "golang.org/x/net/context" ) @@ -111,6 +110,10 @@ type TaskConfig struct { StderrPath string } +func (tc *TaskConfig) GetRawConfig() []byte { + return tc.rawDriverConfig +} + func (tc *TaskConfig) Copy() *TaskConfig { if tc == nil { return nil @@ -145,17 +148,7 @@ func (tc *TaskConfig) TaskDir() *allocdir.TaskDir { } func (tc *TaskConfig) DecodeDriverConfig(t interface{}) error { - ty, err := gocty.ImpliedType(t) - if err != nil { - return err - } - - val, err := msgpack.Unmarshal(tc.rawDriverConfig, ty) - if err != nil { - return err - } - - return gocty.FromCtyValue(val, t) + return base.MsgPackDecode(tc.rawDriverConfig, t) } func (tc *TaskConfig) EncodeDriverConfig(val cty.Value) error { diff --git a/plugins/drivers/utils.go b/plugins/drivers/utils.go index dd0da0614..293e14c45 100644 --- a/plugins/drivers/utils.go +++ b/plugins/drivers/utils.go @@ -146,10 +146,10 @@ func resourcesToProto(r *Resources) *proto.Resources { Memory: int64(r.NomadResources.MemoryMB), Iops: int64(r.NomadResources.IOPS), Disk: int64(r.NomadResources.DiskMB), - Networks: []*proto.NetworkResource{}, + Networks: make([]*proto.NetworkResource, len(r.NomadResources.Networks)), } - for _, network := range r.NomadResources.Networks { + for i, network := range r.NomadResources.Networks { var n proto.NetworkResource n.Device = network.Device n.Ip = network.IP @@ -168,6 +168,7 @@ func resourcesToProto(r *Resources) *proto.Resources { Value: int32(port.Value), }) } + pb.RawResources.Networks[i] = &n } } From edad5e9e6daed4bcce469e235168cc79aa4e11a5 Mon Sep 17 00:00:00 2001 From: Preetha Appan Date: Fri, 12 Oct 2018 14:20:42 -0500 Subject: [PATCH 132/152] Stats collection test --- drivers/rkt/driver_test.go | 42 +++++++++++++++----------------------- 1 file changed, 17 insertions(+), 25 deletions(-) diff --git a/drivers/rkt/driver_test.go b/drivers/rkt/driver_test.go index faa31a644..82194a2ed 100644 --- a/drivers/rkt/driver_test.go +++ b/drivers/rkt/driver_test.go @@ -673,7 +673,8 @@ func TestRktDriver_Exec(t *testing.T) { require.NoError(harness.DestroyTask(task.ID, true)) } -// Verifies executing both correct and incorrect commands inside the container +// Verifies getting resource usage stats +// TODO(preetha) figure out why stats are zero func TestRktDriver_Stats(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsTravis() { @@ -702,6 +703,7 @@ func TestRktDriver_Stats(t *testing.T) { taskConfig := map[string]interface{}{ "trust_prefix": "coreos.com/etcd", "image": "coreos.com/etcd:v2.0.4", + "command": "/etcd", "net": []string{"none"}, } @@ -711,35 +713,25 @@ func TestRktDriver_Stats(t *testing.T) { cleanup := harness.MkAllocDir(task, true) defer cleanup() - _, _, err := harness.StartTask(task) + handle, _, err := harness.StartTask(task) require.NoError(err) - // Run command that should succeed - expected := []byte("etcd version") - testutil.WaitForResult(func() (bool, error) { - res, err := d.ExecTask(task.ID, []string{"/etcd", "--version"}, time.Second) - require.NoError(err) - require.True(res.ExitResult.Successful()) - raw := res.Stdout - return bytes.Contains(raw, expected), fmt.Errorf("expected %q but found:\n%s", expected, raw) - }, func(err error) { - t.Fatalf("err: %v", err) - }) + // Wait for task to start + _, err = harness.WaitTask(context.Background(), handle.Config.ID) + require.NoError(err) - // Run command that should fail - expected = []byte("flag provided but not defined") - testutil.WaitForResult(func() (bool, error) { - res, err := d.ExecTask(task.ID, []string{"/etcd", "--cgdfgdfg"}, time.Second) - require.False(res.ExitResult.Successful()) - fmt.Println(err) - fmt.Println(res.ExitResult.ExitCode) - raw := res.Stdout - return bytes.Contains(raw, expected), fmt.Errorf("expected %q but found:\n%s", expected, raw) - }, func(err error) { - t.Fatalf("err: %v", err) - }) + // Wait until task started + require.NoError(harness.WaitUntilStarted(task.ID, 1*time.Second)) + + resourceUsage, err := d.TaskStats(task.ID) + require.Nil(err) + + //TODO(preetha) why are these zero + fmt.Printf("pid map %v\n", resourceUsage.Pids) + fmt.Printf("CPU:%+v Memory:%+v", resourceUsage.ResourceUsage.CpuStats, resourceUsage.ResourceUsage.MemoryStats) require.NoError(harness.DestroyTask(task.ID, true)) + } func encodeDriverHelper(require *require.Assertions, task *drivers.TaskConfig, taskConfig map[string]interface{}) { From 4f4777d6a618d43fddfe6138774723ae5b2c4958 Mon Sep 17 00:00:00 2001 From: Preetha Appan Date: Mon, 15 Oct 2018 12:27:14 -0500 Subject: [PATCH 133/152] Review comments --- drivers/rkt/driver.go | 1 + plugins/drivers/driver.go | 4 ---- 2 files changed, 1 insertion(+), 4 deletions(-) diff --git a/drivers/rkt/driver.go b/drivers/rkt/driver.go index 988f9d7b6..85a7fccc1 100644 --- a/drivers/rkt/driver.go +++ b/drivers/rkt/driver.go @@ -218,6 +218,7 @@ func (r *RktDriver) Fingerprint(ctx context.Context) (<-chan *drivers.Fingerprin } func (d *RktDriver) handleFingerprint(ctx context.Context, ch chan *drivers.Fingerprint) { + defer close(ch) ticker := time.NewTimer(0) for { select { diff --git a/plugins/drivers/driver.go b/plugins/drivers/driver.go index c50a3042e..f37a10ff0 100644 --- a/plugins/drivers/driver.go +++ b/plugins/drivers/driver.go @@ -110,10 +110,6 @@ type TaskConfig struct { StderrPath string } -func (tc *TaskConfig) GetRawConfig() []byte { - return tc.rawDriverConfig -} - func (tc *TaskConfig) Copy() *TaskConfig { if tc == nil { return nil From 9b3e82a4e753cab6732e915dfb318426210e6f2c Mon Sep 17 00:00:00 2001 From: Preetha Appan Date: Mon, 15 Oct 2018 13:08:41 -0500 Subject: [PATCH 134/152] make port map a slice of maps to match existing rkt driver --- drivers/rkt/driver.go | 42 +++++++++++++++++++++++--------------- drivers/rkt/driver_test.go | 4 ++-- 2 files changed, 28 insertions(+), 18 deletions(-) diff --git a/drivers/rkt/driver.go b/drivers/rkt/driver.go index 85a7fccc1..2a25f9a6f 100644 --- a/drivers/rkt/driver.go +++ b/drivers/rkt/driver.go @@ -86,7 +86,7 @@ var ( "dns_servers": hclspec.NewAttr("dns_servers", "list(string)", false), "dns_search_domains": hclspec.NewAttr("dns_search_domains", "list(string)", false), "net": hclspec.NewAttr("net", "list(string)", false), - "port_map": hclspec.NewAttr("port_map", "map(string)", false), + "port_map": hclspec.NewAttr("port_map", "list(map(string))", false), "volumes": hclspec.NewAttr("volumes", "list(string)", false), "insecure_options": hclspec.NewAttr("insecure_options", "list(string)", false), "no_overlay": hclspec.NewAttr("no_overlay", "bool", false), @@ -118,16 +118,16 @@ type Config struct { // TaskConfig is the driver configuration of a taskConfig within a job type TaskConfig struct { - ImageName string `codec:"image" cty:"image"` - Command string `codec:"command" cty:"command"` - Args []string `codec:"args" cty:"args"` - TrustPrefix string `codec:"trust_prefix" cty:"trust_prefix"` - DNSServers []string `codec:"dns_servers" cty:"dns_servers"` // DNS Server for containers - DNSSearchDomains []string `codec:"dns_search_domains" cty:"dns_search_domains"` // DNS Search domains for containers - Net []string `codec:"net" cty:"net"` // Networks for the containers - PortMap map[string]string `codec:"port_map" cty:"port_map"` // A map of host port and the port name defined in the image manifest file - Volumes []string `codec:"volumes" cty:"volumes"` // Host-Volumes to mount in, syntax: /path/to/host/directory:/destination/path/in/container[:readOnly] - InsecureOptions []string `codec:"insecure_options" cty:"insecure_options"` // list of args for --insecure-options + ImageName string `codec:"image" cty:"image"` + Command string `codec:"command" cty:"command"` + Args []string `codec:"args" cty:"args"` + TrustPrefix string `codec:"trust_prefix" cty:"trust_prefix"` + DNSServers []string `codec:"dns_servers" cty:"dns_servers"` // DNS Server for containers + DNSSearchDomains []string `codec:"dns_search_domains" cty:"dns_search_domains"` // DNS Search domains for containers + Net []string `codec:"net" cty:"net"` // Networks for the containers + PortMap []map[string]string `codec:"port_map" cty:"port_map"` // A map of host port and the port name defined in the image manifest file + Volumes []string `codec:"volumes" cty:"volumes"` // Host-Volumes to mount in, syntax: /path/to/host/directory:/destination/path/in/container[:readOnly] + InsecureOptions []string `codec:"insecure_options" cty:"insecure_options"` // list of args for --insecure-options NoOverlay bool `codec:"no_overlay" cty:"no_overlay"` // disable overlayfs for rkt run Debug bool `codec:"debug" cty:"debug"` // Enable debug option for rkt command @@ -344,8 +344,6 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs } var driverConfig TaskConfig - d.logger.Debug("msgpack encoding", "data", string(cfg.GetRawConfig())) - if err := cfg.DecodeDriverConfig(&driverConfig); err != nil { return nil, nil, fmt.Errorf("failed to decode driver config: %v", err) } @@ -497,6 +495,8 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs runArgs = append(runArgs, fmt.Sprintf("--net=%s", network)) } + mergedPortMap := mapMergeStrStr(driverConfig.PortMap...) + // Setup port mapping and exposed ports if len(cfg.Resources.NomadResources.Networks) == 0 { d.logger.Debug("no network interfaces are available") @@ -511,7 +511,7 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs for _, port := range network.ReservedPorts { var containerPort string - mapped, ok := driverConfig.PortMap[port.Label] + mapped, ok := mergedPortMap[port.Label] if !ok { // If the user doesn't have a mapped port using port_map, driver stops running container. return nil, nil, fmt.Errorf("port_map is not set. When you defined port in the resources, you need to configure port_map.") @@ -529,7 +529,7 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs // By default we will map the allocated port 1:1 to the container var containerPort string - if mapped, ok := driverConfig.PortMap[port.Label]; ok { + if mapped, ok := mergedPortMap[port.Label]; ok { containerPort = mapped } else { // If the user doesn't have mapped a port using port_map, driver stops running container. @@ -666,7 +666,7 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs var driverNetwork *cstructs.DriverNetwork if network != "host" && network != "none" { d.logger.Debug("retrieving network information for pod", "pod", img, "UUID", uuid, "task", cfg.Name) - driverNetwork, err = rktGetDriverNetwork(uuid, driverConfig.PortMap, d.logger) + driverNetwork, err = rktGetDriverNetwork(uuid, mergedPortMap, d.logger) if err != nil && !pluginClient.Exited() { d.logger.Warn("network status retrieval for pod failed", "pod", img, "UUID", uuid, "task", cfg.Name, "error", err) @@ -995,3 +995,13 @@ func (d *RktDriver) handleWait(ctx context.Context, handle *rktTaskHandle, ch ch case ch <- result: } } + +func mapMergeStrStr(maps ...map[string]string) map[string]string { + out := map[string]string{} + for _, in := range maps { + for key, val := range in { + out[key] = val + } + } + return out +} diff --git a/drivers/rkt/driver_test.go b/drivers/rkt/driver_test.go index 82194a2ed..418efef70 100644 --- a/drivers/rkt/driver_test.go +++ b/drivers/rkt/driver_test.go @@ -527,9 +527,9 @@ func TestRktDriver_PortMapping(t *testing.T) { taskConfig := map[string]interface{}{ "image": "docker://redis:3.2-alpine", - "port_map": map[string]string{ + "port_map": []map[string]string{{ "main": "6379-tcp", - }, + }}, "debug": "true", } From f12501d4c376ed93416762700c0b86e72baac36f Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Wed, 10 Oct 2018 12:31:59 -0700 Subject: [PATCH 135/152] tr: implement dispatch payload hook Now passing the TaskDir struct to prestart hooks instead of just the root task dir itself as dispatch needs local/. --- .../allocrunner/interfaces/task_lifecycle.go | 5 +- .../allocrunner/taskrunner/artifact_hook.go | 2 +- .../allocrunner/taskrunner/dispatch_hook.go | 71 +++++++++ .../taskrunner/dispatch_hook_test.go | 144 ++++++++++++++++++ .../taskrunner/task_runner_hooks.go | 5 +- .../allocrunner/taskrunner/template_hook.go | 2 +- client/allocrunner/taskrunner/vault_hook.go | 3 +- 7 files changed, 224 insertions(+), 8 deletions(-) create mode 100644 client/allocrunner/taskrunner/dispatch_hook.go create mode 100644 client/allocrunner/taskrunner/dispatch_hook_test.go diff --git a/client/allocrunner/interfaces/task_lifecycle.go b/client/allocrunner/interfaces/task_lifecycle.go index 6b6314a69..b22ef285b 100644 --- a/client/allocrunner/interfaces/task_lifecycle.go +++ b/client/allocrunner/interfaces/task_lifecycle.go @@ -3,6 +3,7 @@ package interfaces import ( "context" + "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/client/driver/env" cstructs "github.com/hashicorp/nomad/client/structs" @@ -49,8 +50,8 @@ type TaskPrestartRequest struct { // Vault token may optionally be set if a Vault token is available VaultToken string - // TaskDir is the task's directory on the host - TaskDir string + // TaskDir contains the task's directory tree on the host + TaskDir *allocdir.TaskDir // TaskEnv is the task's environment TaskEnv *env.TaskEnv diff --git a/client/allocrunner/taskrunner/artifact_hook.go b/client/allocrunner/taskrunner/artifact_hook.go index 6860a7a8b..d6cfbfe27 100644 --- a/client/allocrunner/taskrunner/artifact_hook.go +++ b/client/allocrunner/taskrunner/artifact_hook.go @@ -39,7 +39,7 @@ func (h *artifactHook) Prestart(ctx context.Context, req *interfaces.TaskPrestar for _, artifact := range req.Task.Artifacts { //XXX add ctx to GetArtifact to allow cancelling long downloads - if err := getter.GetArtifact(req.TaskEnv, artifact, req.TaskDir); err != nil { + if err := getter.GetArtifact(req.TaskEnv, artifact, req.TaskDir.Dir); err != nil { wrapped := fmt.Errorf("failed to download artifact %q: %v", artifact.GetterSource, err) h.logger.Debug(wrapped.Error()) h.eventEmitter.EmitEvent(structs.NewTaskEvent(structs.TaskArtifactDownloadFailed).SetDownloadError(wrapped)) diff --git a/client/allocrunner/taskrunner/dispatch_hook.go b/client/allocrunner/taskrunner/dispatch_hook.go new file mode 100644 index 000000000..25a8f09f2 --- /dev/null +++ b/client/allocrunner/taskrunner/dispatch_hook.go @@ -0,0 +1,71 @@ +package taskrunner + +import ( + "context" + "io/ioutil" + "os" + "path/filepath" + + "github.com/golang/snappy" + hclog "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" + "github.com/hashicorp/nomad/nomad/structs" +) + +// dispatchHook writes a dispatch payload to the task dir +type dispatchHook struct { + payload []byte + + logger hclog.Logger +} + +func newDispatchHook(alloc *structs.Allocation, logger hclog.Logger) *dispatchHook { + h := &dispatchHook{ + payload: alloc.Job.Payload, + } + h.logger = logger.Named(h.Name()) + return h +} + +func (*dispatchHook) Name() string { + return "dispatch_payload" +} + +func (h *dispatchHook) Prestart(ctx context.Context, req *interfaces.TaskPrestartRequest, resp *interfaces.TaskPrestartResponse) error { + if len(h.payload) == 0 || req.Task.DispatchPayload == nil || req.Task.DispatchPayload.File == "" { + // No dispatch payload + resp.Done = true + return nil + } + + err := writeDispatchPayload(req.TaskDir.LocalDir, req.Task.DispatchPayload.File, h.payload) + if err != nil { + return err + } + + h.logger.Trace("dispatch payload written", + "path", req.TaskDir.LocalDir, + "filename", req.Task.DispatchPayload.File, + "bytes", len(h.payload), + ) + + // Dispatch payload written successfully; mark as done + resp.Done = true + return nil +} + +// writeDispatchPayload writes the payload to the given file or returns an +// error. +func writeDispatchPayload(base, filename string, payload []byte) error { + renderTo := filepath.Join(base, filename) + decoded, err := snappy.Decode(nil, payload) + if err != nil { + return err + } + + if err := os.MkdirAll(filepath.Dir(renderTo), 0777); err != nil { + return err + } + + return ioutil.WriteFile(renderTo, decoded, 0777) +} diff --git a/client/allocrunner/taskrunner/dispatch_hook_test.go b/client/allocrunner/taskrunner/dispatch_hook_test.go new file mode 100644 index 000000000..ee7eb7d0f --- /dev/null +++ b/client/allocrunner/taskrunner/dispatch_hook_test.go @@ -0,0 +1,144 @@ +package taskrunner + +import ( + "context" + "io/ioutil" + "path/filepath" + "testing" + + "github.com/golang/snappy" + "github.com/hashicorp/nomad/client/allocdir" + "github.com/hashicorp/nomad/client/allocrunner/interfaces" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/helper/testlog" + "github.com/hashicorp/nomad/nomad/mock" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/stretchr/testify/require" +) + +// Statically assert the stats hook implements the expected interfaces +var _ interfaces.TaskPrestartHook = (*dispatchHook)(nil) + +// TestTaskRunner_DispatchHook_NoPayload asserts that the hook is a noop and is +// marked as done if there is no dispatch payload. +func TestTaskRunner_DispatchHook_NoPayload(t *testing.T) { + t.Parallel() + + require := require.New(t) + ctx := context.Background() + logger := testlog.HCLogger(t) + allocDir := allocdir.NewAllocDir(logger, "nomadtest_nopayload") + defer allocDir.Destroy() + + // Default mock alloc/job is not a dispatch job + alloc := mock.BatchAlloc() + task := alloc.Job.TaskGroups[0].Tasks[0] + taskDir := allocDir.NewTaskDir(task.Name) + require.NoError(taskDir.Build(false, nil, cstructs.FSIsolationNone)) + + h := newDispatchHook(alloc, logger) + + req := interfaces.TaskPrestartRequest{ + Task: task, + TaskDir: taskDir, + } + resp := interfaces.TaskPrestartResponse{} + + // Assert no error and Done=true as this job has no payload + require.NoError(h.Prestart(ctx, &req, &resp)) + require.True(resp.Done) + + // Assert payload directory is empty + files, err := ioutil.ReadDir(req.TaskDir.LocalDir) + require.NoError(err) + require.Empty(files) +} + +// TestTaskRunner_DispatchHook_Ok asserts that dispatch payloads are written to +// a file in the task dir. +func TestTaskRunner_DispatchHook_Ok(t *testing.T) { + t.Parallel() + + require := require.New(t) + ctx := context.Background() + logger := testlog.HCLogger(t) + allocDir := allocdir.NewAllocDir(logger, "nomadtest_dispatchok") + defer allocDir.Destroy() + + // Default mock alloc/job is not a dispatch job; update it + alloc := mock.BatchAlloc() + alloc.Job.ParameterizedJob = &structs.ParameterizedJobConfig{ + Payload: structs.DispatchPayloadRequired, + } + expected := []byte("hello world") + alloc.Job.Payload = snappy.Encode(nil, expected) + + // Set the filename and create the task dir + task := alloc.Job.TaskGroups[0].Tasks[0] + task.DispatchPayload = &structs.DispatchPayloadConfig{ + File: "out", + } + taskDir := allocDir.NewTaskDir(task.Name) + require.NoError(taskDir.Build(false, nil, cstructs.FSIsolationNone)) + + h := newDispatchHook(alloc, logger) + + req := interfaces.TaskPrestartRequest{ + Task: task, + TaskDir: taskDir, + } + resp := interfaces.TaskPrestartResponse{} + require.NoError(h.Prestart(ctx, &req, &resp)) + require.True(resp.Done) + + filename := filepath.Join(req.TaskDir.LocalDir, task.DispatchPayload.File) + result, err := ioutil.ReadFile(filename) + require.NoError(err) + require.Equal(expected, result) +} + +// TestTaskRunner_DispatchHook_Error asserts that on an error dispatch payloads +// are not written and Done=false. +func TestTaskRunner_DispatchHook_Error(t *testing.T) { + t.Parallel() + + require := require.New(t) + ctx := context.Background() + logger := testlog.HCLogger(t) + allocDir := allocdir.NewAllocDir(logger, "nomadtest_dispatcherr") + defer allocDir.Destroy() + + // Default mock alloc/job is not a dispatch job; update it + alloc := mock.BatchAlloc() + alloc.Job.ParameterizedJob = &structs.ParameterizedJobConfig{ + Payload: structs.DispatchPayloadRequired, + } + + // Cause an error by not snappy encoding the payload + alloc.Job.Payload = []byte("hello world") + + // Set the filename and create the task dir + task := alloc.Job.TaskGroups[0].Tasks[0] + task.DispatchPayload = &structs.DispatchPayloadConfig{ + File: "out", + } + taskDir := allocDir.NewTaskDir(task.Name) + require.NoError(taskDir.Build(false, nil, cstructs.FSIsolationNone)) + + h := newDispatchHook(alloc, logger) + + req := interfaces.TaskPrestartRequest{ + Task: task, + TaskDir: taskDir, + } + resp := interfaces.TaskPrestartResponse{} + + // Assert an error was returned and Done=false + require.Error(h.Prestart(ctx, &req, &resp)) + require.False(resp.Done) + + // Assert payload directory is empty + files, err := ioutil.ReadDir(req.TaskDir.LocalDir) + require.NoError(err) + require.Empty(files) +} diff --git a/client/allocrunner/taskrunner/task_runner_hooks.go b/client/allocrunner/taskrunner/task_runner_hooks.go index 6ccb808a4..c785a0245 100644 --- a/client/allocrunner/taskrunner/task_runner_hooks.go +++ b/client/allocrunner/taskrunner/task_runner_hooks.go @@ -24,6 +24,7 @@ func (tr *TaskRunner) initHooks() { newValidateHook(tr.clientConfig, hookLogger), newTaskDirHook(tr, hookLogger), newLogMonHook(tr.logmonHookConfig, hookLogger), + newDispatchHook(tr.Alloc(), hookLogger), newArtifactHook(tr, hookLogger), newShutdownDelayHook(task.ShutdownDelay, hookLogger), newStatsHook(tr, tr.clientConfig.StatsCollectionInterval, hookLogger), @@ -97,7 +98,7 @@ func (tr *TaskRunner) prestart() error { // Build the request req := interfaces.TaskPrestartRequest{ Task: tr.Task(), - TaskDir: tr.taskDir.Dir, + TaskDir: tr.taskDir, TaskEnv: tr.envBuilder.Build(), } @@ -150,7 +151,7 @@ func (tr *TaskRunner) prestart() error { if tr.logger.IsTrace() { end := time.Now() - tr.logger.Trace("finished prestart hooks", "name", name, "end", end, "duration", end.Sub(start)) + tr.logger.Trace("finished prestart hook", "name", name, "end", end, "duration", end.Sub(start)) } } diff --git a/client/allocrunner/taskrunner/template_hook.go b/client/allocrunner/taskrunner/template_hook.go index 3b13443fc..bb920755a 100644 --- a/client/allocrunner/taskrunner/template_hook.go +++ b/client/allocrunner/taskrunner/template_hook.go @@ -73,7 +73,7 @@ func (h *templateHook) Prestart(ctx context.Context, req *interfaces.TaskPrestar } // Store the current Vault token and the task directory - h.taskDir = req.TaskDir + h.taskDir = req.TaskDir.Dir h.vaultToken = req.VaultToken unblockCh, err := h.newManager() if err != nil { diff --git a/client/allocrunner/taskrunner/vault_hook.go b/client/allocrunner/taskrunner/vault_hook.go index d06ecbd81..bb44b0c29 100644 --- a/client/allocrunner/taskrunner/vault_hook.go +++ b/client/allocrunner/taskrunner/vault_hook.go @@ -12,7 +12,6 @@ import ( "github.com/hashicorp/consul-template/signals" log "github.com/hashicorp/go-hclog" - "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner/interfaces" ti "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/client/vaultclient" @@ -130,7 +129,7 @@ func (h *vaultHook) Prestart(ctx context.Context, req *interfaces.TaskPrestartRe // Try to recover a token if it was previously written in the secrets // directory recoveredToken := "" - h.tokenPath = filepath.Join(req.TaskDir, allocdir.TaskSecrets, vaultTokenFile) + h.tokenPath = filepath.Join(req.TaskDir.SecretsDir, vaultTokenFile) data, err := ioutil.ReadFile(h.tokenPath) if err != nil { if !os.IsNotExist(err) { From 528c426c5396b3ea79e180a501108eab38923d05 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 9 Oct 2018 17:27:51 -0700 Subject: [PATCH 136/152] Port client portion of #4392 to new taskrunner PR #4392 was merged to master *after* allocrunnerv2 was branched, so the client-specific portions must be ported from master to arv2. --- client/allocrunner/taskrunner/task_runner.go | 20 ++++++++++++++++++++ 1 file changed, 20 insertions(+) diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index a9d10dc19..c3a49c94a 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -4,6 +4,7 @@ import ( "context" "errors" "fmt" + "strings" "sync" "time" @@ -277,6 +278,25 @@ func (tr *TaskRunner) initLabels() { Value: tr.taskName, }, } + + if tr.alloc.Job.ParentID != "" { + tr.baseLabels = append(tr.baseLabels, metrics.Label{ + Name: "parent_id", + Value: tr.alloc.Job.ParentID, + }) + if strings.Contains(tr.alloc.Job.Name, "/dispatch-") { + tr.baseLabels = append(tr.baseLabels, metrics.Label{ + Name: "dispatch_id", + Value: strings.Split(tr.alloc.Job.Name, "/dispatch-")[1], + }) + } + if strings.Contains(tr.alloc.Job.Name, "/periodic-") { + tr.baseLabels = append(tr.baseLabels, metrics.Label{ + Name: "periodic_id", + Value: strings.Split(tr.alloc.Job.Name, "/periodic-")[1], + }) + } + } } func (tr *TaskRunner) Run() { From 9f564781ce5d5d0aa857a6f5939572a564d428cb Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Fri, 12 Oct 2018 14:31:33 -0400 Subject: [PATCH 137/152] drivers/mock: start mock driver implementation --- drivers/mock/driver.go | 370 +++++++++++++++++++++++++++++++++++++++++ drivers/mock/handle.go | 88 ++++++++++ drivers/mock/state.go | 33 ++++ 3 files changed, 491 insertions(+) create mode 100644 drivers/mock/driver.go create mode 100644 drivers/mock/handle.go create mode 100644 drivers/mock/state.go diff --git a/drivers/mock/driver.go b/drivers/mock/driver.go new file mode 100644 index 000000000..141049865 --- /dev/null +++ b/drivers/mock/driver.go @@ -0,0 +1,370 @@ +package mock + +import ( + "context" + "errors" + "fmt" + "strconv" + "strings" + "time" + + hclog "github.com/hashicorp/go-hclog" + cstructs "github.com/hashicorp/nomad/client/structs" + "github.com/hashicorp/nomad/drivers/shared/eventer" + "github.com/hashicorp/nomad/nomad/structs" + "github.com/hashicorp/nomad/plugins/base" + "github.com/hashicorp/nomad/plugins/drivers" + "github.com/hashicorp/nomad/plugins/shared/hclspec" +) + +const ( + // pluginName is the name of the plugin + pluginName = "mock" + + // fingerprintPeriod is the interval at which the driver will send fingerprint responses + fingerprintPeriod = 30 * time.Second +) + +var ( + // pluginInfo is the response returned for the PluginInfo RPC + pluginInfo = &base.PluginInfoResponse{ + Type: base.PluginTypeDriver, + PluginApiVersion: "0.0.1", + PluginVersion: "0.1.0", + Name: pluginName, + } + + // configSpec is the hcl specification returned by the ConfigSchema RPC + configSpec = hclspec.NewObject(map[string]*hclspec.Spec{ + "shutdown_periodic_after": hclspec.NewDefault( + hclspec.NewAttr("shutdown_periodic_after", "bool", false), + hclspec.NewLiteral("false"), + ), + "shutdown_periodic_duration": hclspec.NewAttr("shutdown_periodic_duration", "number", false), + }) + + // taskConfigSpec is the hcl specification for the driver config section of + // a task within a job. It is returned in the TaskConfigSchema RPC + taskConfigSpec = hclspec.NewObject(map[string]*hclspec.Spec{ + "start_error": hclspec.NewAttr("start_error", "string", false), + "start_error_recoverable": hclspec.NewAttr("start_error_recoverable", "bool", false), + "start_block_for": hclspec.NewAttr("start_block_for", "number", false), + "kill_after": hclspec.NewAttr("kill_after", "number", false), + "run_for": hclspec.NewAttr("run_for", "number", false), + "exit_code": hclspec.NewAttr("exit_code", "number", false), + "exit_signal": hclspec.NewAttr("exit_signal", "number", false), + "exit_err_msg": hclspec.NewAttr("exit_err_msg", "string", false), + "signal_err": hclspec.NewAttr("signal_err", "string", false), + "driver_ip": hclspec.NewAttr("driver_ip", "string", false), + "driver_advertise": hclspec.NewAttr("driver_advertise", "bool", false), + "driver_port_map": hclspec.NewAttr("driver_port_map", "string", false), + "stdout_string": hclspec.NewAttr("stdout_string", "string", false), + "stdout_repeat": hclspec.NewAttr("stdout_repeat", "number", false), + "stdout_repeat_duration": hclspec.NewAttr("stdout_repeat_duration", "number", false), + }) + + // capabilities is returned by the Capabilities RPC and indicates what + // optional features this driver supports + capabilities = &drivers.Capabilities{ + SendSignals: true, + Exec: true, + FSIsolation: cstructs.FSIsolationNone, + } +) + +// Driver is a mock DriverPlugin implementation +type Driver struct { + // eventer is used to handle multiplexing of TaskEvents calls such that an + // event can be broadcast to all callers + eventer *eventer.Eventer + + // config is the driver configuration set by the SetConfig RPC + config *Config + + // tasks is the in memory datastore mapping taskIDs to mockDriverHandles + tasks *taskStore + + // ctx is the context for the driver. It is passed to other subsystems to + // coordinate shutdown + ctx context.Context + + // signalShutdown is called when the driver is shutting down and cancels the + // ctx passed to any subsystems + signalShutdown context.CancelFunc + + shutdownFingerprintTime time.Time + + // logger will log to the plugin output which is usually an 'executor.out' + // file located in the root of the TaskDir + logger hclog.Logger +} + +// Config is the configuration for the driver that applies to all tasks +type Config struct { + // ShutdownPeriodicAfter is a toggle that can be used during tests to + // "stop" a previously-functioning driver, allowing for testing of periodic + // drivers and fingerprinters + ShutdownPeriodicAfter bool `cty:"shutdown_periodic_after"` + + // ShutdownPeriodicDuration is a option that can be used during tests + // to "stop" a previously functioning driver after the specified duration + // for testing of periodic drivers and fingerprinters. + ShutdownPeriodicDuration time.Duration `cty:"shutdown_periodic_duration"` +} + +// TaskConfig is the driver configuration of a task within a job +type TaskConfig struct { + + // StartErr specifies the error that should be returned when starting the + // mock driver. + StartErr string `cty:"start_error"` + + // StartErrRecoverable marks the error returned is recoverable + StartErrRecoverable bool `cty:"start_error_recoverable"` + + // StartBlockFor specifies a duration in which to block before returning + StartBlockFor time.Duration `cty:"start_block_for"` + + // KillAfter is the duration after which the mock driver indicates the task + // has exited after getting the initial SIGINT signal + KillAfter time.Duration `cty:"kill_after"` + + // RunFor is the duration for which the fake task runs for. After this + // period the MockDriver responds to the task running indicating that the + // task has terminated + RunFor time.Duration `cty:"run_for"` + + // ExitCode is the exit code with which the MockDriver indicates the task + // has exited + ExitCode int `cty:"exit_code"` + + // ExitSignal is the signal with which the MockDriver indicates the task has + // been killed + ExitSignal int `cty:"exit_signal"` + + // ExitErrMsg is the error message that the task returns while exiting + ExitErrMsg string `cty:"exit_err_msg"` + + // SignalErr is the error message that the task returns if signalled + SignalErr string `cty:"signal_error"` + + // DriverIP will be returned as the DriverNetwork.IP from Start() + DriverIP string `cty:"driver_ip"` + + // DriverAdvertise will be returned as DriverNetwork.AutoAdvertise from + // Start(). + DriverAdvertise bool `cty:"driver_advertise"` + + // DriverPortMap will parse a label:number pair and return it in + // DriverNetwork.PortMap from Start(). + DriverPortMap string `cty:"driver_port_map"` + + // StdoutString is the string that should be sent to stdout + StdoutString string `cty:"stdout_string"` + + // StdoutRepeat is the number of times the output should be sent. + StdoutRepeat int `cty:"stdout_repeat"` + + // StdoutRepeatDur is the duration between repeated outputs. + StdoutRepeatDur time.Duration `cty:"stdout_repeat_duration"` +} + +type MockTaskState struct { + TaskConfig *drivers.TaskConfig + StartedAt time.Time +} + +func (d *Driver) PluginInfo() (*base.PluginInfoResponse, error) { + return pluginInfo, nil +} + +func (d *Driver) ConfigSchema() (*hclspec.Spec, error) { + return configSpec, nil +} + +func (d *Driver) SetConfig(data []byte) error { + var config Config + if err := base.MsgPackDecode(data, &config); err != nil { + return err + } + + d.config = &config + if d.config.ShutdownPeriodicAfter { + d.shutdownFingerprintTime = time.Now().Add(d.config.ShutdownPeriodicDuration) + } + return nil +} + +func (d *Driver) TaskConfigSchema() (*hclspec.Spec, error) { + return taskConfigSpec, nil +} + +func (d *Driver) Capabilities() (*drivers.Capabilities, error) { + return capabilities, nil +} + +func (d *Driver) Fingerprint(ctx context.Context) (<-chan *drivers.Fingerprint, error) { + ch := make(chan *drivers.Fingerprint) + go d.handleFingerprint(ctx, ch) + return ch, nil +} + +func (d *Driver) handleFingerprint(ctx context.Context, ch chan *drivers.Fingerprint) { + ticker := time.NewTimer(0) + for { + select { + case <-ctx.Done(): + return + case <-d.ctx.Done(): + return + case <-ticker.C: + ticker.Reset(fingerprintPeriod) + ch <- d.buildFingerprint() + } + } +} + +func (d *Driver) buildFingerprint() *drivers.Fingerprint { + var health drivers.HealthState + var desc string + attrs := map[string]string{} + if !d.shutdownFingerprintTime.IsZero() && time.Now().After(d.shutdownFingerprintTime) { + health = drivers.HealthStateUndetected + desc = "mock disabled" + } else { + health = drivers.HealthStateHealthy + attrs["driver.mock"] = "1" + desc = "mock enabled" + } + + return &drivers.Fingerprint{ + Attributes: map[string]string{}, + Health: health, + HealthDescription: desc, + } +} + +func (d *Driver) RecoverTask(*drivers.TaskHandle) error { + panic("not implemented") +} + +func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cstructs.DriverNetwork, error) { + var driverConfig TaskConfig + if err := cfg.DecodeDriverConfig(&driverConfig); err != nil { + return nil, nil, err + } + + if driverConfig.StartBlockFor != 0 { + time.Sleep(driverConfig.StartBlockFor) + } + + if driverConfig.StartErr != "" { + return nil, nil, structs.NewRecoverableError(errors.New(driverConfig.StartErr), driverConfig.StartErrRecoverable) + } + + // Create the driver network + net := &cstructs.DriverNetwork{ + IP: driverConfig.DriverIP, + AutoAdvertise: driverConfig.DriverAdvertise, + } + if raw := driverConfig.DriverPortMap; len(raw) > 0 { + parts := strings.Split(raw, ":") + if len(parts) != 2 { + return nil, nil, fmt.Errorf("malformed port map: %q", raw) + } + port, err := strconv.Atoi(parts[1]) + if err != nil { + return nil, nil, fmt.Errorf("malformed port map: %q -- error: %v", raw, err) + } + net.PortMap = map[string]int{parts[0]: port} + } + + h := &mockTaskHandle{ + task: cfg, + runFor: driverConfig.RunFor, + killAfter: driverConfig.KillAfter, + exitCode: driverConfig.ExitCode, + exitSignal: driverConfig.ExitSignal, + stdoutString: driverConfig.StdoutString, + stdoutRepeat: driverConfig.StdoutRepeat, + stdoutRepeatDur: driverConfig.StdoutRepeatDur, + logger: d.logger, + doneCh: make(chan struct{}), + } + if driverConfig.ExitErrMsg != "" { + h.exitErr = errors.New(driverConfig.ExitErrMsg) + } + if driverConfig.SignalErr != "" { + h.signalErr = fmt.Errorf(driverConfig.SignalErr) + } + + driverState := MockTaskState{ + TaskConfig: cfg, + StartedAt: h.startedAt, + } + handle := drivers.NewTaskHandle(pluginName) + handle.Config = cfg + if err := handle.SetDriverState(&driverState); err != nil { + d.logger.Error("failed to start task, error setting driver state", "error", err) + return nil, nil, fmt.Errorf("failed to set driver state: %v", err) + } + + d.tasks.Set(cfg.ID, h) + + d.logger.Debug("starting task", "name", cfg.Name) + go h.run() + return handle, net, nil + +} + +func (d *Driver) WaitTask(ctx context.Context, taskID string) (<-chan *drivers.ExitResult, error) { + handle, ok := d.tasks.Get(taskID) + if !ok { + return nil, drivers.ErrTaskNotFound + } + + ch := make(chan *drivers.ExitResult) + go d.handleWait(ctx, handle, ch) + + return ch, nil + +} +func (d *Driver) handleWait(ctx context.Context, handle *mockTaskHandle, ch chan *drivers.ExitResult) { + defer close(ch) + + select { + case <-ctx.Done(): + return + case <-d.ctx.Done(): + return + case <-handle.doneCh: + ch <- handle.exitResult + } +} +func (d *Driver) StopTask(taskID string, timeout time.Duration, signal string) error { + panic("not implemented") +} + +func (d *Driver) DestroyTask(taskID string, force bool) error { + panic("not implemented") +} + +func (d *Driver) InspectTask(taskID string) (*drivers.TaskStatus, error) { + panic("not implemented") +} + +func (d *Driver) TaskStats(taskID string) (*cstructs.TaskResourceUsage, error) { + panic("not implemented") +} + +func (d *Driver) TaskEvents(context.Context) (<-chan *drivers.TaskEvent, error) { + panic("not implemented") +} + +func (d *Driver) SignalTask(taskID string, signal string) error { + panic("not implemented") +} + +func (d *Driver) ExecTask(taskID string, cmd []string, timeout time.Duration) (*drivers.ExecTaskResult, error) { + panic("not implemented") +} diff --git a/drivers/mock/handle.go b/drivers/mock/handle.go new file mode 100644 index 000000000..2296610a1 --- /dev/null +++ b/drivers/mock/handle.go @@ -0,0 +1,88 @@ +package mock + +import ( + "io" + "time" + + hclog "github.com/hashicorp/go-hclog" + "github.com/hashicorp/nomad/client/lib/fifo" + "github.com/hashicorp/nomad/plugins/drivers" +) + +// mockTaskHandle is a task handler which supervises a mock task +type mockTaskHandle struct { + logger hclog.Logger + + runFor time.Duration + killAfter time.Duration + killTimeout time.Duration + exitCode int + exitSignal int + exitErr error + signalErr error + stdoutString string + stdoutRepeat int + stdoutRepeatDur time.Duration + + doneCh chan struct{} + + task *drivers.TaskConfig + procState drivers.TaskState + startedAt time.Time + completedAt time.Time + exitResult *drivers.ExitResult +} + +func (h *mockTaskHandle) run() { + + // Setup logging output + if h.stdoutString != "" { + go h.handleLogging() + } + + timer := time.NewTimer(h.runFor) + defer timer.Stop() + for { + select { + case <-timer.C: + select { + case <-h.doneCh: + // already closed + default: + close(h.doneCh) + } + case <-h.doneCh: + h.logger.Debug("finished running task", "name", h.task.Name) + h.exitResult = &drivers.ExitResult{ + ExitCode: h.exitCode, + Signal: h.exitSignal, + Err: h.exitErr, + } + return + } + } +} + +func (h *mockTaskHandle) handleLogging() { + stdout, err := fifo.Open(h.task.StdoutPath) + if err != nil { + h.exitErr = err + close(h.doneCh) + h.logger.Error("failed to write to stdout: %v", err) + return + } + + for i := 0; i < h.stdoutRepeat; i++ { + select { + case <-h.doneCh: + return + case <-time.After(h.stdoutRepeatDur): + if _, err := io.WriteString(stdout, h.stdoutString); err != nil { + h.exitErr = err + close(h.doneCh) + h.logger.Error("failed to write to stdout", "error", err) + return + } + } + } +} diff --git a/drivers/mock/state.go b/drivers/mock/state.go new file mode 100644 index 000000000..b3da44016 --- /dev/null +++ b/drivers/mock/state.go @@ -0,0 +1,33 @@ +package mock + +import ( + "sync" +) + +type taskStore struct { + store map[string]*mockTaskHandle + lock sync.RWMutex +} + +func newTaskStore() *taskStore { + return &taskStore{store: map[string]*mockTaskHandle{}} +} + +func (ts *taskStore) Set(id string, handle *mockTaskHandle) { + ts.lock.Lock() + defer ts.lock.Unlock() + ts.store[id] = handle +} + +func (ts *taskStore) Get(id string) (*mockTaskHandle, bool) { + ts.lock.RLock() + defer ts.lock.RUnlock() + t, ok := ts.store[id] + return t, ok +} + +func (ts *taskStore) Delete(id string) { + ts.lock.Lock() + defer ts.lock.Unlock() + delete(ts.store, id) +} From 118ce7c7825d6edf9b80b11fd1024d2d579c251e Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Fri, 12 Oct 2018 16:56:13 -0700 Subject: [PATCH 138/152] drivers/mock: complete plugin impl --- drivers/mock/driver.go | 99 ++++++++++++++++++++++++++++++++++++------ drivers/mock/handle.go | 61 ++++++++++++++------------ 2 files changed, 119 insertions(+), 41 deletions(-) diff --git a/drivers/mock/driver.go b/drivers/mock/driver.go index 141049865..83baef225 100644 --- a/drivers/mock/driver.go +++ b/drivers/mock/driver.go @@ -15,6 +15,8 @@ import ( "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/drivers" "github.com/hashicorp/nomad/plugins/shared/hclspec" + "github.com/hashicorp/nomad/plugins/shared/loader" + netctx "golang.org/x/net/context" ) const ( @@ -22,10 +24,22 @@ const ( pluginName = "mock" // fingerprintPeriod is the interval at which the driver will send fingerprint responses - fingerprintPeriod = 30 * time.Second + fingerprintPeriod = 500 * time.Millisecond ) var ( + // When the package is loaded the driver is registered as an internal plugin + // with the plugin catalog + PluginID = loader.PluginID{ + Name: pluginName, + PluginType: base.PluginTypeDriver, + } + + PluginConfig = &loader.InternalPluginConfig{ + Config: map[string]interface{}{}, + Factory: func(l hclog.Logger) interface{} { return NewMockDriver(l) }, + } + // pluginInfo is the response returned for the PluginInfo RPC pluginInfo = &base.PluginInfoResponse{ Type: base.PluginTypeDriver, @@ -66,7 +80,7 @@ var ( // capabilities is returned by the Capabilities RPC and indicates what // optional features this driver supports capabilities = &drivers.Capabilities{ - SendSignals: true, + SendSignals: false, Exec: true, FSIsolation: cstructs.FSIsolationNone, } @@ -99,6 +113,20 @@ type Driver struct { logger hclog.Logger } +// NewMockDriver returns a new DriverPlugin implementation +func NewMockDriver(logger hclog.Logger) drivers.DriverPlugin { + ctx, cancel := context.WithCancel(context.Background()) + logger = logger.Named(pluginName) + return &Driver{ + eventer: eventer.NewEventer(ctx, logger), + config: &Config{}, + tasks: newTaskStore(), + ctx: ctx, + signalShutdown: cancel, + logger: logger, + } +} + // Config is the configuration for the driver that applies to all tasks type Config struct { // ShutdownPeriodicAfter is a toggle that can be used during tests to @@ -203,7 +231,7 @@ func (d *Driver) Capabilities() (*drivers.Capabilities, error) { return capabilities, nil } -func (d *Driver) Fingerprint(ctx context.Context) (<-chan *drivers.Fingerprint, error) { +func (d *Driver) Fingerprint(ctx netctx.Context) (<-chan *drivers.Fingerprint, error) { ch := make(chan *drivers.Fingerprint) go d.handleFingerprint(ctx, ch) return ch, nil @@ -245,7 +273,8 @@ func (d *Driver) buildFingerprint() *drivers.Fingerprint { } func (d *Driver) RecoverTask(*drivers.TaskHandle) error { - panic("not implemented") + //TODO is there anything to do here? + return nil } func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cstructs.DriverNetwork, error) { @@ -279,6 +308,8 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cstru net.PortMap = map[string]int{parts[0]: port} } + killCtx, killCancel := context.WithCancel(context.Background()) + h := &mockTaskHandle{ task: cfg, runFor: driverConfig.RunFor, @@ -288,8 +319,10 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cstru stdoutString: driverConfig.StdoutString, stdoutRepeat: driverConfig.StdoutRepeat, stdoutRepeatDur: driverConfig.StdoutRepeatDur, - logger: d.logger, - doneCh: make(chan struct{}), + logger: d.logger.With("task_name", cfg.Name), + waitCh: make(chan struct{}), + killCh: killCtx.Done(), + kill: killCancel, } if driverConfig.ExitErrMsg != "" { h.exitErr = errors.New(driverConfig.ExitErrMsg) @@ -317,7 +350,7 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cstru } -func (d *Driver) WaitTask(ctx context.Context, taskID string) (<-chan *drivers.ExitResult, error) { +func (d *Driver) WaitTask(ctx netctx.Context, taskID string) (<-chan *drivers.ExitResult, error) { handle, ok := d.tasks.Get(taskID) if !ok { return nil, drivers.ErrTaskNotFound @@ -337,16 +370,39 @@ func (d *Driver) handleWait(ctx context.Context, handle *mockTaskHandle, ch chan return case <-d.ctx.Done(): return - case <-handle.doneCh: + case <-handle.waitCh: ch <- handle.exitResult } } func (d *Driver) StopTask(taskID string, timeout time.Duration, signal string) error { - panic("not implemented") + h, ok := d.tasks.Get(taskID) + if !ok { + return drivers.ErrTaskNotFound + } + + d.logger.Debug("killing task", + "task_name", h.task.Name, + "kill_after", h.killAfter, + "kill_timeout", h.killTimeout, + ) + + select { + case <-h.waitCh: + d.logger.Debug("not killing task: already exited", "task_name", h.task.Name) + case <-time.After(h.killAfter): + d.logger.Debug("killing task due to kill_after", "task_name", h.task.Name) + h.kill() + case <-time.After(h.killTimeout): + d.logger.Debug("killing task after kill_timeout", "task_name", h.task.Name) + h.kill() + } + return nil } func (d *Driver) DestroyTask(taskID string, force bool) error { - panic("not implemented") + //TODO is there anything else to do here? + d.tasks.Delete(taskID) + return nil } func (d *Driver) InspectTask(taskID string) (*drivers.TaskStatus, error) { @@ -354,17 +410,32 @@ func (d *Driver) InspectTask(taskID string) (*drivers.TaskStatus, error) { } func (d *Driver) TaskStats(taskID string) (*cstructs.TaskResourceUsage, error) { - panic("not implemented") + //TODO return an error? + return nil, nil } -func (d *Driver) TaskEvents(context.Context) (<-chan *drivers.TaskEvent, error) { +func (d *Driver) TaskEvents(netctx.Context) (<-chan *drivers.TaskEvent, error) { panic("not implemented") } func (d *Driver) SignalTask(taskID string, signal string) error { - panic("not implemented") + h, ok := d.tasks.Get(taskID) + if !ok { + return drivers.ErrTaskNotFound + } + + return h.signalErr } func (d *Driver) ExecTask(taskID string, cmd []string, timeout time.Duration) (*drivers.ExecTaskResult, error) { - panic("not implemented") + h, ok := d.tasks.Get(taskID) + if !ok { + return nil, drivers.ErrTaskNotFound + } + + res := drivers.ExecTaskResult{ + Stdout: []byte(fmt.Sprintf("Exec(%q, %q)", h.task.Name, cmd)), + ExitResult: &drivers.ExitResult{}, + } + return &res, nil } diff --git a/drivers/mock/handle.go b/drivers/mock/handle.go index 2296610a1..214ecba1d 100644 --- a/drivers/mock/handle.go +++ b/drivers/mock/handle.go @@ -1,6 +1,7 @@ package mock import ( + "context" "io" "time" @@ -16,6 +17,7 @@ type mockTaskHandle struct { runFor time.Duration killAfter time.Duration killTimeout time.Duration + waitCh chan struct{} exitCode int exitSignal int exitErr error @@ -24,63 +26,68 @@ type mockTaskHandle struct { stdoutRepeat int stdoutRepeatDur time.Duration - doneCh chan struct{} - task *drivers.TaskConfig procState drivers.TaskState startedAt time.Time completedAt time.Time exitResult *drivers.ExitResult + + // Calling kill closes killCh if it is not already closed + kill context.CancelFunc + killCh <-chan struct{} } func (h *mockTaskHandle) run() { + defer close(h.waitCh) + + errCh := make(chan error, 1) // Setup logging output if h.stdoutString != "" { - go h.handleLogging() + go h.handleLogging(errCh) } timer := time.NewTimer(h.runFor) defer timer.Stop() - for { - select { - case <-timer.C: - select { - case <-h.doneCh: - // already closed - default: - close(h.doneCh) - } - case <-h.doneCh: - h.logger.Debug("finished running task", "name", h.task.Name) - h.exitResult = &drivers.ExitResult{ - ExitCode: h.exitCode, - Signal: h.exitSignal, - Err: h.exitErr, - } - return + + select { + case <-timer.C: + h.logger.Debug("run_for time elapsed; exiting", "run_for", h.runFor) + case <-h.killCh: + h.logger.Debug("killed; exiting") + case err := <-errCh: + h.logger.Error("error running mock task; exiting", "error", err) + h.exitResult = &drivers.ExitResult{ + Err: err, } + return } + + h.exitResult = &drivers.ExitResult{ + ExitCode: h.exitCode, + Signal: h.exitSignal, + Err: h.exitErr, + } + return } -func (h *mockTaskHandle) handleLogging() { +func (h *mockTaskHandle) handleLogging(errCh chan<- error) { stdout, err := fifo.Open(h.task.StdoutPath) if err != nil { - h.exitErr = err - close(h.doneCh) - h.logger.Error("failed to write to stdout: %v", err) + h.logger.Error("failed to write to stdout", "error", err) + errCh <- err return } for i := 0; i < h.stdoutRepeat; i++ { select { - case <-h.doneCh: + case <-h.waitCh: + h.logger.Warn("exiting before done writing output", "i", i, "total", h.stdoutRepeat) return case <-time.After(h.stdoutRepeatDur): if _, err := io.WriteString(stdout, h.stdoutString); err != nil { - h.exitErr = err - close(h.doneCh) h.logger.Error("failed to write to stdout", "error", err) + errCh <- err return } } From 1a29337e48e745dea85d696e913368b7055b27ad Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 15 Oct 2018 10:58:09 -0700 Subject: [PATCH 139/152] register drivers by default Do not register mock_driver on release builds. --- command/agent/plugin_internal_catalog.go | 12 -------- drivers/rawexec/driver.go | 33 +++++++++++----------- plugins/shared/catalog/register.go | 10 +++++++ plugins/shared/catalog/register_testing.go | 12 ++++++++ 4 files changed, 39 insertions(+), 28 deletions(-) delete mode 100644 command/agent/plugin_internal_catalog.go create mode 100644 plugins/shared/catalog/register.go create mode 100644 plugins/shared/catalog/register_testing.go diff --git a/command/agent/plugin_internal_catalog.go b/command/agent/plugin_internal_catalog.go deleted file mode 100644 index 034dea2cf..000000000 --- a/command/agent/plugin_internal_catalog.go +++ /dev/null @@ -1,12 +0,0 @@ -package agent - -import ( - - // Each internal plugin has an init func which registers itself with the - // plugin catalog. Since the plugin implementations are not imported by the - // client or server they must be referenced here so plugin registration - // occures. - - // raw_exec driver - _ "github.com/hashicorp/nomad/drivers/rawexec" -) diff --git a/drivers/rawexec/driver.go b/drivers/rawexec/driver.go index 7e8fc2643..fce320c8d 100644 --- a/drivers/rawexec/driver.go +++ b/drivers/rawexec/driver.go @@ -17,7 +17,6 @@ import ( "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/drivers" "github.com/hashicorp/nomad/plugins/drivers/utils" - "github.com/hashicorp/nomad/plugins/shared/catalog" "github.com/hashicorp/nomad/plugins/shared/hclspec" "github.com/hashicorp/nomad/plugins/shared/loader" "golang.org/x/net/context" @@ -25,25 +24,27 @@ import ( // When the package is loaded the driver is registered as an internal plugin // with the plugin catalog -func init() { - catalog.RegisterDeferredConfig(loader.PluginID{ +var ( + PluginID = loader.PluginID{ Name: pluginName, PluginType: base.PluginTypeDriver, - }, &loader.InternalPluginConfig{ + } + + PluginConfig = &loader.InternalPluginConfig{ Config: map[string]interface{}{}, Factory: func(l hclog.Logger) interface{} { return NewRawExecDriver(l) }, - }, - func(opts map[string]string) (map[string]interface{}, error) { - conf := map[string]interface{}{} - if v, err := strconv.ParseBool(opts["driver.raw_exec.enable"]); err == nil { - conf["enabled"] = v - } - if v, err := strconv.ParseBool(opts["driver.raw_exec.no_cgroups"]); err == nil { - conf["no_cgroups"] = v - } - return conf, nil - }, - ) + } +) + +func PluginLoader(opts map[string]string) (map[string]interface{}, error) { + conf := map[string]interface{}{} + if v, err := strconv.ParseBool(opts["driver.raw_exec.enable"]); err == nil { + conf["enabled"] = v + } + if v, err := strconv.ParseBool(opts["driver.raw_exec.no_cgroups"]); err == nil { + conf["no_cgroups"] = v + } + return conf, nil } const ( diff --git a/plugins/shared/catalog/register.go b/plugins/shared/catalog/register.go new file mode 100644 index 000000000..553ff2aef --- /dev/null +++ b/plugins/shared/catalog/register.go @@ -0,0 +1,10 @@ +package catalog + +import "github.com/hashicorp/nomad/drivers/rawexec" + +// This file is where all builtin plugins should be registered in the catalog. +// Plugins with build restrictions should be placed in the appropriate +// register_XXX.go file. +func init() { + RegisterDeferredConfig(rawexec.PluginID, rawexec.PluginConfig, rawexec.PluginLoader) +} diff --git a/plugins/shared/catalog/register_testing.go b/plugins/shared/catalog/register_testing.go new file mode 100644 index 000000000..d3d58c925 --- /dev/null +++ b/plugins/shared/catalog/register_testing.go @@ -0,0 +1,12 @@ +// +build !release + +package catalog + +import "github.com/hashicorp/nomad/drivers/mock" + +// Register the mock driver with the builtin driver plugin catalog. All builtin +// plugins that are intended for production use should be registered in +// register.go as this file is not built as part of a release. +func init() { + Register(mock.PluginID, mock.PluginConfig) +} From ed5641055f18e14a03331b83ea866d9d1422f3e0 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 15 Oct 2018 14:08:12 -0700 Subject: [PATCH 140/152] fixup comments, logging, and missing method impls from #4777 comments --- drivers/mock/driver.go | 33 ++++++++++++++++++--------------- drivers/mock/handle.go | 29 +++++++++++++++++++++++++---- drivers/rawexec/driver.go | 22 ++++++++++++---------- 3 files changed, 55 insertions(+), 29 deletions(-) diff --git a/drivers/mock/driver.go b/drivers/mock/driver.go index 83baef225..4daba649a 100644 --- a/drivers/mock/driver.go +++ b/drivers/mock/driver.go @@ -28,13 +28,15 @@ const ( ) var ( - // When the package is loaded the driver is registered as an internal plugin - // with the plugin catalog + // PluginID is the mock driver plugin metadata registered in the plugin + // catalog. PluginID = loader.PluginID{ Name: pluginName, PluginType: base.PluginTypeDriver, } + // PluginConfig is the mock driver factory function registered in the + // plugin catalog. PluginConfig = &loader.InternalPluginConfig{ Config: map[string]interface{}{}, Factory: func(l hclog.Logger) interface{} { return NewMockDriver(l) }, @@ -338,13 +340,13 @@ func (d *Driver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cstru handle := drivers.NewTaskHandle(pluginName) handle.Config = cfg if err := handle.SetDriverState(&driverState); err != nil { - d.logger.Error("failed to start task, error setting driver state", "error", err) + d.logger.Error("failed to start task, error setting driver state", "error", err, "task_name", cfg.Name) return nil, nil, fmt.Errorf("failed to set driver state: %v", err) } d.tasks.Set(cfg.ID, h) - d.logger.Debug("starting task", "name", cfg.Name) + d.logger.Debug("starting task", "task_name", cfg.Name) go h.run() return handle, net, nil @@ -380,11 +382,7 @@ func (d *Driver) StopTask(taskID string, timeout time.Duration, signal string) e return drivers.ErrTaskNotFound } - d.logger.Debug("killing task", - "task_name", h.task.Name, - "kill_after", h.killAfter, - "kill_timeout", h.killTimeout, - ) + d.logger.Debug("killing task", "task_name", h.task.Name, "kill_after", h.killAfter) select { case <-h.waitCh: @@ -392,15 +390,20 @@ func (d *Driver) StopTask(taskID string, timeout time.Duration, signal string) e case <-time.After(h.killAfter): d.logger.Debug("killing task due to kill_after", "task_name", h.task.Name) h.kill() - case <-time.After(h.killTimeout): - d.logger.Debug("killing task after kill_timeout", "task_name", h.task.Name) - h.kill() } return nil } func (d *Driver) DestroyTask(taskID string, force bool) error { - //TODO is there anything else to do here? + handle, ok := d.tasks.Get(taskID) + if !ok { + return drivers.ErrTaskNotFound + } + + if handle.IsRunning() && !force { + return fmt.Errorf("cannot destroy running task") + } + d.tasks.Delete(taskID) return nil } @@ -414,8 +417,8 @@ func (d *Driver) TaskStats(taskID string) (*cstructs.TaskResourceUsage, error) { return nil, nil } -func (d *Driver) TaskEvents(netctx.Context) (<-chan *drivers.TaskEvent, error) { - panic("not implemented") +func (d *Driver) TaskEvents(ctx netctx.Context) (<-chan *drivers.TaskEvent, error) { + return d.eventer.TaskEvents(ctx) } func (d *Driver) SignalTask(taskID string, signal string) error { diff --git a/drivers/mock/handle.go b/drivers/mock/handle.go index 214ecba1d..e20215943 100644 --- a/drivers/mock/handle.go +++ b/drivers/mock/handle.go @@ -3,6 +3,7 @@ package mock import ( "context" "io" + "sync" "time" hclog "github.com/hashicorp/go-hclog" @@ -16,7 +17,6 @@ type mockTaskHandle struct { runFor time.Duration killAfter time.Duration - killTimeout time.Duration waitCh chan struct{} exitCode int exitSignal int @@ -26,8 +26,12 @@ type mockTaskHandle struct { stdoutRepeat int stdoutRepeatDur time.Duration - task *drivers.TaskConfig - procState drivers.TaskState + task *drivers.TaskConfig + + // stateLock guards the procState field + stateLock sync.Mutex + procState drivers.TaskState + startedAt time.Time completedAt time.Time exitResult *drivers.ExitResult @@ -37,8 +41,25 @@ type mockTaskHandle struct { killCh <-chan struct{} } +func (h *mockTaskHandle) IsRunning() bool { + h.stateLock.Lock() + defer h.stateLock.Unlock() + return h.procState == drivers.TaskStateRunning +} + func (h *mockTaskHandle) run() { - defer close(h.waitCh) + defer func() { + h.stateLock.Lock() + h.procState = drivers.TaskStateExited + h.stateLock.Unlock() + + h.completedAt = time.Now() + close(h.waitCh) + }() + + h.stateLock.Lock() + h.procState = drivers.TaskStateRunning + h.stateLock.Unlock() errCh := make(chan error, 1) diff --git a/drivers/rawexec/driver.go b/drivers/rawexec/driver.go index fce320c8d..12da75609 100644 --- a/drivers/rawexec/driver.go +++ b/drivers/rawexec/driver.go @@ -22,14 +22,24 @@ import ( "golang.org/x/net/context" ) -// When the package is loaded the driver is registered as an internal plugin -// with the plugin catalog +const ( + // pluginName is the name of the plugin + pluginName = "raw_exec" + + // fingerprintPeriod is the interval at which the driver will send fingerprint responses + fingerprintPeriod = 30 * time.Second +) + var ( + // PluginID is the rawexec plugin metadata registered in the plugin + // catalog. PluginID = loader.PluginID{ Name: pluginName, PluginType: base.PluginTypeDriver, } + // PluginConfig is the rawexec factory function registered in the + // plugin catalog. PluginConfig = &loader.InternalPluginConfig{ Config: map[string]interface{}{}, Factory: func(l hclog.Logger) interface{} { return NewRawExecDriver(l) }, @@ -47,14 +57,6 @@ func PluginLoader(opts map[string]string) (map[string]interface{}, error) { return conf, nil } -const ( - // pluginName is the name of the plugin - pluginName = "raw_exec" - - // fingerprintPeriod is the interval at which the driver will send fingerprint responses - fingerprintPeriod = 30 * time.Second -) - var ( // pluginInfo is the response returned for the PluginInfo RPC pluginInfo = &base.PluginInfoResponse{ From 02b1f26d3442b500e120474af281ffc9fae92bb1 Mon Sep 17 00:00:00 2001 From: Preetha Appan Date: Mon, 15 Oct 2018 17:40:38 -0500 Subject: [PATCH 141/152] Switch back to using map[string]string for port map --- drivers/rkt/driver.go | 41 ++++++++++++++------------------------ drivers/rkt/driver_test.go | 4 ++-- 2 files changed, 17 insertions(+), 28 deletions(-) diff --git a/drivers/rkt/driver.go b/drivers/rkt/driver.go index 2a25f9a6f..05f89aec0 100644 --- a/drivers/rkt/driver.go +++ b/drivers/rkt/driver.go @@ -86,7 +86,7 @@ var ( "dns_servers": hclspec.NewAttr("dns_servers", "list(string)", false), "dns_search_domains": hclspec.NewAttr("dns_search_domains", "list(string)", false), "net": hclspec.NewAttr("net", "list(string)", false), - "port_map": hclspec.NewAttr("port_map", "list(map(string))", false), + "port_map": hclspec.NewAttr("port_map", "map(string)", false), "volumes": hclspec.NewAttr("volumes", "list(string)", false), "insecure_options": hclspec.NewAttr("insecure_options", "list(string)", false), "no_overlay": hclspec.NewAttr("no_overlay", "bool", false), @@ -118,16 +118,16 @@ type Config struct { // TaskConfig is the driver configuration of a taskConfig within a job type TaskConfig struct { - ImageName string `codec:"image" cty:"image"` - Command string `codec:"command" cty:"command"` - Args []string `codec:"args" cty:"args"` - TrustPrefix string `codec:"trust_prefix" cty:"trust_prefix"` - DNSServers []string `codec:"dns_servers" cty:"dns_servers"` // DNS Server for containers - DNSSearchDomains []string `codec:"dns_search_domains" cty:"dns_search_domains"` // DNS Search domains for containers - Net []string `codec:"net" cty:"net"` // Networks for the containers - PortMap []map[string]string `codec:"port_map" cty:"port_map"` // A map of host port and the port name defined in the image manifest file - Volumes []string `codec:"volumes" cty:"volumes"` // Host-Volumes to mount in, syntax: /path/to/host/directory:/destination/path/in/container[:readOnly] - InsecureOptions []string `codec:"insecure_options" cty:"insecure_options"` // list of args for --insecure-options + ImageName string `codec:"image" cty:"image"` + Command string `codec:"command" cty:"command"` + Args []string `codec:"args" cty:"args"` + TrustPrefix string `codec:"trust_prefix" cty:"trust_prefix"` + DNSServers []string `codec:"dns_servers" cty:"dns_servers"` // DNS Server for containers + DNSSearchDomains []string `codec:"dns_search_domains" cty:"dns_search_domains"` // DNS Search domains for containers + Net []string `codec:"net" cty:"net"` // Networks for the containers + PortMap map[string]string `codec:"port_map" cty:"port_map"` // A map of host port and the port name defined in the image manifest file + Volumes []string `codec:"volumes" cty:"volumes"` // Host-Volumes to mount in, syntax: /path/to/host/directory:/destination/path/in/container[:readOnly] + InsecureOptions []string `codec:"insecure_options" cty:"insecure_options"` // list of args for --insecure-options NoOverlay bool `codec:"no_overlay" cty:"no_overlay"` // disable overlayfs for rkt run Debug bool `codec:"debug" cty:"debug"` // Enable debug option for rkt command @@ -344,6 +344,7 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs } var driverConfig TaskConfig + if err := cfg.DecodeDriverConfig(&driverConfig); err != nil { return nil, nil, fmt.Errorf("failed to decode driver config: %v", err) } @@ -495,8 +496,6 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs runArgs = append(runArgs, fmt.Sprintf("--net=%s", network)) } - mergedPortMap := mapMergeStrStr(driverConfig.PortMap...) - // Setup port mapping and exposed ports if len(cfg.Resources.NomadResources.Networks) == 0 { d.logger.Debug("no network interfaces are available") @@ -511,7 +510,7 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs for _, port := range network.ReservedPorts { var containerPort string - mapped, ok := mergedPortMap[port.Label] + mapped, ok := driverConfig.PortMap[port.Label] if !ok { // If the user doesn't have a mapped port using port_map, driver stops running container. return nil, nil, fmt.Errorf("port_map is not set. When you defined port in the resources, you need to configure port_map.") @@ -529,7 +528,7 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs // By default we will map the allocated port 1:1 to the container var containerPort string - if mapped, ok := mergedPortMap[port.Label]; ok { + if mapped, ok := driverConfig.PortMap[port.Label]; ok { containerPort = mapped } else { // If the user doesn't have mapped a port using port_map, driver stops running container. @@ -666,7 +665,7 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs var driverNetwork *cstructs.DriverNetwork if network != "host" && network != "none" { d.logger.Debug("retrieving network information for pod", "pod", img, "UUID", uuid, "task", cfg.Name) - driverNetwork, err = rktGetDriverNetwork(uuid, mergedPortMap, d.logger) + driverNetwork, err = rktGetDriverNetwork(uuid, driverConfig.PortMap, d.logger) if err != nil && !pluginClient.Exited() { d.logger.Warn("network status retrieval for pod failed", "pod", img, "UUID", uuid, "task", cfg.Name, "error", err) @@ -995,13 +994,3 @@ func (d *RktDriver) handleWait(ctx context.Context, handle *rktTaskHandle, ch ch case ch <- result: } } - -func mapMergeStrStr(maps ...map[string]string) map[string]string { - out := map[string]string{} - for _, in := range maps { - for key, val := range in { - out[key] = val - } - } - return out -} diff --git a/drivers/rkt/driver_test.go b/drivers/rkt/driver_test.go index 418efef70..82194a2ed 100644 --- a/drivers/rkt/driver_test.go +++ b/drivers/rkt/driver_test.go @@ -527,9 +527,9 @@ func TestRktDriver_PortMapping(t *testing.T) { taskConfig := map[string]interface{}{ "image": "docker://redis:3.2-alpine", - "port_map": []map[string]string{{ + "port_map": map[string]string{ "main": "6379-tcp", - }}, + }, "debug": "true", } From 3280727c4ec456b4e46c0811bebed6f1b4aba318 Mon Sep 17 00:00:00 2001 From: Preetha Appan Date: Mon, 15 Oct 2018 17:49:19 -0500 Subject: [PATCH 142/152] Address review comments --- drivers/rkt/driver.go | 7 ++----- drivers/rkt/driver_test.go | 2 ++ 2 files changed, 4 insertions(+), 5 deletions(-) diff --git a/drivers/rkt/driver.go b/drivers/rkt/driver.go index 05f89aec0..a63e14401 100644 --- a/drivers/rkt/driver.go +++ b/drivers/rkt/driver.go @@ -10,7 +10,6 @@ import ( "os/exec" "path/filepath" "regexp" - "runtime" "strings" "syscall" "time" @@ -104,8 +103,6 @@ var ( reRktVersion = regexp.MustCompile(`rkt [vV]ersion[:]? (\d[.\d]+)`) reAppcVersion = regexp.MustCompile(`appc [vV]ersion[:]? (\d[.\d]+)`) - - _ drivers.DriverPlugin = (*RktDriver)(nil) ) // Config is the client configuration for the driver @@ -240,8 +237,8 @@ func (d *RktDriver) buildFingerprint() *drivers.Fingerprint { HealthDescription: "healthy", } - // Only enable if we are root when running on non-windows systems. - if runtime.GOOS != "windows" && syscall.Geteuid() != 0 { + // Only enable if we are root + if syscall.Geteuid() != 0 { d.logger.Debug("must run as root user, disabling") fingerprint.Health = drivers.HealthStateUndetected fingerprint.HealthDescription = "driver must run as root user" diff --git a/drivers/rkt/driver_test.go b/drivers/rkt/driver_test.go index 82194a2ed..f3240d6bd 100644 --- a/drivers/rkt/driver_test.go +++ b/drivers/rkt/driver_test.go @@ -29,6 +29,8 @@ import ( "golang.org/x/net/context" ) +var _ drivers.DriverPlugin = (*RktDriver)(nil) + func TestRktVersionRegex(t *testing.T) { ctestutil.RktCompatible(t) t.Parallel() From d634157158be5dc6d8d0426c88f98a316b8bf9b4 Mon Sep 17 00:00:00 2001 From: Preetha Appan Date: Mon, 15 Oct 2018 17:59:25 -0500 Subject: [PATCH 143/152] Address review comments around logging task properly --- drivers/rkt/driver.go | 18 ++++++++---------- 1 file changed, 8 insertions(+), 10 deletions(-) diff --git a/drivers/rkt/driver.go b/drivers/rkt/driver.go index a63e14401..a06c8b9d0 100644 --- a/drivers/rkt/driver.go +++ b/drivers/rkt/driver.go @@ -153,7 +153,7 @@ type RktDriver struct { // config is the driver configuration set by the SetConfig RPC config *Config - // tasks is the in memory datastore mapping taskIDs to execDriverHandles + // tasks is the in memory datastore mapping taskIDs to rktTaskHandles tasks *taskStore // ctx is the context for the driver. It is passed to other subsystems to @@ -373,7 +373,7 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs return nil, nil, fmt.Errorf("Error running rkt trust: %s\n\nOutput: %s\n\nError: %s", err, outBuf.String(), errBuf.String()) } - d.logger.Debug("added trust prefix", "trust_prefix", trustPrefix) + d.logger.Debug("added trust prefix", "trust_prefix", trustPrefix, "task_name", cfg.Name) } else { // Disable signature verification if the trust command was not run. insecure = true @@ -501,7 +501,7 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs } } else if network == "host" { // Port mapping is skipped when host networking is used. - d.logger.Debug("driver.rkt: Ignoring port_map when using --net=host") + d.logger.Debug("Ignoring port_map when using --net=host", "task_name", cfg.Name) } else { network := cfg.Resources.NomadResources.Networks[0] for _, port := range network.ReservedPorts { @@ -534,7 +534,7 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs hostPortStr := strconv.Itoa(port.Value) - d.logger.Debug("driver.rkt: exposed port", "containerPort", containerPort) + d.logger.Debug("exposed port", "containerPort", containerPort, "task_name", cfg.Name) // Add port option to rkt run arguments. rkt allows multiple port args prepareArgs = append(prepareArgs, fmt.Sprintf("--port=%s:%s", containerPort, hostPortStr)) } @@ -646,7 +646,7 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs } if err := handle.SetDriverState(&rktDriverState); err != nil { - d.logger.Error("failed to start task, error setting driver state", "error", err) + d.logger.Error("failed to start task, error setting driver state", "error", err, "task_name", cfg.Name) execImpl.Shutdown("", 0) pluginClient.Kill() return nil, nil, fmt.Errorf("failed to set driver state: %v", err) @@ -661,10 +661,10 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs // https://coreos.com/rkt/docs/latest/networking/overview.html#no-loopback-only-networking var driverNetwork *cstructs.DriverNetwork if network != "host" && network != "none" { - d.logger.Debug("retrieving network information for pod", "pod", img, "UUID", uuid, "task", cfg.Name) + d.logger.Debug("retrieving network information for pod", "pod", img, "UUID", uuid, "task_name", cfg.Name) driverNetwork, err = rktGetDriverNetwork(uuid, driverConfig.PortMap, d.logger) if err != nil && !pluginClient.Exited() { - d.logger.Warn("network status retrieval for pod failed", "pod", img, "UUID", uuid, "task", cfg.Name, "error", err) + d.logger.Warn("network status retrieval for pod failed", "pod", img, "UUID", uuid, "task_name", cfg.Name, "error", err) // If a portmap was given, this turns into a fatal error if len(driverConfig.PortMap) != 0 { @@ -775,7 +775,7 @@ func (d *RktDriver) SignalTask(taskID string, signal string) error { sig := os.Interrupt if s, ok := signals.SignalLookup[signal]; ok { - d.logger.Warn("signal to send to task unknown, using SIGINT", "signal", signal, "task_id", handle.taskConfig.ID) + d.logger.Warn("signal to send to task unknown, using SIGINT", "signal", signal, "task_id", handle.taskConfig.ID, "task_name", handle.taskConfig.Name) sig = s } return handle.exec.Signal(sig) @@ -985,9 +985,7 @@ func (d *RktDriver) handleWait(ctx context.Context, handle *rktTaskHandle, ch ch select { case <-ctx.Done(): - return case <-d.ctx.Done(): - return case ch <- result: } } From fd2db90569ad780b827b0776f00f54e092930003 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 15 Oct 2018 20:03:05 -0700 Subject: [PATCH 144/152] plugins: msgpack codec should use cty tags --- plugins/base/plugin.go | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/plugins/base/plugin.go b/plugins/base/plugin.go index 32a369900..05afef607 100644 --- a/plugins/base/plugin.go +++ b/plugins/base/plugin.go @@ -52,7 +52,9 @@ func (p *PluginBase) GRPCClient(ctx context.Context, broker *plugin.GRPCBroker, // MsgpackHandle is a shared handle for encoding/decoding of structs var MsgpackHandle = func() *codec.MsgpackHandle { - h := &codec.MsgpackHandle{RawToString: true} + h := &codec.MsgpackHandle{} + h.RawToString = true + h.TypeInfos = codec.NewTypeInfos([]string{"cty", "codec"}) h.MapType = reflect.TypeOf(map[string]interface{}(nil)) return h }() From 64605a3957ebf386ba1f7b688541702324d9a5ad Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 15 Oct 2018 20:03:51 -0700 Subject: [PATCH 145/152] drivers/mock: fix plugin name Was mock_driver before plugins, so keep the name. --- drivers/mock/driver.go | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/drivers/mock/driver.go b/drivers/mock/driver.go index 4daba649a..3e8bd5818 100644 --- a/drivers/mock/driver.go +++ b/drivers/mock/driver.go @@ -21,7 +21,7 @@ import ( const ( // pluginName is the name of the plugin - pluginName = "mock" + pluginName = "mock_driver" // fingerprintPeriod is the interval at which the driver will send fingerprint responses fingerprintPeriod = 500 * time.Millisecond From 4ea5217d72590fe4928faa25c148c1b0a108c7ec Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Mon, 15 Oct 2018 20:06:35 -0700 Subject: [PATCH 146/152] tr: remove unused DriverHandle interface was causing typed nil interface panics and served no purpose --- .../taskrunner/interfaces/handle.go | 36 ------------------- client/allocrunner/taskrunner/task_runner.go | 3 +- .../taskrunner/task_runner_getters.go | 3 +- 3 files changed, 2 insertions(+), 40 deletions(-) diff --git a/client/allocrunner/taskrunner/interfaces/handle.go b/client/allocrunner/taskrunner/interfaces/handle.go index a0737c74f..3fafaba15 100644 --- a/client/allocrunner/taskrunner/interfaces/handle.go +++ b/client/allocrunner/taskrunner/interfaces/handle.go @@ -1,45 +1,9 @@ package interfaces import ( - "context" "time" - - cstructs "github.com/hashicorp/nomad/client/structs" - "github.com/hashicorp/nomad/nomad/structs" - "github.com/hashicorp/nomad/plugins/drivers" ) -// DriverHandle wraps operations to a driver such that they are operated on a specific -// task -type DriverHandle interface { - // ID returns the task ID - ID() string - - // WaitCh is used to return a channel used to wait for task completion - WaitCh(context.Context) (<-chan *drivers.ExitResult, error) - - // Update is used to update the task if possible and update task related - // configurations. - Update(task *structs.Task) error - - // Kill is used to stop the task - Kill() error - - // Stats returns aggregated stats of the driver - Stats() (*cstructs.TaskResourceUsage, error) - - // Signal is used to send a signal to the task - Signal(s string) error - - // ScriptExecutor is an interface used to execute commands such as - // health check scripts in the a DriverHandle's context. - ScriptExecutor - - // Network returns the driver's network or nil if the driver did not - // create a network. - Network() *cstructs.DriverNetwork -} - // ScriptExecutor is an interface that supports Exec()ing commands in the // driver's context. Split out of DriverHandle to ease testing. type ScriptExecutor interface { diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index c3a49c94a..49c560b50 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -15,7 +15,6 @@ import ( "github.com/hashicorp/hcl2/hcldec" "github.com/hashicorp/nomad/client/allocdir" "github.com/hashicorp/nomad/client/allocrunner/interfaces" - tinterfaces "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/client/allocrunner/taskrunner/restarts" "github.com/hashicorp/nomad/client/allocrunner/taskrunner/state" "github.com/hashicorp/nomad/client/config" @@ -548,7 +547,7 @@ func (tr *TaskRunner) initDriver() error { // handleDestroy will retry with an exponential backoff and will give up at a // given limit. It returns whether the task was destroyed and the error // associated with the last kill attempt. -func (tr *TaskRunner) handleDestroy(handle tinterfaces.DriverHandle) (destroyed bool, err error) { +func (tr *TaskRunner) handleDestroy(handle *DriverHandle) (destroyed bool, err error) { // Cap the number of times we attempt to kill the task. for i := 0; i < killFailureLimit; i++ { if err = handle.Kill(); err != nil { diff --git a/client/allocrunner/taskrunner/task_runner_getters.go b/client/allocrunner/taskrunner/task_runner_getters.go index 8a5661be0..61eaf9711 100644 --- a/client/allocrunner/taskrunner/task_runner_getters.go +++ b/client/allocrunner/taskrunner/task_runner_getters.go @@ -1,7 +1,6 @@ package taskrunner import ( - "github.com/hashicorp/nomad/client/allocrunner/taskrunner/interfaces" "github.com/hashicorp/nomad/nomad/structs" ) @@ -51,7 +50,7 @@ func (tr *TaskRunner) setVaultToken(token string) { // getDriverHandle returns a driver handle and its result proxy. Use the // result proxy instead of the handle's WaitCh. -func (tr *TaskRunner) getDriverHandle() interfaces.DriverHandle { +func (tr *TaskRunner) getDriverHandle() *DriverHandle { tr.handleLock.Lock() defer tr.handleLock.Unlock() return tr.handle From 448a709603096541ef853ee44c92517496fe1f86 Mon Sep 17 00:00:00 2001 From: Nick Ethier Date: Tue, 16 Oct 2018 13:49:50 -0400 Subject: [PATCH 147/152] driver/docker: rename container logger --- drivers/docker/docklog/client.go | 10 +- .../docklog/{docklog.go => docker_logger.go} | 10 +- ...{docklog_test.go => docker_logger_test.go} | 29 ++--- drivers/docker/docklog/plugin.go | 22 ++-- .../{docklog.pb.go => docker_logger.pb.go} | 110 +++++++++--------- .../{docklog.proto => docker_logger.proto} | 2 +- drivers/docker/docklog/server.go | 14 +-- 7 files changed, 100 insertions(+), 97 deletions(-) rename drivers/docker/docklog/{docklog.go => docker_logger.go} (91%) rename drivers/docker/docklog/{docklog_test.go => docker_logger_test.go} (81%) rename drivers/docker/docklog/proto/{docklog.pb.go => docker_logger.pb.go} (68%) rename drivers/docker/docklog/proto/{docklog.proto => docker_logger.proto} (95%) diff --git a/drivers/docker/docklog/client.go b/drivers/docker/docklog/client.go index d07cf3339..3bb1b8312 100644 --- a/drivers/docker/docklog/client.go +++ b/drivers/docker/docklog/client.go @@ -6,13 +6,13 @@ import ( "github.com/hashicorp/nomad/drivers/docker/docklog/proto" ) -// docklogClient implements the Docklog interface for client side requests -type docklogClient struct { - client proto.DocklogClient +// dockerLoggerClient implements the dockerLogger interface for client side requests +type dockerLoggerClient struct { + client proto.DockerLoggerClient } // Start proxies the Start client side func to the protobuf interface -func (c *docklogClient) Start(opts *StartOpts) error { +func (c *dockerLoggerClient) Start(opts *StartOpts) error { req := &proto.StartRequest{ Endpoint: opts.Endpoint, ContainerId: opts.ContainerID, @@ -28,7 +28,7 @@ func (c *docklogClient) Start(opts *StartOpts) error { } // Stop proxies the Stop client side func to the protobuf interface -func (c *docklogClient) Stop() error { +func (c *dockerLoggerClient) Stop() error { req := &proto.StopRequest{} _, err := c.client.Stop(context.Background(), req) return err diff --git a/drivers/docker/docklog/docklog.go b/drivers/docker/docklog/docker_logger.go similarity index 91% rename from drivers/docker/docklog/docklog.go rename to drivers/docker/docklog/docker_logger.go index 5557dcb8f..afa1c89c7 100644 --- a/drivers/docker/docklog/docklog.go +++ b/drivers/docker/docklog/docker_logger.go @@ -11,9 +11,9 @@ import ( "golang.org/x/net/context" ) -// Docklog is a small utility to forward logs from a docker container to a target +// DockerLogger is a small utility to forward logs from a docker container to a target // destination -type Docklog interface { +type DockerLogger interface { Start(*StartOpts) error Stop() error } @@ -37,12 +37,12 @@ type StartOpts struct { TLSCA string } -// NewDocklog returns an implementation of the Docklog interface -func NewDocklog(logger hclog.Logger) Docklog { +// NewDockerLogger returns an implementation of the DockerLogger interface +func NewDockerLogger(logger hclog.Logger) DockerLogger { return &dockerLogger{logger: logger} } -// dockerLogger implements the Docklog interface +// dockerLogger implements the DockerLogger interface type dockerLogger struct { logger hclog.Logger diff --git a/drivers/docker/docklog/docklog_test.go b/drivers/docker/docklog/docker_logger_test.go similarity index 81% rename from drivers/docker/docklog/docklog_test.go rename to drivers/docker/docklog/docker_logger_test.go index 7c4848537..9a09d8796 100644 --- a/drivers/docker/docklog/docklog_test.go +++ b/drivers/docker/docklog/docker_logger_test.go @@ -4,7 +4,6 @@ import ( "bytes" "fmt" "testing" - "time" docker "github.com/fsouza/go-dockerclient" "github.com/hashicorp/nomad/helper/testlog" @@ -13,7 +12,7 @@ import ( "golang.org/x/net/context" ) -func TestDocklog(t *testing.T) { +func TestDockerLogger(t *testing.T) { t.Parallel() require := require.New(t) @@ -43,20 +42,23 @@ func TestDocklog(t *testing.T) { err = client.StartContainer(container.ID, nil) require.NoError(err) - var count int - for !container.State.Running { - if count > 10 { - t.Fatal("timeout waiting for container to start") - } - time.Sleep(100 * time.Millisecond) + testutil.WaitForResult(func() (bool, error) { container, err = client.InspectContainer(container.ID) - count++ - } + if err != nil { + return false, err + } + if !container.State.Running { + return false, fmt.Errorf("container not running") + } + return true, nil + }, func(err error) { + require.NoError(err) + }) - stdout := &noopCloser{bytes.NewBufferString("")} - stderr := &noopCloser{bytes.NewBufferString("")} + stdout := &noopCloser{bytes.NewBuffer(nil)} + stderr := &noopCloser{bytes.NewBuffer(nil)} - dl := NewDocklog(testlog.HCLogger(t)).(*dockerLogger) + dl := NewDockerLogger(testlog.HCLogger(t)).(*dockerLogger) dl.stdout = stdout dl.stderr = stderr require.NoError(dl.Start(&StartOpts{ @@ -66,7 +68,6 @@ func TestDocklog(t *testing.T) { echoToContainer(t, client, container.ID, "abc") echoToContainer(t, client, container.ID, "123") - time.Sleep(2000 * time.Millisecond) testutil.WaitForResult(func() (bool, error) { act := stdout.String() if "abc\n123\n" != act { diff --git a/drivers/docker/docklog/plugin.go b/drivers/docker/docklog/plugin.go index 97498525a..b44c88146 100644 --- a/drivers/docker/docklog/plugin.go +++ b/drivers/docker/docklog/plugin.go @@ -12,10 +12,12 @@ import ( "google.golang.org/grpc" ) -// LaunchDocklog launches an instance of docklog +const pluginName = "docker_logger" + +// LaunchDockerLogger launches an instance of DockerLogger // TODO: Integrate with base plugin loader -func LaunchDocklog(logger hclog.Logger) (Docklog, *plugin.Client, error) { - logger = logger.Named("docklog-launcher") +func LaunchDockerLogger(logger hclog.Logger) (DockerLogger, *plugin.Client, error) { + logger = logger.Named(pluginName) bin, err := discover.NomadExecutable() if err != nil { return nil, nil, err @@ -24,9 +26,9 @@ func LaunchDocklog(logger hclog.Logger) (Docklog, *plugin.Client, error) { client := plugin.NewClient(&plugin.ClientConfig{ HandshakeConfig: base.Handshake, Plugins: map[string]plugin.Plugin{ - "docklog": &Plugin{impl: NewDocklog(hclog.L().Named("docklog"))}, + pluginName: &Plugin{impl: NewDockerLogger(hclog.L().Named(pluginName))}, }, - Cmd: exec.Command(bin, "docklog"), + Cmd: exec.Command(bin, pluginName), AllowedProtocols: []plugin.Protocol{ plugin.ProtocolGRPC, }, @@ -37,12 +39,12 @@ func LaunchDocklog(logger hclog.Logger) (Docklog, *plugin.Client, error) { return nil, nil, err } - raw, err := rpcClient.Dispense("docklog") + raw, err := rpcClient.Dispense(pluginName) if err != nil { return nil, nil, err } - l := raw.(Docklog) + l := raw.(DockerLogger) return l, client, nil } @@ -50,12 +52,12 @@ func LaunchDocklog(logger hclog.Logger) (Docklog, *plugin.Client, error) { // Plugin is the go-plugin implementation type Plugin struct { plugin.NetRPCUnsupportedPlugin - impl Docklog + impl DockerLogger } // GRPCServer registered the server side implementation with the grpc server func (p *Plugin) GRPCServer(broker *plugin.GRPCBroker, s *grpc.Server) error { - proto.RegisterDocklogServer(s, &docklogServer{ + proto.RegisterDockerLoggerServer(s, &dockerLoggerServer{ impl: p.impl, broker: broker, }) @@ -64,5 +66,5 @@ func (p *Plugin) GRPCServer(broker *plugin.GRPCBroker, s *grpc.Server) error { // GRPCClient returns a client side implementation of the plugin func (p *Plugin) GRPCClient(ctx context.Context, broker *plugin.GRPCBroker, c *grpc.ClientConn) (interface{}, error) { - return &docklogClient{client: proto.NewDocklogClient(c)}, nil + return &dockerLoggerClient{client: proto.NewDockerLoggerClient(c)}, nil } diff --git a/drivers/docker/docklog/proto/docklog.pb.go b/drivers/docker/docklog/proto/docker_logger.pb.go similarity index 68% rename from drivers/docker/docklog/proto/docklog.pb.go rename to drivers/docker/docklog/proto/docker_logger.pb.go index 506455821..9de0a1de7 100644 --- a/drivers/docker/docklog/proto/docklog.pb.go +++ b/drivers/docker/docklog/proto/docker_logger.pb.go @@ -1,5 +1,5 @@ // Code generated by protoc-gen-go. DO NOT EDIT. -// source: docklog.proto +// source: docker_logger.proto package proto @@ -40,7 +40,7 @@ func (m *StartRequest) Reset() { *m = StartRequest{} } func (m *StartRequest) String() string { return proto.CompactTextString(m) } func (*StartRequest) ProtoMessage() {} func (*StartRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_docklog_6cb3edde8d295224, []int{0} + return fileDescriptor_docker_logger_54dce3369d4ecf54, []int{0} } func (m *StartRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StartRequest.Unmarshal(m, b) @@ -119,7 +119,7 @@ func (m *StartResponse) Reset() { *m = StartResponse{} } func (m *StartResponse) String() string { return proto.CompactTextString(m) } func (*StartResponse) ProtoMessage() {} func (*StartResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_docklog_6cb3edde8d295224, []int{1} + return fileDescriptor_docker_logger_54dce3369d4ecf54, []int{1} } func (m *StartResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StartResponse.Unmarshal(m, b) @@ -149,7 +149,7 @@ func (m *StopRequest) Reset() { *m = StopRequest{} } func (m *StopRequest) String() string { return proto.CompactTextString(m) } func (*StopRequest) ProtoMessage() {} func (*StopRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_docklog_6cb3edde8d295224, []int{2} + return fileDescriptor_docker_logger_54dce3369d4ecf54, []int{2} } func (m *StopRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StopRequest.Unmarshal(m, b) @@ -179,7 +179,7 @@ func (m *StopResponse) Reset() { *m = StopResponse{} } func (m *StopResponse) String() string { return proto.CompactTextString(m) } func (*StopResponse) ProtoMessage() {} func (*StopResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_docklog_6cb3edde8d295224, []int{3} + return fileDescriptor_docker_logger_54dce3369d4ecf54, []int{3} } func (m *StopResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StopResponse.Unmarshal(m, b) @@ -214,125 +214,125 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// DocklogClient is the client API for Docklog service. +// DockerLoggerClient is the client API for DockerLogger service. // // For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. -type DocklogClient interface { +type DockerLoggerClient interface { Start(ctx context.Context, in *StartRequest, opts ...grpc.CallOption) (*StartResponse, error) Stop(ctx context.Context, in *StopRequest, opts ...grpc.CallOption) (*StopResponse, error) } -type docklogClient struct { +type dockerLoggerClient struct { cc *grpc.ClientConn } -func NewDocklogClient(cc *grpc.ClientConn) DocklogClient { - return &docklogClient{cc} +func NewDockerLoggerClient(cc *grpc.ClientConn) DockerLoggerClient { + return &dockerLoggerClient{cc} } -func (c *docklogClient) Start(ctx context.Context, in *StartRequest, opts ...grpc.CallOption) (*StartResponse, error) { +func (c *dockerLoggerClient) Start(ctx context.Context, in *StartRequest, opts ...grpc.CallOption) (*StartResponse, error) { out := new(StartResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.drivers.docker.docklog.proto.Docklog/Start", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.drivers.docker.docklog.proto.DockerLogger/Start", in, out, opts...) if err != nil { return nil, err } return out, nil } -func (c *docklogClient) Stop(ctx context.Context, in *StopRequest, opts ...grpc.CallOption) (*StopResponse, error) { +func (c *dockerLoggerClient) Stop(ctx context.Context, in *StopRequest, opts ...grpc.CallOption) (*StopResponse, error) { out := new(StopResponse) - err := c.cc.Invoke(ctx, "/hashicorp.nomad.drivers.docker.docklog.proto.Docklog/Stop", in, out, opts...) + err := c.cc.Invoke(ctx, "/hashicorp.nomad.drivers.docker.docklog.proto.DockerLogger/Stop", in, out, opts...) if err != nil { return nil, err } return out, nil } -// DocklogServer is the server API for Docklog service. -type DocklogServer interface { +// DockerLoggerServer is the server API for DockerLogger service. +type DockerLoggerServer interface { Start(context.Context, *StartRequest) (*StartResponse, error) Stop(context.Context, *StopRequest) (*StopResponse, error) } -func RegisterDocklogServer(s *grpc.Server, srv DocklogServer) { - s.RegisterService(&_Docklog_serviceDesc, srv) +func RegisterDockerLoggerServer(s *grpc.Server, srv DockerLoggerServer) { + s.RegisterService(&_DockerLogger_serviceDesc, srv) } -func _Docklog_Start_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { +func _DockerLogger_Start_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { in := new(StartRequest) if err := dec(in); err != nil { return nil, err } if interceptor == nil { - return srv.(DocklogServer).Start(ctx, in) + return srv.(DockerLoggerServer).Start(ctx, in) } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.drivers.docker.docklog.proto.Docklog/Start", + FullMethod: "/hashicorp.nomad.drivers.docker.docklog.proto.DockerLogger/Start", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { - return srv.(DocklogServer).Start(ctx, req.(*StartRequest)) + return srv.(DockerLoggerServer).Start(ctx, req.(*StartRequest)) } return interceptor(ctx, in, info, handler) } -func _Docklog_Stop_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { +func _DockerLogger_Stop_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { in := new(StopRequest) if err := dec(in); err != nil { return nil, err } if interceptor == nil { - return srv.(DocklogServer).Stop(ctx, in) + return srv.(DockerLoggerServer).Stop(ctx, in) } info := &grpc.UnaryServerInfo{ Server: srv, - FullMethod: "/hashicorp.nomad.drivers.docker.docklog.proto.Docklog/Stop", + FullMethod: "/hashicorp.nomad.drivers.docker.docklog.proto.DockerLogger/Stop", } handler := func(ctx context.Context, req interface{}) (interface{}, error) { - return srv.(DocklogServer).Stop(ctx, req.(*StopRequest)) + return srv.(DockerLoggerServer).Stop(ctx, req.(*StopRequest)) } return interceptor(ctx, in, info, handler) } -var _Docklog_serviceDesc = grpc.ServiceDesc{ - ServiceName: "hashicorp.nomad.drivers.docker.docklog.proto.Docklog", - HandlerType: (*DocklogServer)(nil), +var _DockerLogger_serviceDesc = grpc.ServiceDesc{ + ServiceName: "hashicorp.nomad.drivers.docker.docklog.proto.DockerLogger", + HandlerType: (*DockerLoggerServer)(nil), Methods: []grpc.MethodDesc{ { MethodName: "Start", - Handler: _Docklog_Start_Handler, + Handler: _DockerLogger_Start_Handler, }, { MethodName: "Stop", - Handler: _Docklog_Stop_Handler, + Handler: _DockerLogger_Stop_Handler, }, }, Streams: []grpc.StreamDesc{}, - Metadata: "docklog.proto", + Metadata: "docker_logger.proto", } -func init() { proto.RegisterFile("docklog.proto", fileDescriptor_docklog_6cb3edde8d295224) } +func init() { proto.RegisterFile("docker_logger.proto", fileDescriptor_docker_logger_54dce3369d4ecf54) } -var fileDescriptor_docklog_6cb3edde8d295224 = []byte{ - // 311 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x90, 0xb1, 0x4e, 0xf3, 0x40, - 0x10, 0x84, 0x7f, 0xe7, 0x8f, 0xed, 0xb0, 0x49, 0x40, 0x3a, 0x09, 0x61, 0xd2, 0x00, 0xae, 0x28, - 0xd0, 0x15, 0x50, 0x11, 0x3a, 0x40, 0x48, 0x88, 0x2e, 0xe9, 0x68, 0x2c, 0xe3, 0x3b, 0x93, 0x53, - 0x8c, 0xd7, 0xec, 0x6d, 0x90, 0x52, 0x21, 0xf1, 0x02, 0xbc, 0x1f, 0x4f, 0x83, 0x7c, 0x36, 0x16, - 0x94, 0x49, 0x65, 0xef, 0xcc, 0x37, 0xa7, 0x9d, 0x85, 0xb1, 0xc2, 0x6c, 0x59, 0xe0, 0xb3, 0xac, - 0x08, 0x19, 0xc5, 0xd9, 0x22, 0xb5, 0x0b, 0x93, 0x21, 0x55, 0xb2, 0xc4, 0x97, 0x54, 0x49, 0x45, - 0xe6, 0x4d, 0x93, 0x95, 0x35, 0xa6, 0x49, 0xfe, 0xa1, 0xe3, 0x2f, 0x0f, 0x46, 0x73, 0x4e, 0x89, - 0x67, 0xfa, 0x75, 0xa5, 0x2d, 0x8b, 0x09, 0x0c, 0x74, 0xa9, 0x2a, 0x34, 0x25, 0x47, 0xde, 0xb1, - 0x77, 0xba, 0x33, 0xeb, 0x66, 0x71, 0x02, 0xa3, 0x0c, 0x4b, 0x4e, 0x4d, 0xa9, 0x29, 0x31, 0x2a, - 0xea, 0x39, 0x7f, 0xd8, 0x69, 0xf7, 0x4a, 0x1c, 0xc1, 0xd0, 0xb2, 0xc2, 0x15, 0x27, 0xb9, 0xc9, - 0x31, 0xfa, 0xef, 0x08, 0x68, 0xa4, 0x3b, 0x93, 0x63, 0x0b, 0x68, 0xa2, 0x06, 0xe8, 0x77, 0x80, - 0x26, 0x72, 0xc0, 0x21, 0x0c, 0xb8, 0xb0, 0x49, 0xa6, 0x89, 0x23, 0xdf, 0xb9, 0x21, 0x17, 0xf6, - 0x46, 0x13, 0x8b, 0x03, 0xa8, 0x7f, 0x93, 0xa5, 0x5e, 0x47, 0x81, 0x73, 0x02, 0x2e, 0xec, 0x83, - 0x5e, 0x8b, 0x7d, 0x08, 0x5c, 0x26, 0x8d, 0x42, 0xa7, 0xfb, 0x75, 0x22, 0x8d, 0xf7, 0x60, 0xdc, - 0x76, 0xb3, 0x15, 0x96, 0x56, 0xc7, 0x63, 0x18, 0xce, 0x19, 0xab, 0xb6, 0x6b, 0xbc, 0x5b, 0x77, - 0xaf, 0xc7, 0xc6, 0x3e, 0xff, 0xec, 0x41, 0x78, 0xdb, 0x9c, 0x47, 0x7c, 0x78, 0xe0, 0xbb, 0xb0, - 0x98, 0xca, 0x4d, 0x2e, 0x2a, 0x7f, 0x5f, 0x73, 0x72, 0xb5, 0x55, 0xb6, 0xdd, 0xf6, 0x9f, 0x78, - 0x87, 0x7e, 0xbd, 0xa0, 0xb8, 0xdc, 0xf4, 0x99, 0xae, 0xe3, 0x64, 0xba, 0x4d, 0xf4, 0x67, 0x81, - 0xeb, 0xf0, 0xd1, 0x77, 0xfa, 0x53, 0xe0, 0x3e, 0x17, 0xdf, 0x01, 0x00, 0x00, 0xff, 0xff, 0xd7, - 0xfa, 0x4e, 0xd3, 0x6d, 0x02, 0x00, 0x00, +var fileDescriptor_docker_logger_54dce3369d4ecf54 = []byte{ + // 319 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x50, 0xbd, 0x4e, 0xe3, 0x40, + 0x10, 0x3e, 0xe7, 0x62, 0x27, 0x37, 0x49, 0xee, 0xa4, 0x3d, 0x21, 0x4c, 0x1a, 0xc0, 0x15, 0x05, + 0x72, 0x01, 0x15, 0xa1, 0x03, 0x84, 0x84, 0xa0, 0x4a, 0x3a, 0x1a, 0xcb, 0x78, 0x27, 0xc9, 0x2a, + 0xc6, 0x63, 0x66, 0x27, 0x48, 0xa9, 0x90, 0x78, 0x06, 0x9e, 0x8e, 0xa7, 0x41, 0x5e, 0x1b, 0x8b, + 0x36, 0xa9, 0x76, 0xe7, 0xfb, 0x19, 0xcd, 0xf7, 0xc1, 0x7f, 0x4d, 0xd9, 0x0a, 0x39, 0xc9, 0x69, + 0xb1, 0x40, 0x8e, 0x4b, 0x26, 0x21, 0x75, 0xba, 0x4c, 0xed, 0xd2, 0x64, 0xc4, 0x65, 0x5c, 0xd0, + 0x73, 0xaa, 0x63, 0xcd, 0xe6, 0x15, 0xd9, 0xc6, 0xb5, 0xd8, 0x3d, 0x39, 0x2d, 0x6a, 0x75, 0xf4, + 0xe9, 0xc1, 0x70, 0x26, 0x29, 0xcb, 0x14, 0x5f, 0xd6, 0x68, 0x45, 0x8d, 0xa1, 0x8f, 0x85, 0x2e, + 0xc9, 0x14, 0x12, 0x7a, 0x47, 0xde, 0xc9, 0x9f, 0x69, 0x3b, 0xab, 0x63, 0x18, 0x66, 0x54, 0x48, + 0x6a, 0x0a, 0xe4, 0xc4, 0xe8, 0xb0, 0xe3, 0xf8, 0x41, 0x8b, 0xdd, 0x69, 0x75, 0x08, 0x03, 0x2b, + 0x9a, 0xd6, 0x92, 0xcc, 0xcd, 0x9c, 0xc2, 0xdf, 0x4e, 0x01, 0x35, 0x74, 0x6b, 0xe6, 0xd4, 0x08, + 0x90, 0xb9, 0x16, 0x74, 0x5b, 0x01, 0x32, 0x3b, 0xc1, 0x01, 0xf4, 0x25, 0xb7, 0x49, 0x86, 0x2c, + 0xa1, 0xef, 0xd8, 0x9e, 0xe4, 0xf6, 0x1a, 0x59, 0xd4, 0x3e, 0x54, 0xdf, 0x64, 0x85, 0x9b, 0x30, + 0x70, 0x4c, 0x20, 0xb9, 0xbd, 0xc7, 0x8d, 0xda, 0x83, 0xc0, 0x79, 0xd2, 0xb0, 0xe7, 0x70, 0xbf, + 0x72, 0xa4, 0xd1, 0x3f, 0x18, 0x35, 0xd9, 0x6c, 0x49, 0x85, 0xc5, 0x68, 0x04, 0x83, 0x99, 0x50, + 0xd9, 0x64, 0x8d, 0xfe, 0x56, 0xd9, 0xab, 0xb1, 0xa6, 0xcf, 0x3e, 0x3a, 0x30, 0xbc, 0x71, 0x2d, + 0x3d, 0xb8, 0x46, 0xd5, 0xbb, 0x07, 0xbe, 0xdb, 0xa0, 0x26, 0xf1, 0x36, 0xb5, 0xc6, 0x3f, 0x2b, + 0x1d, 0x5f, 0xee, 0xe4, 0x6d, 0x4e, 0xfe, 0xa5, 0xde, 0xa0, 0x5b, 0x5d, 0xa9, 0x2e, 0xb6, 0x5d, + 0xd3, 0x06, 0x1d, 0x4f, 0x76, 0xb1, 0x7e, 0x1f, 0x70, 0xd5, 0x7b, 0xf4, 0x1d, 0xfe, 0x14, 0xb8, + 0xe7, 0xfc, 0x2b, 0x00, 0x00, 0xff, 0xff, 0x2e, 0xe1, 0xdf, 0xc9, 0x78, 0x02, 0x00, 0x00, } diff --git a/drivers/docker/docklog/proto/docklog.proto b/drivers/docker/docklog/proto/docker_logger.proto similarity index 95% rename from drivers/docker/docklog/proto/docklog.proto rename to drivers/docker/docklog/proto/docker_logger.proto index ef77ea1c2..7f71c657e 100644 --- a/drivers/docker/docklog/proto/docklog.proto +++ b/drivers/docker/docklog/proto/docker_logger.proto @@ -2,7 +2,7 @@ syntax = "proto3"; package hashicorp.nomad.drivers.docker.docklog.proto; option go_package = "proto"; -service Docklog { +service DockerLogger { rpc Start(StartRequest) returns (StartResponse) {} rpc Stop(StopRequest) returns (StopResponse) {} } diff --git a/drivers/docker/docklog/server.go b/drivers/docker/docklog/server.go index 2602e2163..1bbf09f03 100644 --- a/drivers/docker/docklog/server.go +++ b/drivers/docker/docklog/server.go @@ -7,14 +7,14 @@ import ( "github.com/hashicorp/nomad/drivers/docker/docklog/proto" ) -// docklogServer is the server side translation between the protobuf and native interfaces -type docklogServer struct { +// dockerLoggerServer is the server side translation between the protobuf and native interfaces +type dockerLoggerServer struct { broker *plugin.GRPCBroker - impl Docklog + impl DockerLogger } -// Start proxies the protobuf Start RPC to the Start fun of the Docklog interface -func (s *docklogServer) Start(ctx context.Context, req *proto.StartRequest) (*proto.StartResponse, error) { +// Start proxies the protobuf Start RPC to the Start fun of the DockerLogger interface +func (s *dockerLoggerServer) Start(ctx context.Context, req *proto.StartRequest) (*proto.StartResponse, error) { opts := &StartOpts{ Endpoint: req.Endpoint, ContainerID: req.ContainerId, @@ -33,7 +33,7 @@ func (s *docklogServer) Start(ctx context.Context, req *proto.StartRequest) (*pr return resp, nil } -// Stop proxies the protobuf Stop RPC to the Stop fun of the Docklog interface -func (s *docklogServer) Stop(ctx context.Context, req *proto.StopRequest) (*proto.StopResponse, error) { +// Stop proxies the protobuf Stop RPC to the Stop fun of the DockerLogger interface +func (s *dockerLoggerServer) Stop(ctx context.Context, req *proto.StopRequest) (*proto.StopResponse, error) { return &proto.StopResponse{}, s.impl.Stop() } From e27c82ea4d802027a89123674dc80c0216e3e91b Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 16 Oct 2018 16:17:24 -0700 Subject: [PATCH 148/152] client: remove unused handleproxy --- client/allocrunner/taskrunner/handleproxy.go | 49 ------------ .../taskrunner/handleproxy_test.go | 76 ------------------- 2 files changed, 125 deletions(-) delete mode 100644 client/allocrunner/taskrunner/handleproxy.go delete mode 100644 client/allocrunner/taskrunner/handleproxy_test.go diff --git a/client/allocrunner/taskrunner/handleproxy.go b/client/allocrunner/taskrunner/handleproxy.go deleted file mode 100644 index 3b24f7d2f..000000000 --- a/client/allocrunner/taskrunner/handleproxy.go +++ /dev/null @@ -1,49 +0,0 @@ -package taskrunner - -import ( - "context" - - "github.com/hashicorp/nomad/client/driver/structs" -) - -// handleResult multiplexes a single WaitResult to multiple waiters. Useful -// because DriverHandle.WaitCh is closed after it returns a single WaitResult. -type handleResult struct { - doneCh <-chan struct{} - result *structs.WaitResult -} - -func newHandleResult(waitCh <-chan *structs.WaitResult) *handleResult { - doneCh := make(chan struct{}) - - h := &handleResult{ - doneCh: doneCh, - } - - go func() { - // Wait for result - res := <-waitCh - - // Set result - h.result = res - - // Notify waiters - close(doneCh) - - }() - - return h -} - -// Wait blocks until a task's result is available or the passed-in context is -// canceled. Safe for concurrent callers. -func (h *handleResult) Wait(ctx context.Context) *structs.WaitResult { - // Block until done or canceled - select { - case <-h.doneCh: - case <-ctx.Done(): - return nil - } - - return h.result -} diff --git a/client/allocrunner/taskrunner/handleproxy_test.go b/client/allocrunner/taskrunner/handleproxy_test.go deleted file mode 100644 index b43029407..000000000 --- a/client/allocrunner/taskrunner/handleproxy_test.go +++ /dev/null @@ -1,76 +0,0 @@ -package taskrunner - -import ( - "context" - "testing" - "time" - - "github.com/hashicorp/nomad/client/driver/structs" - "github.com/stretchr/testify/require" -) - -// TestHandleResult_Wait_Result asserts multiple waiters on a handleResult all -// receive the wait result. -func TestHandleResult_Wait_Result(t *testing.T) { - t.Parallel() - - waitCh := make(chan *structs.WaitResult) - h := newHandleResult(waitCh) - - outCh1 := make(chan *structs.WaitResult) - outCh2 := make(chan *structs.WaitResult) - - // Create two receivers - go func() { - outCh1 <- h.Wait(context.Background()) - }() - go func() { - outCh2 <- h.Wait(context.Background()) - }() - - // Send a single result - go func() { - waitCh <- &structs.WaitResult{ExitCode: 1} - }() - - // Assert both receivers got the result - assert := func(outCh chan *structs.WaitResult) { - select { - case result := <-outCh: - require.NotNil(t, result) - require.Equal(t, 1, result.ExitCode) - case <-time.After(time.Second): - t.Fatalf("timeout waiting for result") - } - } - - assert(outCh1) - assert(outCh2) -} - -// TestHandleResult_Wait_Cancel asserts that canceling the context unblocks the -// waiter. -func TestHandleResult_Wait_Cancel(t *testing.T) { - t.Parallel() - - waitCh := make(chan *structs.WaitResult) - h := newHandleResult(waitCh) - - ctx, cancel := context.WithCancel(context.Background()) - outCh := make(chan *structs.WaitResult) - - go func() { - outCh <- h.Wait(ctx) - }() - - // Cancelling the context should unblock the Wait - cancel() - - // Assert the result is nil - select { - case result := <-outCh: - require.Nil(t, result) - case <-time.After(time.Second): - t.Fatalf("timeout waiting for result") - } -} From bb5b3da7fb1d6d76d5bfb5a5ac0f9e68ac4aee38 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 16 Oct 2018 16:31:22 -0700 Subject: [PATCH 149/152] changelog: add client refactor feature branch merge --- CHANGELOG.md | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/CHANGELOG.md b/CHANGELOG.md index 60f849bff..d64a665b2 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,11 +1,15 @@ ## 0.9.0 (Unreleased) +__BACKWARDS INCOMPATIBILITIES:__ + * core: Switch to structured logging using [go-hclog](https://github.com/hashicorp/go-hclog) + IMPROVEMENTS: * core: Added advertise address to client node meta data [[GH-4390](https://github.com/hashicorp/nomad/issues/4390)] * core: Added support for specifying node affinities. Affinities allow job operators to specify weighted placement preferences according to different node attributes [[GH-4512](https://github.com/hashicorp/nomad/issues/4512)] * core: Added support for spreading allocations across a specific attribute. Operators can specify spread target percentages across failure domains such as datacenter or rack [[GH-4512](https://github.com/hashicorp/nomad/issues/4512)] + * client: Refactor client to support plugins and improve state handling [[GH-4792](https://github.com/hashicorp/nomad/pull/4792)] * client: Extend timeout to 60 seconds for Windows CPU fingerprinting [[GH-4441](https://github.com/hashicorp/nomad/pull/4441)] * driver/docker: Add support for specifying `cpu_cfs_period` in the Docker driver [[GH-4462](https://github.com/hashicorp/nomad/issues/4462)] * telemetry: All client metrics include a new `node_class` tag [[GH-3882](https://github.com/hashicorp/nomad/issues/3882)] From 838ddf4d4ad49e40dfe3f07d34e1361dac5c7279 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 16 Oct 2018 16:42:19 -0700 Subject: [PATCH 150/152] fix linter errors --- client/allocrunner/taskrunner/lifecycle.go | 2 +- client/allocrunner/taskrunner/task_runner.go | 2 +- drivers/rkt/driver.go | 4 ++-- 3 files changed, 4 insertions(+), 4 deletions(-) diff --git a/client/allocrunner/taskrunner/lifecycle.go b/client/allocrunner/taskrunner/lifecycle.go index a561ec8da..7a3146480 100644 --- a/client/allocrunner/taskrunner/lifecycle.go +++ b/client/allocrunner/taskrunner/lifecycle.go @@ -91,7 +91,7 @@ func (tr *TaskRunner) Kill(ctx context.Context, event *structs.TaskEvent) error waitCh, err := handle.WaitCh(ctx) // The error should be nil or TaskNotFound, if it's something else then a - // failure in the driver or transport layer occured + // failure in the driver or transport layer occurred if err != nil { if err == drivers.ErrTaskNotFound { return nil diff --git a/client/allocrunner/taskrunner/task_runner.go b/client/allocrunner/taskrunner/task_runner.go index 49c560b50..8dffd65a4 100644 --- a/client/allocrunner/taskrunner/task_runner.go +++ b/client/allocrunner/taskrunner/task_runner.go @@ -735,7 +735,7 @@ func (tr *TaskRunner) appendEvent(event *structs.TaskEvent) error { // Ensure the event is populated with human readable strings event.PopulateEventDisplayMessage() - // Propogate failure from event to task state + // Propagate failure from event to task state if event.FailsTask { tr.state.Failed = true } diff --git a/drivers/rkt/driver.go b/drivers/rkt/driver.go index a06c8b9d0..211a1075a 100644 --- a/drivers/rkt/driver.go +++ b/drivers/rkt/driver.go @@ -462,10 +462,10 @@ func (d *RktDriver) StartTask(cfg *drivers.TaskConfig) (*drivers.TaskHandle, *cs } // Add memory isolator - prepareArgs = append(prepareArgs, fmt.Sprintf("--memory=%v", int64(cfg.Resources.LinuxResources.MemoryLimitBytes))) + prepareArgs = append(prepareArgs, fmt.Sprintf("--memory=%v", cfg.Resources.LinuxResources.MemoryLimitBytes)) // Add CPU isolator - prepareArgs = append(prepareArgs, fmt.Sprintf("--cpu-shares=%v", int64(cfg.Resources.LinuxResources.CPUShares))) + prepareArgs = append(prepareArgs, fmt.Sprintf("--cpu-shares=%v", cfg.Resources.LinuxResources.CPUShares)) // Add DNS servers if len(driverConfig.DNSServers) == 1 && (driverConfig.DNSServers[0] == "host" || driverConfig.DNSServers[0] == "none") { From 0baaba8b09f77e7491f9110b18b6de335cbbf7e1 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 16 Oct 2018 16:44:18 -0700 Subject: [PATCH 151/152] templates: fix tests --- client/allocrunner/taskrunner/template/template_test.go | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/client/allocrunner/taskrunner/template/template_test.go b/client/allocrunner/taskrunner/template/template_test.go index f4fefeb8f..8561a8f2f 100644 --- a/client/allocrunner/taskrunner/template/template_test.go +++ b/client/allocrunner/taskrunner/template/template_test.go @@ -34,7 +34,7 @@ type MockTaskHooks struct { Restarts int RestartCh chan struct{} - Signals []os.Signal + Signals []string SignalCh chan struct{} // SignalError is returned when Signal is called on the mock hook @@ -67,7 +67,7 @@ func (m *MockTaskHooks) Restart(ctx context.Context, event *structs.TaskEvent, f return nil } -func (m *MockTaskHooks) Signal(event *structs.TaskEvent, s os.Signal) error { +func (m *MockTaskHooks) Signal(event *structs.TaskEvent, s string) error { m.Signals = append(m.Signals, s) select { case m.SignalCh <- struct{}{}: From 2b1b3d7e1ec7dca0636cc5e872d4499b07dc8788 Mon Sep 17 00:00:00 2001 From: Michael Schurter Date: Tue, 16 Oct 2018 16:52:58 -0700 Subject: [PATCH 152/152] tests: get tests building if not yet passing --- command/agent/consul/script_test.go | 8 +++++--- command/agent/consul/unit_test.go | 4 +++- plugins/drivers/plugin_test.go | 7 ++++--- plugins/drivers/testing_test.go | 7 ++++--- 4 files changed, 16 insertions(+), 10 deletions(-) diff --git a/command/agent/consul/script_test.go b/command/agent/consul/script_test.go index 4a029ca0f..3cc833753 100644 --- a/command/agent/consul/script_test.go +++ b/command/agent/consul/script_test.go @@ -34,8 +34,10 @@ func newBlockingScriptExec() *blockingScriptExec { return &blockingScriptExec{running: make(chan struct{})} } -func (b *blockingScriptExec) Exec(ctx context.Context, _ string, _ []string) ([]byte, int, error) { +func (b *blockingScriptExec) Exec(dur time.Duration, _ string, _ []string) ([]byte, int, error) { b.running <- struct{}{} + ctx, cancel := context.WithTimeout(context.Background(), dur) + defer cancel() cmd := exec.CommandContext(ctx, testtask.Path(), "sleep", "9000h") testtask.SetCmdEnv(cmd) err := cmd.Run() @@ -145,7 +147,7 @@ func TestConsulScript_Exec_Timeout(t *testing.T) { // sleeperExec sleeps for 100ms but returns successfully to allow testing timeout conditions type sleeperExec struct{} -func (sleeperExec) Exec(context.Context, string, []string) ([]byte, int, error) { +func (sleeperExec) Exec(time.Duration, string, []string) ([]byte, int, error) { time.Sleep(100 * time.Millisecond) return []byte{}, 0, nil } @@ -185,7 +187,7 @@ type simpleExec struct { err error } -func (s simpleExec) Exec(context.Context, string, []string) ([]byte, int, error) { +func (s simpleExec) Exec(time.Duration, string, []string) ([]byte, int, error) { return []byte(fmt.Sprintf("code=%d err=%v", s.code, s.err)), s.code, s.err } diff --git a/command/agent/consul/unit_test.go b/command/agent/consul/unit_test.go index 25dbf2bfb..5626b8403 100644 --- a/command/agent/consul/unit_test.go +++ b/command/agent/consul/unit_test.go @@ -67,7 +67,7 @@ func newMockExec() *mockExec { } } -func (m *mockExec) Exec(ctx context.Context, cmd string, args []string) ([]byte, int, error) { +func (m *mockExec) Exec(dur time.Duration, cmd string, args []string) ([]byte, int, error) { select { case m.execs <- 1: default: @@ -76,6 +76,8 @@ func (m *mockExec) Exec(ctx context.Context, cmd string, args []string) ([]byte, // Default impl is just "ok" return []byte("ok"), 0, nil } + ctx, cancel := context.WithTimeout(context.Background(), dur) + defer cancel() return m.ExecFunc(ctx, cmd, args) } diff --git a/plugins/drivers/plugin_test.go b/plugins/drivers/plugin_test.go index c08d14150..ce15d074b 100644 --- a/plugins/drivers/plugin_test.go +++ b/plugins/drivers/plugin_test.go @@ -6,6 +6,7 @@ import ( "testing" "time" + cstructs "github.com/hashicorp/nomad/client/structs" "github.com/hashicorp/nomad/nomad/structs" "github.com/stretchr/testify/require" "github.com/ugorji/go/codec" @@ -118,18 +119,18 @@ func TestBaseDriver_StartTask(t *testing.T) { state := &testDriverState{Pid: 1, Log: "log"} var handle *TaskHandle impl := &MockDriver{ - StartTaskF: func(c *TaskConfig) (*TaskHandle, error) { + StartTaskF: func(c *TaskConfig) (*TaskHandle, *cstructs.DriverNetwork, error) { handle = NewTaskHandle("test") handle.Config = c handle.State = TaskStateRunning handle.SetDriverState(state) - return handle, nil + return handle, nil, nil }, } harness := NewDriverHarness(t, impl) defer harness.Kill() - resp, err := harness.StartTask(cfg) + resp, _, err := harness.StartTask(cfg) require.NoError(err) require.Equal(cfg.ID, resp.Config.ID) require.Equal(handle.State, resp.State) diff --git a/plugins/drivers/testing_test.go b/plugins/drivers/testing_test.go index ac4141f8d..faa83c616 100644 --- a/plugins/drivers/testing_test.go +++ b/plugins/drivers/testing_test.go @@ -3,6 +3,7 @@ package drivers import ( "testing" + cstructs "github.com/hashicorp/nomad/client/structs" "github.com/stretchr/testify/require" ) @@ -12,13 +13,13 @@ var _ DriverPlugin = (*MockDriver)(nil) func TestDriverHarness(t *testing.T) { handle := &TaskHandle{Config: &TaskConfig{Name: "mock"}} d := &MockDriver{ - StartTaskF: func(task *TaskConfig) (*TaskHandle, error) { - return handle, nil + StartTaskF: func(task *TaskConfig) (*TaskHandle, *cstructs.DriverNetwork, error) { + return handle, nil, nil }, } harness := NewDriverHarness(t, d) defer harness.Kill() - actual, err := harness.StartTask(&TaskConfig{}) + actual, _, err := harness.StartTask(&TaskConfig{}) require.NoError(t, err) require.Equal(t, handle.Config.Name, actual.Config.Name) }